实用的图像/视频修复算法。

Xintao 685d429c81 v0.2.5.0 3 years ago
.github 35ee6f781e improve codes comments 3 years ago
.vscode 571b89257a add no-response workflow, vscode format setting, update requirements 3 years ago
assets 2b4e485eb0 Update ReadMe (#259) 3 years ago
docs 13c95fe094 update readme 3 years ago
experiments 248cbedbce add readme for training 3 years ago
inputs 20355e0c79 Update readme for anime video models; add video demo (#181) 3 years ago
options 35ee6f781e improve codes comments 3 years ago
realesrgan 192f672f91 add inference_realesrgan_video 3 years ago
scripts 7dd860a881 catch more specific errors 3 years ago
tests 42110857ef add unittest for model and utils 3 years ago
.gitignore 571b89257a add no-response workflow, vscode format setting, update requirements 3 years ago
.pre-commit-config.yaml 772923e207 add codespell to pre-commit hook 3 years ago
CODE_OF_CONDUCT.md 01aeba2f7a Add CODE_OF_CONDUCT.md 3 years ago
CONTRIBUTING.md c9023b3d7a Update README_CN.md (#142) 3 years ago
FAQ.md cddc2ff658 update readme 3 years ago
LICENSE 0573f32dd0 Create LICENSE 3 years ago
MANIFEST.in 32a4fa1772 add publish-pip action 3 years ago
README.md 13c95fe094 update readme 3 years ago
README_CN.md 13c95fe094 update readme 3 years ago
Training.md 501efe3da6 update ReadMe 3 years ago
VERSION 685d429c81 v0.2.5.0 3 years ago
feedback.md c9023b3d7a Update README_CN.md (#142) 3 years ago
inference_realesrgan.py cddc2ff658 update readme 3 years ago
inference_realesrgan_video.py 98add035f2 support realesr-animevideov3 3 years ago
requirements.txt f5ccd64ce5 support finetune with paired data 3 years ago
setup.cfg 1d180efaf3 add unittest for dataset and archs 3 years ago
setup.py 3338b31f48 update setup.py, V0.2.2.5 3 years ago

README.md

download PyPI Open issue Closed issue LICENSE python lint Publish-pip

:fire: Update the RealESRGAN AnimeVideo-v3 model 更新动漫视频的小模型. Please see anime video models and comparisons for more details.

  1. Colab Demo for Real-ESRGAN google colab logo.
  2. Colab Demo for Real-ESRGAN (anime videos) google colab logo.
  3. Portable Windows / Linux / MacOS executable files for Intel/AMD/Nvidia GPU. You can find more information here. The ncnn implementation is in Real-ESRGAN-ncnn-vulkan.

Real-ESRGAN aims at developing Practical Algorithms for General Image/Video Restoration.
We extend the powerful ESRGAN to a practical restoration application (namely, Real-ESRGAN), which is trained with pure synthetic data.

:art: Real-ESRGAN needs your contributions. Any contributions are welcome, such as new features/models/typo fixes/suggestions/maintenance, etc. See CONTRIBUTING.md. All contributors are list here.

:question: Frequently Asked Questions can be found in FAQ.md.

:milky_way: Thanks for your valuable feedbacks/suggestions. All the feedbacks are updated in feedback.md.


If Real-ESRGAN is helpful in your photos/projects, please help to :star: this repo or recommend it to your friends. Thanks:blush:
Other recommended projects:
:arrow_forward: GFPGAN: A practical algorithm for real-world face restoration
:arrow_forward: BasicSR: An open-source image and video restoration toolbox
:arrow_forward: facexlib: A collection that provides useful face-relation functions.
:arrow_forward: HandyView: A PyQt5-based image viewer that is handy for view and comparison.


🚩Updates - ✅ Update the **RealESRGAN AnimeVideo-v3** model. Please see [anime video models](docs/anime_video_model.md) and [comparisons](docs/anime_comparisons.md) for more details. - ✅ Add small models for anime videos. More details are in [anime video models](docs/anime_video_model.md). - ✅ Add the ncnn implementation [Real-ESRGAN-ncnn-vulkan](https://github.com/xinntao/Real-ESRGAN-ncnn-vulkan). - ✅ Add [*RealESRGAN_x4plus_anime_6B.pth*](https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.2.4/RealESRGAN_x4plus_anime_6B.pth), which is optimized for **anime** images with much smaller model size. More details and comparisons with [waifu2x](https://github.com/nihui/waifu2x-ncnn-vulkan) are in [**anime_model.md**](docs/anime_model.md) - ✅ Support finetuning on your own data or paired data (*i.e.*, finetuning ESRGAN). See [here](Training.md#Finetune-Real-ESRGAN-on-your-own-dataset) - ✅ Integrate [GFPGAN](https://github.com/TencentARC/GFPGAN) to support **face enhancement**. - ✅ Integrated to [Huggingface Spaces](https://huggingface.co/spaces) with [Gradio](https://github.com/gradio-app/gradio). See [Gradio Web Demo](https://huggingface.co/spaces/akhaliq/Real-ESRGAN). Thanks [@AK391](https://github.com/AK391) - ✅ Support arbitrary scale with `--outscale` (It actually further resizes outputs with `LANCZOS4`). Add *RealESRGAN_x2plus.pth* model. - ✅ [The inference code](inference_realesrgan.py) supports: 1) **tile** options; 2) images with **alpha channel**; 3) **gray** images; 4) **16-bit** images. - ✅ The training codes have been released. A detailed guide can be found in [Training.md](Training.md).
🧩Projects that use Real-ESRGAN     👋 If you develop/use Real-ESRGAN in your projects, welcome to let me know. - NCNN-Android: [RealSR-NCNN-Android](https://github.com/tumuyan/RealSR-NCNN-Android) by [tumuyan](https://github.com/tumuyan) - VapourSynth: [vs-realesrgan](https://github.com/HolyWu/vs-realesrgan) by [HolyWu](https://github.com/HolyWu) - NCNN: [Real-ESRGAN-ncnn-vulkan](https://github.com/xinntao/Real-ESRGAN-ncnn-vulkan)     **GUI** - [Waifu2x-Extension-GUI](https://github.com/AaronFeng753/Waifu2x-Extension-GUI) by [AaronFeng753](https://github.com/AaronFeng753) - [Squirrel-RIFE](https://github.com/Justin62628/Squirrel-RIFE) by [Justin62628](https://github.com/Justin62628) - [Real-GUI](https://github.com/scifx/Real-GUI) by [scifx](https://github.com/scifx) - [Real-ESRGAN_GUI](https://github.com/net2cn/Real-ESRGAN_GUI) by [net2cn](https://github.com/net2cn) - [Real-ESRGAN-EGUI](https://github.com/WGzeyu/Real-ESRGAN-EGUI) by [WGzeyu](https://github.com/WGzeyu) - [anime_upscaler](https://github.com/shangar21/anime_upscaler) by [shangar21](https://github.com/shangar21)
👀Demo videos - [大闹天宫片段](https://www.bilibili.com/video/BV1ja41117zb)

:book: Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data

[Paper]   [Project Page]   [YouTube Video]   [B站讲解]   [Poster]   [PPT slides]
Xintao Wang, Liangbin Xie, Chao Dong, Ying Shan
Tencent ARC Lab; Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences


We have provided a pretrained model (RealESRGAN_x4plus.pth) with upsampling X4.
Note that RealESRGAN may still fail in some cases as the real-world degradations are really too complex.
Moreover, it may not perform well on human faces, text, etc, which will be optimized later.

Real-ESRGAN will be a long-term supported project (in my current plan :smiley:). It will be continuously updated in my spare time.

Here is a TODO list in the near future:

  • optimize for human faces
  • optimize for texts
  • optimize for anime images
  • support more scales
  • support controllable restoration strength

If you have any good ideas or demands, please open an issue/discussion to let me know.
If you have some images that Real-ESRGAN could not well restored, please also open an issue/discussion. I will record it (but I cannot guarantee to resolve it:stuck_out_tongue:). If necessary, I will open a page to specially record these real-world cases that need to be solved, but the current technology is difficult to handle well.


Portable executable files

You can download Windows / Linux / MacOS executable files for Intel/AMD/Nvidia GPU.

This executable file is portable and includes all the binaries and models required. No CUDA or PyTorch environment is needed.

You can simply run the following command (the Windows example, more information is in the README.md of each executable files):

./realesrgan-ncnn-vulkan.exe -i input.jpg -o output.png -n model_name

We have provided five models:

  1. realesrgan-x4plus (default)
  2. realesrnet-x4plus
  3. realesrgan-x4plus-anime (optimized for anime images, small model size)
  4. realesr-animevideov3 (animation video)

You can use the -n argument for other models, for example, ./realesrgan-ncnn-vulkan.exe -i input.jpg -o output.png -n realesrnet-x4plus

Usage of executable files

  1. Please refer to Real-ESRGAN-ncnn-vulkan for more details.
  2. Note that it does not support all the functions (such as outscale) as the python script inference_realesrgan.py.

    Usage: realesrgan-ncnn-vulkan.exe -i infile -o outfile [options]...
    
    -h                   show this help
    -i input-path        input image path (jpg/png/webp) or directory
    -o output-path       output image path (jpg/png/webp) or directory
    -s scale             upscale ratio (can be 2, 3, 4. default=4)
    -t tile-size         tile size (>=32/0=auto, default=0) can be 0,0,0 for multi-gpu
    -m model-path        folder path to the pre-trained models. default=models
    -n model-name        model name (default=realesr-animevideov3, can be realesr-animevideov3 | realesrgan-x4plus | realesrgan-x4plus-anime | realesrnet-x4plus)
    -g gpu-id            gpu device to use (default=auto) can be 0,1,2 for multi-gpu
    -j load:proc:save    thread count for load/proc/save (default=1:2:2) can be 1:2,2,2:2 for multi-gpu
    -x                   enable tta mode"
    -f format            output image format (jpg/png/webp, default=ext/png)
    -v                   verbose output
    

Note that it may introduce block inconsistency (and also generate slightly different results from the PyTorch implementation), because this executable file first crops the input image into several tiles, and then processes them separately, finally stitches together.


:wrench: Dependencies and Installation

Installation

  1. Clone repo

    git clone https://github.com/xinntao/Real-ESRGAN.git
    cd Real-ESRGAN
    
  2. Install dependent packages

    # Install basicsr - https://github.com/xinntao/BasicSR
    # We use BasicSR for both training and inference
    pip install basicsr
    # facexlib and gfpgan are for face enhancement
    pip install facexlib
    pip install gfpgan
    pip install -r requirements.txt
    python setup.py develop
    

:zap: Quick Inference

Inference general images

Download pre-trained models: RealESRGAN_x4plus.pth

wget https://github.com/xinntao/Real-ESRGAN/releases/download/v0.1.0/RealESRGAN_x4plus.pth -P experiments/pretrained_models

Inference!

python inference_realesrgan.py -n RealESRGAN_x4plus -i inputs --face_enhance

Results are in the results folder

Inference anime images

Pre-trained models: RealESRGAN_x4plus_anime_6B
More details and comparisons with waifu2x are in anime_model.md

# download model
wget https://github.com/xinntao/Real-ESRGAN/releases/download/v0.2.2.4/RealESRGAN_x4plus_anime_6B.pth -P experiments/pretrained_models
# inference
python inference_realesrgan.py -n RealESRGAN_x4plus_anime_6B -i inputs

Results are in the results folder

Usage of python script

  1. You can use X4 model for arbitrary output size with the argument outscale. The program will further perform cheap resize operation after the Real-ESRGAN output.

    Usage: python inference_realesrgan.py -n RealESRGAN_x4plus -i infile -o outfile [options]...
    
    A common command: python inference_realesrgan.py -n RealESRGAN_x4plus -i infile --outscale 3.5 --face_enhance
    
    -h                   show this help
    -i --input           Input image or folder. Default: inputs
    -o --output          Output folder. Default: results
    -n --model_name      Model name. Default: RealESRGAN_x4plus
    -s, --outscale       The final upsampling scale of the image. Default: 4
    --suffix             Suffix of the restored image. Default: out
    -t, --tile           Tile size, 0 for no tile during testing. Default: 0
    --face_enhance       Whether to use GFPGAN to enhance face. Default: False
    --fp32               Use fp32 precision during inference. Default: fp16 (half precision).
    --ext                Image extension. Options: auto | jpg | png, auto means using the same extension as inputs. Default: auto
    

:european_castle: Model Zoo

Please see docs/model_zoo.md

:computer: Training and Finetuning on your own dataset

A detailed guide can be found in Training.md.

BibTeX

@InProceedings{wang2021realesrgan,
    author    = {Xintao Wang and Liangbin Xie and Chao Dong and Ying Shan},
    title     = {Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data},
    booktitle = {International Conference on Computer Vision Workshops (ICCVW)},
    date      = {2021}
}

:e-mail: Contact

If you have any question, please email xintao.wang@outlook.com or xintaowang@tencent.com.

:hugs: Acknowledgement

Thanks for all the contributors.