Voix commited on
Commit
0bdbb84
1 Parent(s): 55b2ffd

Upload 9 files

Browse files
Files changed (9) hide show
  1. .gitattributes +3 -35
  2. .gitignore +12 -0
  3. LICENSE +201 -0
  4. QA.md +34 -0
  5. README.md +202 -0
  6. RRDBNet_arch.py +78 -0
  7. net_interp.py +21 -0
  8. test.py +37 -0
  9. transer_RRDB_models.py +55 -0
.gitattributes CHANGED
@@ -1,35 +1,3 @@
1
- *.7z filter=lfs diff=lfs merge=lfs -text
2
- *.arrow filter=lfs diff=lfs merge=lfs -text
3
- *.bin filter=lfs diff=lfs merge=lfs -text
4
- *.bz2 filter=lfs diff=lfs merge=lfs -text
5
- *.ckpt filter=lfs diff=lfs merge=lfs -text
6
- *.ftz filter=lfs diff=lfs merge=lfs -text
7
- *.gz filter=lfs diff=lfs merge=lfs -text
8
- *.h5 filter=lfs diff=lfs merge=lfs -text
9
- *.joblib filter=lfs diff=lfs merge=lfs -text
10
- *.lfs.* filter=lfs diff=lfs merge=lfs -text
11
- *.mlmodel filter=lfs diff=lfs merge=lfs -text
12
- *.model filter=lfs diff=lfs merge=lfs -text
13
- *.msgpack filter=lfs diff=lfs merge=lfs -text
14
- *.npy filter=lfs diff=lfs merge=lfs -text
15
- *.npz filter=lfs diff=lfs merge=lfs -text
16
- *.onnx filter=lfs diff=lfs merge=lfs -text
17
- *.ot filter=lfs diff=lfs merge=lfs -text
18
- *.parquet filter=lfs diff=lfs merge=lfs -text
19
- *.pb filter=lfs diff=lfs merge=lfs -text
20
- *.pickle filter=lfs diff=lfs merge=lfs -text
21
- *.pkl filter=lfs diff=lfs merge=lfs -text
22
- *.pt filter=lfs diff=lfs merge=lfs -text
23
- *.pth filter=lfs diff=lfs merge=lfs -text
24
- *.rar filter=lfs diff=lfs merge=lfs -text
25
- *.safetensors filter=lfs diff=lfs merge=lfs -text
26
- saved_model/**/* filter=lfs diff=lfs merge=lfs -text
27
- *.tar.* filter=lfs diff=lfs merge=lfs -text
28
- *.tar filter=lfs diff=lfs merge=lfs -text
29
- *.tflite filter=lfs diff=lfs merge=lfs -text
30
- *.tgz filter=lfs diff=lfs merge=lfs -text
31
- *.wasm filter=lfs diff=lfs merge=lfs -text
32
- *.xz filter=lfs diff=lfs merge=lfs -text
33
- *.zip filter=lfs diff=lfs merge=lfs -text
34
- *.zst filter=lfs diff=lfs merge=lfs -text
35
- *tfevents* filter=lfs diff=lfs merge=lfs -text
 
1
+ *.gif filter=lfs diff=lfs merge=lfs -text
2
+ *.png filter=lfs diff=lfs merge=lfs -text
3
+ *.mp4 filter=lfs diff=lfs merge=lfs -text
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
.gitignore ADDED
@@ -0,0 +1,12 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # folder
2
+ .vscode
3
+
4
+ # file type
5
+ *.svg
6
+ *.pyc
7
+ *.pth
8
+ *.t7
9
+ *.caffemodel
10
+ *.mat
11
+ *.npy
12
+
LICENSE ADDED
@@ -0,0 +1,201 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ Apache License
2
+ Version 2.0, January 2004
3
+ http://www.apache.org/licenses/
4
+
5
+ TERMS AND CONDITIONS FOR USE, REPRODUCTION, AND DISTRIBUTION
6
+
7
+ 1. Definitions.
8
+
9
+ "License" shall mean the terms and conditions for use, reproduction,
10
+ and distribution as defined by Sections 1 through 9 of this document.
11
+
12
+ "Licensor" shall mean the copyright owner or entity authorized by
13
+ the copyright owner that is granting the License.
14
+
15
+ "Legal Entity" shall mean the union of the acting entity and all
16
+ other entities that control, are controlled by, or are under common
17
+ control with that entity. For the purposes of this definition,
18
+ "control" means (i) the power, direct or indirect, to cause the
19
+ direction or management of such entity, whether by contract or
20
+ otherwise, or (ii) ownership of fifty percent (50%) or more of the
21
+ outstanding shares, or (iii) beneficial ownership of such entity.
22
+
23
+ "You" (or "Your") shall mean an individual or Legal Entity
24
+ exercising permissions granted by this License.
25
+
26
+ "Source" form shall mean the preferred form for making modifications,
27
+ including but not limited to software source code, documentation
28
+ source, and configuration files.
29
+
30
+ "Object" form shall mean any form resulting from mechanical
31
+ transformation or translation of a Source form, including but
32
+ not limited to compiled object code, generated documentation,
33
+ and conversions to other media types.
34
+
35
+ "Work" shall mean the work of authorship, whether in Source or
36
+ Object form, made available under the License, as indicated by a
37
+ copyright notice that is included in or attached to the work
38
+ (an example is provided in the Appendix below).
39
+
40
+ "Derivative Works" shall mean any work, whether in Source or Object
41
+ form, that is based on (or derived from) the Work and for which the
42
+ editorial revisions, annotations, elaborations, or other modifications
43
+ represent, as a whole, an original work of authorship. For the purposes
44
+ of this License, Derivative Works shall not include works that remain
45
+ separable from, or merely link (or bind by name) to the interfaces of,
46
+ the Work and Derivative Works thereof.
47
+
48
+ "Contribution" shall mean any work of authorship, including
49
+ the original version of the Work and any modifications or additions
50
+ to that Work or Derivative Works thereof, that is intentionally
51
+ submitted to Licensor for inclusion in the Work by the copyright owner
52
+ or by an individual or Legal Entity authorized to submit on behalf of
53
+ the copyright owner. For the purposes of this definition, "submitted"
54
+ means any form of electronic, verbal, or written communication sent
55
+ to the Licensor or its representatives, including but not limited to
56
+ communication on electronic mailing lists, source code control systems,
57
+ and issue tracking systems that are managed by, or on behalf of, the
58
+ Licensor for the purpose of discussing and improving the Work, but
59
+ excluding communication that is conspicuously marked or otherwise
60
+ designated in writing by the copyright owner as "Not a Contribution."
61
+
62
+ "Contributor" shall mean Licensor and any individual or Legal Entity
63
+ on behalf of whom a Contribution has been received by Licensor and
64
+ subsequently incorporated within the Work.
65
+
66
+ 2. Grant of Copyright License. Subject to the terms and conditions of
67
+ this License, each Contributor hereby grants to You a perpetual,
68
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
69
+ copyright license to reproduce, prepare Derivative Works of,
70
+ publicly display, publicly perform, sublicense, and distribute the
71
+ Work and such Derivative Works in Source or Object form.
72
+
73
+ 3. Grant of Patent License. Subject to the terms and conditions of
74
+ this License, each Contributor hereby grants to You a perpetual,
75
+ worldwide, non-exclusive, no-charge, royalty-free, irrevocable
76
+ (except as stated in this section) patent license to make, have made,
77
+ use, offer to sell, sell, import, and otherwise transfer the Work,
78
+ where such license applies only to those patent claims licensable
79
+ by such Contributor that are necessarily infringed by their
80
+ Contribution(s) alone or by combination of their Contribution(s)
81
+ with the Work to which such Contribution(s) was submitted. If You
82
+ institute patent litigation against any entity (including a
83
+ cross-claim or counterclaim in a lawsuit) alleging that the Work
84
+ or a Contribution incorporated within the Work constitutes direct
85
+ or contributory patent infringement, then any patent licenses
86
+ granted to You under this License for that Work shall terminate
87
+ as of the date such litigation is filed.
88
+
89
+ 4. Redistribution. You may reproduce and distribute copies of the
90
+ Work or Derivative Works thereof in any medium, with or without
91
+ modifications, and in Source or Object form, provided that You
92
+ meet the following conditions:
93
+
94
+ (a) You must give any other recipients of the Work or
95
+ Derivative Works a copy of this License; and
96
+
97
+ (b) You must cause any modified files to carry prominent notices
98
+ stating that You changed the files; and
99
+
100
+ (c) You must retain, in the Source form of any Derivative Works
101
+ that You distribute, all copyright, patent, trademark, and
102
+ attribution notices from the Source form of the Work,
103
+ excluding those notices that do not pertain to any part of
104
+ the Derivative Works; and
105
+
106
+ (d) If the Work includes a "NOTICE" text file as part of its
107
+ distribution, then any Derivative Works that You distribute must
108
+ include a readable copy of the attribution notices contained
109
+ within such NOTICE file, excluding those notices that do not
110
+ pertain to any part of the Derivative Works, in at least one
111
+ of the following places: within a NOTICE text file distributed
112
+ as part of the Derivative Works; within the Source form or
113
+ documentation, if provided along with the Derivative Works; or,
114
+ within a display generated by the Derivative Works, if and
115
+ wherever such third-party notices normally appear. The contents
116
+ of the NOTICE file are for informational purposes only and
117
+ do not modify the License. You may add Your own attribution
118
+ notices within Derivative Works that You distribute, alongside
119
+ or as an addendum to the NOTICE text from the Work, provided
120
+ that such additional attribution notices cannot be construed
121
+ as modifying the License.
122
+
123
+ You may add Your own copyright statement to Your modifications and
124
+ may provide additional or different license terms and conditions
125
+ for use, reproduction, or distribution of Your modifications, or
126
+ for any such Derivative Works as a whole, provided Your use,
127
+ reproduction, and distribution of the Work otherwise complies with
128
+ the conditions stated in this License.
129
+
130
+ 5. Submission of Contributions. Unless You explicitly state otherwise,
131
+ any Contribution intentionally submitted for inclusion in the Work
132
+ by You to the Licensor shall be under the terms and conditions of
133
+ this License, without any additional terms or conditions.
134
+ Notwithstanding the above, nothing herein shall supersede or modify
135
+ the terms of any separate license agreement you may have executed
136
+ with Licensor regarding such Contributions.
137
+
138
+ 6. Trademarks. This License does not grant permission to use the trade
139
+ names, trademarks, service marks, or product names of the Licensor,
140
+ except as required for reasonable and customary use in describing the
141
+ origin of the Work and reproducing the content of the NOTICE file.
142
+
143
+ 7. Disclaimer of Warranty. Unless required by applicable law or
144
+ agreed to in writing, Licensor provides the Work (and each
145
+ Contributor provides its Contributions) on an "AS IS" BASIS,
146
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or
147
+ implied, including, without limitation, any warranties or conditions
148
+ of TITLE, NON-INFRINGEMENT, MERCHANTABILITY, or FITNESS FOR A
149
+ PARTICULAR PURPOSE. You are solely responsible for determining the
150
+ appropriateness of using or redistributing the Work and assume any
151
+ risks associated with Your exercise of permissions under this License.
152
+
153
+ 8. Limitation of Liability. In no event and under no legal theory,
154
+ whether in tort (including negligence), contract, or otherwise,
155
+ unless required by applicable law (such as deliberate and grossly
156
+ negligent acts) or agreed to in writing, shall any Contributor be
157
+ liable to You for damages, including any direct, indirect, special,
158
+ incidental, or consequential damages of any character arising as a
159
+ result of this License or out of the use or inability to use the
160
+ Work (including but not limited to damages for loss of goodwill,
161
+ work stoppage, computer failure or malfunction, or any and all
162
+ other commercial damages or losses), even if such Contributor
163
+ has been advised of the possibility of such damages.
164
+
165
+ 9. Accepting Warranty or Additional Liability. While redistributing
166
+ the Work or Derivative Works thereof, You may choose to offer,
167
+ and charge a fee for, acceptance of support, warranty, indemnity,
168
+ or other liability obligations and/or rights consistent with this
169
+ License. However, in accepting such obligations, You may act only
170
+ on Your own behalf and on Your sole responsibility, not on behalf
171
+ of any other Contributor, and only if You agree to indemnify,
172
+ defend, and hold each Contributor harmless for any liability
173
+ incurred by, or claims asserted against, such Contributor by reason
174
+ of your accepting any such warranty or additional liability.
175
+
176
+ END OF TERMS AND CONDITIONS
177
+
178
+ APPENDIX: How to apply the Apache License to your work.
179
+
180
+ To apply the Apache License to your work, attach the following
181
+ boilerplate notice, with the fields enclosed by brackets "[]"
182
+ replaced with your own identifying information. (Don't include
183
+ the brackets!) The text should be enclosed in the appropriate
184
+ comment syntax for the file format. We also recommend that a
185
+ file or class name and description of purpose be included on the
186
+ same "printed page" as the copyright notice for easier
187
+ identification within third-party archives.
188
+
189
+ Copyright [yyyy] [name of copyright owner]
190
+
191
+ Licensed under the Apache License, Version 2.0 (the "License");
192
+ you may not use this file except in compliance with the License.
193
+ You may obtain a copy of the License at
194
+
195
+ http://www.apache.org/licenses/LICENSE-2.0
196
+
197
+ Unless required by applicable law or agreed to in writing, software
198
+ distributed under the License is distributed on an "AS IS" BASIS,
199
+ WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
200
+ See the License for the specific language governing permissions and
201
+ limitations under the License.
QA.md ADDED
@@ -0,0 +1,34 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ # Frequently Asked Questions
2
+
3
+ ### 1. How to reproduce your results in the [PIRM18-SR Challenge](https://www.pirm2018.org/PIRM-SR.html) (with low perceptual index)?
4
+
5
+ First, the released ESRGAN model in the GitHub (`RRDB_ESRGAN_x4.pth`) is **different** from the model we submitted in the competition.
6
+ We found that the lower perceptual index does not always guarantee a better visual quality.
7
+ The aims for the competition and our ESRGAN work will be a bit different.
8
+ We think the aim for the competition is the lower perceptual index and the aim for our ESRGAN work is the better visual quality.
9
+ > More analyses can be found in Sec 4.1 and Sec 5 in [PIRM18-SR Chanllenge report](https://arxiv.org/pdf/1809.07517.pdf).
10
+ > It points out that PI (perceptual index) is well correlated with the human-opinion-scores on a coarse scale, but it is not always well-correlated with these scores on a finer scale. This highlights the urgent need for better perceptual quality metrics.)
11
+
12
+ Therefore, in the PIRM18-SR Challenge competition, we used several tricks for the best perceptual index (see Section 4.5 in the [paper](https://arxiv.org/abs/1809.00219)).
13
+
14
+ Here, we provid the models and codes used in the competition, which is able to produce the results on the `PIRM test dataset` (we use MATLAB 2016b/2017a):
15
+
16
+ | Group | Perceptual index | RMSE |
17
+ | ------------- |:-------------:| -----:|
18
+ | SuperSR | 1.978 | 15.30 |
19
+
20
+ > 1. Download the model and codes from [GoogleDrive](https://drive.google.com/file/d/1l0gBRMqhVLpL_-7R7aN-q-3hnv5ADFSM/view?usp=sharing)
21
+ > 2. Put LR input images in the `LR` folder
22
+ > 3. Run `python test.py`
23
+ > 4. Run `main_reverse_filter.m` in MATLAB as a post processing
24
+ > 5. The results on my computer are: Perceptual index: **1.9777** and RMSE: **15.304**
25
+
26
+
27
+ ### 2. How do you get the perceptual index in your ESRGAN paper?
28
+ In our paper, we provide the perceptual index in two places.
29
+
30
+ 1). In the Fig. 2, the perceptual index on PIRM self validation dataset is obtained with the **model we submitted in the competition**.
31
+ Since the pupose of this figure is to show the perception-distortion plane. And we also use the post-precessing here same as in the competition.
32
+
33
+ 2). In the Fig.7, the perceptual indexs are provided as references and they are tested on the data generated by the released ESRGAN model `RRDB_ESRGAN_x4.pth` in the GiuHub.
34
+ Also, there is **no** post-processing when testing the ESRGAN model for better visual quality.
README.md ADDED
@@ -0,0 +1,202 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ## ESRGAN (Enhanced SRGAN) [:rocket: [BasicSR](https://github.com/xinntao/BasicSR)] [[Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN)]
2
+
3
+ :sparkles: **New Updates.**
4
+
5
+ We have extended ESRGAN to [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN), which is a **more practical algorithm for real-world image restoration**. For example, it can also remove annoying JPEG compression artifacts. <br> You are recommended to have a try :smiley:
6
+
7
+ In the [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo,
8
+
9
+ - You can still use the original ESRGAN model or your re-trained ESRGAN model. [The model zoo in Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN#european_castle-model-zoo).
10
+ - We provide a more handy inference script, which supports 1) **tile** inference; 2) images with **alpha channel**; 3) **gray** images; 4) **16-bit** images.
11
+ - We also provide a **Windows executable file** `RealESRGAN-ncnn-vulkan` for easier use without installing the environment. This executable file also includes the original ESRGAN model.
12
+ - The full training codes are also released in the [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo.
13
+
14
+ Welcome to open issues or open discussions in the [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo.
15
+
16
+ - If you have any question, you can open an issue in the [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo.
17
+ - If you have any good ideas or demands, please open an issue/discussion in the [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo to let me know.
18
+ - If you have some images that Real-ESRGAN could not well restored, please also open an issue/discussion in the [Real-ESRGAN](https://github.com/xinntao/Real-ESRGAN) repo. I will record it (but I cannot guarantee to resolve it😛).
19
+
20
+ Here are some examples for Real-ESRGAN:
21
+
22
+ <p align="center">
23
+ <img src="https://raw.githubusercontent.com/xinntao/Real-ESRGAN/master/assets/teaser.jpg">
24
+ </p>
25
+ :book: Real-ESRGAN: Training Real-World Blind Super-Resolution with Pure Synthetic Data
26
+
27
+ > [[Paper](https://arxiv.org/abs/2107.10833)] <br>
28
+ > [Xintao Wang](https://xinntao.github.io/), Liangbin Xie, [Chao Dong](https://scholar.google.com.hk/citations?user=OSDCB0UAAAAJ), [Ying Shan](https://scholar.google.com/citations?user=4oXBp9UAAAAJ&hl=en) <br>
29
+ > Applied Research Center (ARC), Tencent PCG<br>
30
+ > Shenzhen Institutes of Advanced Technology, Chinese Academy of Sciences
31
+
32
+ -----
33
+
34
+ As there may be some repos have dependency on this ESRGAN repo, we will not modify this ESRGAN repo (especially the codes).
35
+
36
+ The following is the original README:
37
+
38
+ #### The training codes are in :rocket: [BasicSR](https://github.com/xinntao/BasicSR). This repo only provides simple testing codes, pretrained models and the network interpolation demo.
39
+
40
+ [BasicSR](https://github.com/xinntao/BasicSR) is an **open source** image and video super-resolution toolbox based on PyTorch (will extend to more restoration tasks in the future). <br>
41
+ It includes methods such as **EDSR, RCAN, SRResNet, SRGAN, ESRGAN, EDVR**, etc. It now also supports **StyleGAN2**.
42
+
43
+ ### Enhanced Super-Resolution Generative Adversarial Networks
44
+ By Xintao Wang, [Ke Yu](https://yuke93.github.io/), Shixiang Wu, [Jinjin Gu](http://www.jasongt.com/), Yihao Liu, [Chao Dong](https://scholar.google.com.hk/citations?user=OSDCB0UAAAAJ&hl=en), [Yu Qiao](http://mmlab.siat.ac.cn/yuqiao/), [Chen Change Loy](http://personal.ie.cuhk.edu.hk/~ccloy/)
45
+
46
+ We won the first place in [PIRM2018-SR competition](https://www.pirm2018.org/PIRM-SR.html) (region 3) and got the best perceptual index.
47
+ The paper is accepted to [ECCV2018 PIRM Workshop](https://pirm2018.org/).
48
+
49
+ :triangular_flag_on_post: Add [Frequently Asked Questions](https://github.com/xinntao/ESRGAN/blob/master/QA.md).
50
+
51
+ > For instance,
52
+ > 1. How to reproduce your results in the PIRM18-SR Challenge (with low perceptual index)?
53
+ > 2. How do you get the perceptual index in your ESRGAN paper?
54
+
55
+ #### BibTeX
56
+
57
+ @InProceedings{wang2018esrgan,
58
+ author = {Wang, Xintao and Yu, Ke and Wu, Shixiang and Gu, Jinjin and Liu, Yihao and Dong, Chao and Qiao, Yu and Loy, Chen Change},
59
+ title = {ESRGAN: Enhanced super-resolution generative adversarial networks},
60
+ booktitle = {The European Conference on Computer Vision Workshops (ECCVW)},
61
+ month = {September},
62
+ year = {2018}
63
+ }
64
+
65
+ <p align="center">
66
+ <img src="figures/baboon.jpg">
67
+ </p>
68
+
69
+ The **RRDB_PSNR** PSNR_oriented model trained with DF2K dataset (a merged dataset with [DIV2K](https://data.vision.ee.ethz.ch/cvl/DIV2K/) and [Flickr2K](http://cv.snu.ac.kr/research/EDSR/Flickr2K.tar) (proposed in [EDSR](https://github.com/LimBee/NTIRE2017))) is also able to achive high PSNR performance.
70
+
71
+ | <sub>Method</sub> | <sub>Training dataset</sub> | <sub>Set5</sub> | <sub>Set14</sub> | <sub>BSD100</sub> | <sub>Urban100</sub> | <sub>Manga109</sub> |
72
+ |:---:|:---:|:---:|:---:|:---:|:---:|:---:|
73
+ | <sub>[SRCNN](http://mmlab.ie.cuhk.edu.hk/projects/SRCNN.html)</sub>| <sub>291</sub>| <sub>30.48/0.8628</sub> |<sub>27.50/0.7513</sub>|<sub>26.90/0.7101</sub>|<sub>24.52/0.7221</sub>|<sub>27.58/0.8555</sub>|
74
+ | <sub>[EDSR](https://github.com/thstkdgus35/EDSR-PyTorch)</sub> | <sub>DIV2K</sub> | <sub>32.46/0.8968</sub> | <sub>28.80/0.7876</sub> | <sub>27.71/0.7420</sub> | <sub>26.64/0.8033</sub> | <sub>31.02/0.9148</sub> |
75
+ | <sub>[RCAN](https://github.com/yulunzhang/RCAN)</sub> | <sub>DIV2K</sub> | <sub>32.63/0.9002</sub> | <sub>28.87/0.7889</sub> | <sub>27.77/0.7436</sub> | <sub>26.82/ 0.8087</sub>| <sub>31.22/ 0.9173</sub>|
76
+ |<sub>RRDB(ours)</sub>| <sub>DF2K</sub>| <sub>**32.73/0.9011**</sub> |<sub>**28.99/0.7917**</sub> |<sub>**27.85/0.7455**</sub> |<sub>**27.03/0.8153**</sub> |<sub>**31.66/0.9196**</sub>|
77
+
78
+ ## Quick Test
79
+ #### Dependencies
80
+ - Python 3
81
+ - [PyTorch >= 1.0](https://pytorch.org/) (CUDA version >= 7.5 if installing with CUDA. [More details](https://pytorch.org/get-started/previous-versions/))
82
+ - Python packages: `pip install numpy opencv-python`
83
+
84
+ ### Test models
85
+ 1. Clone this github repo.
86
+ ```
87
+ git clone https://github.com/xinntao/ESRGAN
88
+ cd ESRGAN
89
+ ```
90
+ 2. Place your own **low-resolution images** in `./LR` folder. (There are two sample images - baboon and comic).
91
+ 3. Download pretrained models from [Google Drive](https://drive.google.com/drive/u/0/folders/17VYV_SoZZesU6mbxz2dMAIccSSlqLecY) or [Baidu Drive](https://pan.baidu.com/s/1-Lh6ma-wXzfH8NqeBtPaFQ). Place the models in `./models`. We provide two models with high perceptual quality and high PSNR performance (see [model list](https://github.com/xinntao/ESRGAN/tree/master/models)).
92
+ 4. Run test. We provide ESRGAN model and RRDB_PSNR model and you can config in the `test.py`.
93
+ ```
94
+ python test.py
95
+ ```
96
+ 5. The results are in `./results` folder.
97
+ ### Network interpolation demo
98
+ You can interpolate the RRDB_ESRGAN and RRDB_PSNR models with alpha in [0, 1].
99
+
100
+ 1. Run `python net_interp.py 0.8`, where *0.8* is the interpolation parameter and you can change it to any value in [0,1].
101
+ 2. Run `python test.py models/interp_08.pth`, where *models/interp_08.pth* is the model path.
102
+
103
+ <p align="center">
104
+ <img height="400" src="figures/43074.gif">
105
+ </p>
106
+
107
+ ## Perceptual-driven SR Results
108
+
109
+ You can download all the resutls from [Google Drive](https://drive.google.com/drive/folders/1iaM-c6EgT1FNoJAOKmDrK7YhEhtlKcLx?usp=sharing). (:heavy_check_mark: included; :heavy_minus_sign: not included; :o: TODO)
110
+
111
+ HR images can be downloaed from [BasicSR-Datasets](https://github.com/xinntao/BasicSR#datasets).
112
+
113
+ | Datasets |LR | [*ESRGAN*](https://arxiv.org/abs/1809.00219) | [SRGAN](https://arxiv.org/abs/1609.04802) | [EnhanceNet](http://openaccess.thecvf.com/content_ICCV_2017/papers/Sajjadi_EnhanceNet_Single_Image_ICCV_2017_paper.pdf) | [CX](https://arxiv.org/abs/1803.04626) |
114
+ |:---:|:---:|:---:|:---:|:---:|:---:|
115
+ | Set5 |:heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark:| :o: |
116
+ | Set14 | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark:| :o: |
117
+ | BSDS100 | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark: | :heavy_check_mark:| :o: |
118
+ | [PIRM](https://pirm.github.io/) <br><sup>(val, test)</sup> | :heavy_check_mark: | :heavy_check_mark: | :heavy_minus_sign: | :heavy_check_mark:| :heavy_check_mark: |
119
+ | [OST300](https://arxiv.org/pdf/1804.02815.pdf) |:heavy_check_mark: | :heavy_check_mark: | :heavy_minus_sign: | :heavy_check_mark:| :o: |
120
+ | urban100 | :heavy_check_mark: | :heavy_check_mark: | :heavy_minus_sign: | :heavy_check_mark:| :o: |
121
+ | [DIV2K](https://data.vision.ee.ethz.ch/cvl/DIV2K/) <br><sup>(val, test)</sup> | :heavy_check_mark: | :heavy_check_mark: | :heavy_minus_sign: | :heavy_check_mark:| :o: |
122
+
123
+ ## ESRGAN
124
+ We improve the [SRGAN](https://arxiv.org/abs/1609.04802) from three aspects:
125
+ 1. adopt a deeper model using Residual-in-Residual Dense Block (RRDB) without batch normalization layers.
126
+ 2. employ [Relativistic average GAN](https://ajolicoeur.wordpress.com/relativisticgan/) instead of the vanilla GAN.
127
+ 3. improve the perceptual loss by using the features before activation.
128
+
129
+ In contrast to SRGAN, which claimed that **deeper models are increasingly difficult to train**, our deeper ESRGAN model shows its superior performance with easy training.
130
+
131
+ <p align="center">
132
+ <img height="120" src="figures/architecture.jpg">
133
+ </p>
134
+ <p align="center">
135
+ <img height="180" src="figures/RRDB.png">
136
+ </p>
137
+
138
+ ## Network Interpolation
139
+ We propose the **network interpolation strategy** to balance the visual quality and PSNR.
140
+
141
+ <p align="center">
142
+ <img height="500" src="figures/net_interp.jpg">
143
+ </p>
144
+
145
+ We show the smooth animation with the interpolation parameters changing from 0 to 1.
146
+ Interestingly, it is observed that the network interpolation strategy provides a smooth control of the RRDB_PSNR model and the fine-tuned ESRGAN model.
147
+
148
+ <p align="center">
149
+ <img height="480" src="figures/81.gif">
150
+ &nbsp &nbsp
151
+ <img height="480" src="figures/102061.gif">
152
+ </p>
153
+
154
+ ## Qualitative Results
155
+ PSNR (evaluated on the Y channel) and the perceptual index used in the PIRM-SR challenge are also provided for reference.
156
+
157
+ <p align="center">
158
+ <img src="figures/qualitative_cmp_01.jpg">
159
+ </p>
160
+ <p align="center">
161
+ <img src="figures/qualitative_cmp_02.jpg">
162
+ </p>
163
+ <p align="center">
164
+ <img src="figures/qualitative_cmp_03.jpg">
165
+ </p>
166
+ <p align="center">
167
+ <img src="figures/qualitative_cmp_04.jpg">
168
+ </p>
169
+
170
+ ## Ablation Study
171
+ Overall visual comparisons for showing the effects of each component in
172
+ ESRGAN. Each column represents a model with its configurations in the top.
173
+ The red sign indicates the main improvement compared with the previous model.
174
+ <p align="center">
175
+ <img src="figures/abalation_study.png">
176
+ </p>
177
+
178
+ ## BN artifacts
179
+ We empirically observe that BN layers tend to bring artifacts. These artifacts,
180
+ namely BN artifacts, occasionally appear among iterations and different settings,
181
+ violating the needs for a stable performance over training. We find that
182
+ the network depth, BN position, training dataset and training loss
183
+ have impact on the occurrence of BN artifacts.
184
+ <p align="center">
185
+ <img src="figures/BN_artifacts.jpg">
186
+ </p>
187
+
188
+ ## Useful techniques to train a very deep network
189
+ We find that residual scaling and smaller initialization can help to train a very deep network. More details are in the Supplementary File attached in our [paper](https://arxiv.org/abs/1809.00219).
190
+
191
+ <p align="center">
192
+ <img height="250" src="figures/train_deeper_neta.png">
193
+ <img height="250" src="figures/train_deeper_netb.png">
194
+ </p>
195
+
196
+ ## The influence of training patch size
197
+ We observe that training a deeper network benefits from a larger patch size. Moreover, the deeper model achieves more improvement (∼0.12dB) than the shallower one (∼0.04dB) since larger model capacity is capable of taking full advantage of
198
+ larger training patch size. (Evaluated on Set5 dataset with RGB channels.)
199
+ <p align="center">
200
+ <img height="250" src="figures/patch_a.png">
201
+ <img height="250" src="figures/patch_b.png">
202
+ </p>
RRDBNet_arch.py ADDED
@@ -0,0 +1,78 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import functools
2
+ import torch
3
+ import torch.nn as nn
4
+ import torch.nn.functional as F
5
+
6
+
7
+ def make_layer(block, n_layers):
8
+ layers = []
9
+ for _ in range(n_layers):
10
+ layers.append(block())
11
+ return nn.Sequential(*layers)
12
+
13
+
14
+ class ResidualDenseBlock_5C(nn.Module):
15
+ def __init__(self, nf=64, gc=32, bias=True):
16
+ super(ResidualDenseBlock_5C, self).__init__()
17
+ # gc: growth channel, i.e. intermediate channels
18
+ self.conv1 = nn.Conv2d(nf, gc, 3, 1, 1, bias=bias)
19
+ self.conv2 = nn.Conv2d(nf + gc, gc, 3, 1, 1, bias=bias)
20
+ self.conv3 = nn.Conv2d(nf + 2 * gc, gc, 3, 1, 1, bias=bias)
21
+ self.conv4 = nn.Conv2d(nf + 3 * gc, gc, 3, 1, 1, bias=bias)
22
+ self.conv5 = nn.Conv2d(nf + 4 * gc, nf, 3, 1, 1, bias=bias)
23
+ self.lrelu = nn.LeakyReLU(negative_slope=0.2, inplace=True)
24
+
25
+ # initialization
26
+ # mutil.initialize_weights([self.conv1, self.conv2, self.conv3, self.conv4, self.conv5], 0.1)
27
+
28
+ def forward(self, x):
29
+ x1 = self.lrelu(self.conv1(x))
30
+ x2 = self.lrelu(self.conv2(torch.cat((x, x1), 1)))
31
+ x3 = self.lrelu(self.conv3(torch.cat((x, x1, x2), 1)))
32
+ x4 = self.lrelu(self.conv4(torch.cat((x, x1, x2, x3), 1)))
33
+ x5 = self.conv5(torch.cat((x, x1, x2, x3, x4), 1))
34
+ return x5 * 0.2 + x
35
+
36
+
37
+ class RRDB(nn.Module):
38
+ '''Residual in Residual Dense Block'''
39
+
40
+ def __init__(self, nf, gc=32):
41
+ super(RRDB, self).__init__()
42
+ self.RDB1 = ResidualDenseBlock_5C(nf, gc)
43
+ self.RDB2 = ResidualDenseBlock_5C(nf, gc)
44
+ self.RDB3 = ResidualDenseBlock_5C(nf, gc)
45
+
46
+ def forward(self, x):
47
+ out = self.RDB1(x)
48
+ out = self.RDB2(out)
49
+ out = self.RDB3(out)
50
+ return out * 0.2 + x
51
+
52
+
53
+ class RRDBNet(nn.Module):
54
+ def __init__(self, in_nc, out_nc, nf, nb, gc=32):
55
+ super(RRDBNet, self).__init__()
56
+ RRDB_block_f = functools.partial(RRDB, nf=nf, gc=gc)
57
+
58
+ self.conv_first = nn.Conv2d(in_nc, nf, 3, 1, 1, bias=True)
59
+ self.RRDB_trunk = make_layer(RRDB_block_f, nb)
60
+ self.trunk_conv = nn.Conv2d(nf, nf, 3, 1, 1, bias=True)
61
+ #### upsampling
62
+ self.upconv1 = nn.Conv2d(nf, nf, 3, 1, 1, bias=True)
63
+ self.upconv2 = nn.Conv2d(nf, nf, 3, 1, 1, bias=True)
64
+ self.HRconv = nn.Conv2d(nf, nf, 3, 1, 1, bias=True)
65
+ self.conv_last = nn.Conv2d(nf, out_nc, 3, 1, 1, bias=True)
66
+
67
+ self.lrelu = nn.LeakyReLU(negative_slope=0.2, inplace=True)
68
+
69
+ def forward(self, x):
70
+ fea = self.conv_first(x)
71
+ trunk = self.trunk_conv(self.RRDB_trunk(fea))
72
+ fea = fea + trunk
73
+
74
+ fea = self.lrelu(self.upconv1(F.interpolate(fea, scale_factor=2, mode='nearest')))
75
+ fea = self.lrelu(self.upconv2(F.interpolate(fea, scale_factor=2, mode='nearest')))
76
+ out = self.conv_last(self.lrelu(self.HRconv(fea)))
77
+
78
+ return out
net_interp.py ADDED
@@ -0,0 +1,21 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import sys
2
+ import torch
3
+ from collections import OrderedDict
4
+
5
+ alpha = float(sys.argv[1])
6
+
7
+ net_PSNR_path = './models/RRDB_PSNR_x4.pth'
8
+ net_ESRGAN_path = './models/RRDB_ESRGAN_x4.pth'
9
+ net_interp_path = './models/interp_{:02d}.pth'.format(int(alpha*10))
10
+
11
+ net_PSNR = torch.load(net_PSNR_path)
12
+ net_ESRGAN = torch.load(net_ESRGAN_path)
13
+ net_interp = OrderedDict()
14
+
15
+ print('Interpolating with alpha = ', alpha)
16
+
17
+ for k, v_PSNR in net_PSNR.items():
18
+ v_ESRGAN = net_ESRGAN[k]
19
+ net_interp[k] = (1 - alpha) * v_PSNR + alpha * v_ESRGAN
20
+
21
+ torch.save(net_interp, net_interp_path)
test.py ADDED
@@ -0,0 +1,37 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os.path as osp
2
+ import glob
3
+ import cv2
4
+ import numpy as np
5
+ import torch
6
+ import RRDBNet_arch as arch
7
+
8
+ model_path = 'models/RRDB_ESRGAN_x4.pth' # models/RRDB_ESRGAN_x4.pth OR models/RRDB_PSNR_x4.pth
9
+ device = torch.device('cpu') # if you want to run on CPU, change 'cuda' -> cpu
10
+ # device = torch.device('cpu')
11
+
12
+ test_img_folder = 'LR/*'
13
+
14
+ model = arch.RRDBNet(3, 3, 64, 23, gc=32)
15
+ model.load_state_dict(torch.load(model_path), strict=True)
16
+ model.eval()
17
+ model = model.to(device)
18
+
19
+ print('Model path {:s}. \nTesting...'.format(model_path))
20
+
21
+ idx = 0
22
+ for path in glob.glob(test_img_folder):
23
+ idx += 1
24
+ base = osp.splitext(osp.basename(path))[0]
25
+ print(idx, base)
26
+ # read images
27
+ img = cv2.imread(path, cv2.IMREAD_COLOR)
28
+ img = img * 1.0 / 255
29
+ img = torch.from_numpy(np.transpose(img[:, :, [2, 1, 0]], (2, 0, 1))).float()
30
+ img_LR = img.unsqueeze(0)
31
+ img_LR = img_LR.to(device)
32
+
33
+ with torch.no_grad():
34
+ output = model(img_LR).data.squeeze().float().cpu().clamp_(0, 1).numpy()
35
+ output = np.transpose(output[[2, 1, 0], :, :], (1, 2, 0))
36
+ output = (output * 255.0).round()
37
+ cv2.imwrite('results/{:s}_rlt.png'.format(base), output)
transer_RRDB_models.py ADDED
@@ -0,0 +1,55 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ import os
2
+ import torch
3
+ import RRDBNet_arch as arch
4
+
5
+ pretrained_net = torch.load('./models/RRDB_ESRGAN_x4.pth')
6
+ save_path = './models/RRDB_ESRGAN_x4.pth'
7
+
8
+ crt_model = arch.RRDBNet(3, 3, 64, 23, gc=32)
9
+ crt_net = crt_model.state_dict()
10
+
11
+ load_net_clean = {}
12
+ for k, v in pretrained_net.items():
13
+ if k.startswith('module.'):
14
+ load_net_clean[k[7:]] = v
15
+ else:
16
+ load_net_clean[k] = v
17
+ pretrained_net = load_net_clean
18
+
19
+ print('###################################\n')
20
+ tbd = []
21
+ for k, v in crt_net.items():
22
+ tbd.append(k)
23
+
24
+ # directly copy
25
+ for k, v in crt_net.items():
26
+ if k in pretrained_net and pretrained_net[k].size() == v.size():
27
+ crt_net[k] = pretrained_net[k]
28
+ tbd.remove(k)
29
+
30
+ crt_net['conv_first.weight'] = pretrained_net['model.0.weight']
31
+ crt_net['conv_first.bias'] = pretrained_net['model.0.bias']
32
+
33
+ for k in tbd.copy():
34
+ if 'RDB' in k:
35
+ ori_k = k.replace('RRDB_trunk.', 'model.1.sub.')
36
+ if '.weight' in k:
37
+ ori_k = ori_k.replace('.weight', '.0.weight')
38
+ elif '.bias' in k:
39
+ ori_k = ori_k.replace('.bias', '.0.bias')
40
+ crt_net[k] = pretrained_net[ori_k]
41
+ tbd.remove(k)
42
+
43
+ crt_net['trunk_conv.weight'] = pretrained_net['model.1.sub.23.weight']
44
+ crt_net['trunk_conv.bias'] = pretrained_net['model.1.sub.23.bias']
45
+ crt_net['upconv1.weight'] = pretrained_net['model.3.weight']
46
+ crt_net['upconv1.bias'] = pretrained_net['model.3.bias']
47
+ crt_net['upconv2.weight'] = pretrained_net['model.6.weight']
48
+ crt_net['upconv2.bias'] = pretrained_net['model.6.bias']
49
+ crt_net['HRconv.weight'] = pretrained_net['model.8.weight']
50
+ crt_net['HRconv.bias'] = pretrained_net['model.8.bias']
51
+ crt_net['conv_last.weight'] = pretrained_net['model.10.weight']
52
+ crt_net['conv_last.bias'] = pretrained_net['model.10.bias']
53
+
54
+ torch.save(crt_net, save_path)
55
+ print('Saving to ', save_path)