Update README.md
Browse files
README.md
CHANGED
@@ -1,3 +1,41 @@
|
|
1 |
-
---
|
2 |
-
license: cc-by-4.0
|
3 |
-
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
---
|
2 |
+
license: cc-by-4.0
|
3 |
+
pipeline_tag: image-to-image
|
4 |
+
tags:
|
5 |
+
- pytorch
|
6 |
+
- super-resolution
|
7 |
+
---
|
8 |
+
|
9 |
+
[Link to Github Release](https://github.com/Phhofm/models/releases/tag/4xLSDIRCompactC)
|
10 |
+
|
11 |
+
# 4xLSDIRCompactC
|
12 |
+
|
13 |
+
Name: 4xLSDIRCompactC
|
14 |
+
Author: Philip Hofmann
|
15 |
+
Release Date: 17.03.2023
|
16 |
+
License: CC BY 4.0
|
17 |
+
Network: SRVGGNetCompact
|
18 |
+
Scale: 4
|
19 |
+
Purpose: 4x photo upscaler that handler jpg compression
|
20 |
+
|
21 |
+
Iterations: 190000
|
22 |
+
batch_size: Variable(1-5)
|
23 |
+
HR_size: 256
|
24 |
+
Dataset: LSDIR
|
25 |
+
Dataset_size: 84991
|
26 |
+
OTF Training No
|
27 |
+
Pretrained_Model_G: 4xLSDIRCompact.pth
|
28 |
+
|
29 |
+
Description: Trying to extend my previous model to be able to handle compression (JPG 100-30) by manually altering the training dataset, since 4xLSDIRCompact cant handle compression. Use this instead of 4xLSDIRCompact if your photo has compression (like an image from the web).
|
30 |
+
|
31 |
+
---
|
32 |
+
|
33 |
+
Here is a comparison to show that 4xLSDIRCompact cannot handle compression artifacts, and that these two models will produce better output for that specific scenario. These models are not ‘better’ than the previous one, they are just meant to handle a different use case: https://imgsli.com/MTYyODY3
|
34 |
+
|
35 |
+
|
36 |
+
![Example1](https://github.com/Phhofm/models/assets/14755670/68be7b9e-472a-4eab-b0ec-a19346f6ac0d)
|
37 |
+
![Example2](https://github.com/Phhofm/models/assets/14755670/b3f59497-82e5-48d1-a15e-842ebfbcbf8a)
|
38 |
+
![Example3](https://github.com/Phhofm/models/assets/14755670/c0ddd288-52fe-4786-841a-264fe5098904)
|
39 |
+
![Example4](https://github.com/Phhofm/models/assets/14755670/292e2c49-5b99-4255-9068-bb1ed33f58cd)
|
40 |
+
![Example5](https://github.com/Phhofm/models/assets/14755670/bba3fb8c-d3f8-438a-9e9c-a3517a88ab5b)
|
41 |
+
|