|
--- |
|
license: openrail++ |
|
tags: |
|
- stable-diffusion |
|
- text-to-image |
|
--- |
|
|
|
# Clean Diffusion 2.0 PoC Model Card |
|
|
|
> That's one small step for artists, one giant leap for engineers. |
|
|
|
![ocean](ocean.png) |
|
|
|
Clean Diffusion 2.0 PoC is Latent Diffusion Model made of public domain images. |
|
Clean Diffusion 2.0 PoC is for the proof of the concept: Stable Diffusion can be made of public domain images. |
|
Therefore, the model can only express the ocean. |
|
If you are Japanese, I recommend Clean Diffusion For Japanese (TBA) instead of Clean Diffusion (For Global). |
|
The model is more powerful than this global version. |
|
|
|
# Note |
|
> With great power comes great responsibility. |
|
|
|
If you **CANNOT UNDERSTAND THESE WORDS**, I recommend that **YOU SHOULD NOT USE ALL OF DIFFUSION MODELS** what have great powers. |
|
|
|
# Getting Started |
|
You would be able to use Clean Diffusion by the following code soon. |
|
|
|
```python |
|
from diffusers import StableDiffusionPipeline |
|
import torch |
|
|
|
model_id = "alfredplpl/clean-diffusion-2-0-poc" |
|
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float32) |
|
pipe = pipe.to("cuda") |
|
|
|
prompt = "Cartoon, ocean." |
|
image = pipe(prompt).images[0] |
|
|
|
image.save("ocean.png") |
|
``` |
|
|
|
# Tuning |
|
Clean Diffusion is less powerful than Stable Diffusion. |
|
Therefore, I recommend to tune Clean Diffusion like Stable Diffusion |
|
because Clean Diffusion of the network architecture and Stable Diffusion of the network architecture are same. |
|
And I repeat the words before I explain the detail. |
|
|
|
> With great power comes great responsibility. |
|
|
|
Please consider the words before you tune Clean Diffusion. |
|
|
|
## Textual Inversion |
|
TBA on Colab. |
|
|
|
## Dreambooth on Stable Diffusion |
|
TBA on Colab. |
|
|
|
## Pure fine-tuning |
|
TBA |
|
|
|
# Transparency of Clean Diffusion |
|
I proof that clean diffusion is clean by following explanation. |
|
|
|
## Legal information |
|
TBA |
|
|
|
### Training |
|
Clean Diffusion is legal and ethical. |
|
|
|
Clean Diffusion is MADE IN JAPAN. |
|
Therefore, Clean Diffusion is subject to [Japanese copyright laws](https://en.wikipedia.org/wiki/Copyright_law_of_Japan). |
|
|
|
TBA |
|
|
|
### Generating |
|
|
|
TBA |
|
|
|
## Training Images |
|
TBA |
|
|
|
### List of works |
|
- [ArtBench](https://github.com/liaopeiyuan/artbench) (public domain is True) |
|
- Popeye the Sailor Meets Sindbad the Sailor |
|
|
|
### Tiny training images |
|
I would like to the all training raw images because these images are public domain. |
|
However, these images are huge (70GB+). |
|
Therefore, I have opened the tiny version like this. |
|
|
|
[Tiny Images](https://1drv.ms/u/s!ApxVlgxlqLRliLpSC58y5qyAlt52tQ?e=3Yfwbt) |
|
|
|
### Training Process of VAE |
|
|
|
TBA |
|
|
|
## Training text-image pairs |
|
TBA |
|
|
|
## Trainning code and config |
|
TBA |
|
|
|
# Acknowledgement |
|
> Standing on the shoulders of giants |
|
|
|
# Citations |
|
|
|
```bibtex |
|
@misc{rombach2021highresolution, |
|
title={High-Resolution Image Synthesis with Latent Diffusion Models}, |
|
author={Robin Rombach and Andreas Blattmann and Dominik Lorenz and Patrick Esser and Björn Ommer}, |
|
year={2021}, |
|
eprint={2112.10752}, |
|
archivePrefix={arXiv}, |
|
primaryClass={cs.CV} |
|
} |
|
``` |
|
```bibtex |
|
@article{liao2022artbench, |
|
title={The ArtBench Dataset: Benchmarking Generative Models with Artworks}, |
|
author={Liao, Peiyuan and Li, Xiuyu and Liu, Xihui and Keutzer, Kurt}, |
|
journal={arXiv preprint arXiv:2206.11404}, |
|
year={2022} |
|
} |
|
``` |
|
|
|
|