alfredplpl's picture
first
8ded8b2
|
raw
history blame
3.28 kB
---
license: openrail++
tags:
- stable-diffusion
- text-to-image
---
# Clean Diffusion 2.0 PoC Model Card
> That's one small step for artists, one giant leap for engineers.
![ocean](ocean.png)
Clean Diffusion 2.0 PoC is Latent Diffusion Model made of public domain images.
Clean Diffusion 2.0 PoC is for the proof of the concept: Stable Diffusion can be made of public domain images.
Therefore, the model can only express the ocean.
If you are Japanese, I recommend Clean Diffusion For Japanese (TBA) instead of Clean Diffusion (For Global).
The model is more powerful than this global version.
# Note
> With great power comes great responsibility.
If you **CANNOT UNDERSTAND THESE WORDS**, I recommend that **YOU SHOULD NOT USE ALL OF DIFFUSION MODELS** what have great powers.
# Getting Started
You would be able to use Clean Diffusion by the following code soon.
```python
from diffusers import StableDiffusionPipeline
import torch
model_id = "alfredplpl/clean-diffusion-2-0-poc"
pipe = StableDiffusionPipeline.from_pretrained(model_id, torch_dtype=torch.float32)
pipe = pipe.to("cuda")
prompt = "Cartoon, ocean."
image = pipe(prompt).images[0]
image.save("ocean.png")
```
# Tuning
Clean Diffusion is less powerful than Stable Diffusion.
Therefore, I recommend to tune Clean Diffusion like Stable Diffusion
because Clean Diffusion of the network architecture and Stable Diffusion of the network architecture are same.
And I repeat the words before I explain the detail.
> With great power comes great responsibility.
Please consider the words before you tune Clean Diffusion.
## Textual Inversion
TBA on Colab.
## Dreambooth on Stable Diffusion
TBA on Colab.
## Pure fine-tuning
TBA
# Transparency of Clean Diffusion
I proof that clean diffusion is clean by following explanation.
## Legal information
TBA
### Training
Clean Diffusion is legal and ethical.
Clean Diffusion is MADE IN JAPAN.
Therefore, Clean Diffusion is subject to [Japanese copyright laws](https://en.wikipedia.org/wiki/Copyright_law_of_Japan).
TBA
### Generating
TBA
## Training Images
TBA
### List of works
- [ArtBench](https://github.com/liaopeiyuan/artbench) (public domain is True)
- Popeye the Sailor Meets Sindbad the Sailor
### Tiny training images
I would like to the all training raw images because these images are public domain.
However, these images are huge (70GB+).
Therefore, I have opened the tiny version like this.
[Tiny Images](https://1drv.ms/u/s!ApxVlgxlqLRliLpSC58y5qyAlt52tQ?e=3Yfwbt)
### Training Process of VAE
TBA
## Training text-image pairs
TBA
## Trainning code and config
TBA
# Acknowledgement
> Standing on the shoulders of giants
# Citations
```bibtex
@misc{rombach2021highresolution,
title={High-Resolution Image Synthesis with Latent Diffusion Models},
author={Robin Rombach and Andreas Blattmann and Dominik Lorenz and Patrick Esser and Björn Ommer},
year={2021},
eprint={2112.10752},
archivePrefix={arXiv},
primaryClass={cs.CV}
}
```
```bibtex
@article{liao2022artbench,
title={The ArtBench Dataset: Benchmarking Generative Models with Artworks},
author={Liao, Peiyuan and Li, Xiuyu and Liu, Xihui and Keutzer, Kurt},
journal={arXiv preprint arXiv:2206.11404},
year={2022}
}
```