Upload model_card.md
Browse files- model_card.md +41 -0
model_card.md
ADDED
@@ -0,0 +1,41 @@
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
1 |
+
# Model Card: DALL路E dVAE
|
2 |
+
|
3 |
+
Following [Model Cards for Model Reporting (Mitchell et al.)](https://arxiv.org/abs/1810.03993) and [Lessons from
|
4 |
+
Archives (Jo & Gebru)](https://arxiv.org/pdf/1912.10389.pdf), we're providing some information about about the discrete
|
5 |
+
VAE (dVAE) that was used to train DALL路E.
|
6 |
+
|
7 |
+
## Model Details
|
8 |
+
|
9 |
+
The dVAE was developed by researchers at OpenAI to reduce the memory footprint of the transformer trained on the
|
10 |
+
text-to-image generation task. The details involved in training the dVAE are described in [the paper][dalle_paper]. This
|
11 |
+
model card describes the first version of the model, released in February 2021. The model consists of a convolutional
|
12 |
+
encoder and decoder whose architectures are described [here](dall_e/encoder.py) and [here](dall_e/decoder.py), respectively.
|
13 |
+
For questions or comments about the models or the code release, please file a Github issue.
|
14 |
+
|
15 |
+
## Model Use
|
16 |
+
|
17 |
+
### Intended Use
|
18 |
+
|
19 |
+
The model is intended for others to use for training their own generative models.
|
20 |
+
|
21 |
+
### Out-of-Scope Use Cases
|
22 |
+
|
23 |
+
This model is inappropriate for high-fidelity image processing applications. We also do not recommend its use as a
|
24 |
+
general-purpose image compressor.
|
25 |
+
|
26 |
+
## Training Data
|
27 |
+
|
28 |
+
The model was trained on publicly available text-image pairs collected from the internet. This data consists partly of
|
29 |
+
[Conceptual Captions][cc] and a filtered subset of [YFCC100M][yfcc100m]. We used a subset of the filters described in
|
30 |
+
[Sharma et al.][cc_paper] to construct this dataset; further details are described in [our paper][dalle_paper]. We will
|
31 |
+
not be releasing the dataset.
|
32 |
+
|
33 |
+
## Performance and Limitations
|
34 |
+
|
35 |
+
The heavy compression from the encoding process results in a noticeable loss of detail in the reconstructed images. This
|
36 |
+
renders it inappropriate for applications that require fine-grained details of the image to be preserved.
|
37 |
+
|
38 |
+
[dalle_paper]: https://arxiv.org/abs/2102.12092
|
39 |
+
[cc]: https://ai.google.com/research/ConceptualCaptions
|
40 |
+
[cc_paper]: https://www.aclweb.org/anthology/P18-1238/
|
41 |
+
[yfcc100m]: http://projects.dfki.uni-kl.de/yfcc100m/
|