update_checkpoint
#6
by
wenmengzhou
- opened
- README.md +5 -20
- text2video_pytorch_model.pth +1 -1
README.md
CHANGED
@@ -5,11 +5,11 @@ pipeline_tag: text-to-video
|
|
5 |
|
6 |
The original repo is [here](https://modelscope.cn/models/damo/text-to-video-synthesis/summary).
|
7 |
|
8 |
-
|
9 |
|
10 |
If you're looking for an exciting challenge and the opportunity to work with cutting-edge technologies in AIGC and large-scale pretraining, then we are the place for you. We are looking for talented, motivated and creative individuals to join our team. If you are interested, please send your CV to us.
|
11 |
|
12 |
-
EMAIL:
|
13 |
|
14 |
This model is based on a multi-stage text-to-video generation diffusion model, which inputs a description text and returns a video that matches the text description. Only English input is supported.
|
15 |
|
@@ -25,17 +25,14 @@ This model has a wide range of applications and can reason and generate videos b
|
|
25 |
|
26 |
## How to use
|
27 |
|
28 |
-
|
29 |
-
The model has been launched on [ModelScope Studio](https://modelscope.cn/studios/damo/text-to-video-synthesis/summary) and [huggingface](https://huggingface.co/spaces/damo-vilab/modelscope-text-to-video-synthesis), you can experience it directly; you can also refer to [Colab page](https://colab.research.google.com/drive/1uW1ZqswkQ9Z9bp5Nbo5z59cAn7I0hE6R?usp=sharing#scrollTo=bSluBq99ObSk) to build it yourself.
|
30 |
-
In order to facilitate the experience of the model, users can refer to the [Aliyun Notebook Tutorial](https://modelscope.cn/headlines/detail/26) to quickly develop this Text-to-Video model.
|
31 |
-
|
32 |
-
This demo requires about 16GB CPU RAM and 16GB GPU RAM. Under the ModelScope framework, the current model can be used by calling a simple Pipeline, where the input must be in dictionary format, the legal key value is 'text', and the content is a short text. This model currently only supports inference on the GPU. Enter specific code examples as follows:
|
33 |
|
|
|
34 |
|
35 |
### Operating environment (Python Package)
|
36 |
|
37 |
```
|
38 |
-
pip install modelscope
|
39 |
pip install open_clip_torch
|
40 |
pip install pytorch-lightning
|
41 |
```
|
@@ -85,15 +82,3 @@ The output mp4 file can be viewed by [VLC media player](https://www.videolan.org
|
|
85 |
## Training data
|
86 |
|
87 |
The training data includes [LAION5B](https://huggingface.co/datasets/laion/laion2B-en), [ImageNet](https://www.image-net.org/), [Webvid](https://m-bain.github.io/webvid-dataset/) and other public datasets. Image and video filtering is performed after pre-training such as aesthetic score, watermark score, and deduplication.
|
88 |
-
|
89 |
-
## Citation
|
90 |
-
|
91 |
-
```bibtex
|
92 |
-
@InProceedings{VideoFusion,
|
93 |
-
author = {Luo, Zhengxiong and Chen, Dayou and Zhang, Yingya and Huang, Yan and Wang, Liang and Shen, Yujun and Zhao, Deli and Zhou, Jingren and Tan, Tieniu},
|
94 |
-
title = {VideoFusion: Decomposed Diffusion Models for High-Quality Video Generation},
|
95 |
-
booktitle = {Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR)},
|
96 |
-
month = {June},
|
97 |
-
year = {2023}
|
98 |
-
}
|
99 |
-
```
|
|
|
5 |
|
6 |
The original repo is [here](https://modelscope.cn/models/damo/text-to-video-synthesis/summary).
|
7 |
|
8 |
+
We Are Hiring! (Based on Beijing / Hangzhou, China.)
|
9 |
|
10 |
If you're looking for an exciting challenge and the opportunity to work with cutting-edge technologies in AIGC and large-scale pretraining, then we are the place for you. We are looking for talented, motivated and creative individuals to join our team. If you are interested, please send your CV to us.
|
11 |
|
12 |
+
EMAIL: wangjiuniu.wjn@alibaba-inc.com
|
13 |
|
14 |
This model is based on a multi-stage text-to-video generation diffusion model, which inputs a description text and returns a video that matches the text description. Only English input is supported.
|
15 |
|
|
|
25 |
|
26 |
## How to use
|
27 |
|
28 |
+
Under the ModelScope framework, the current model can be used by calling a simple Pipeline, where the input must be in dictionary format, the legal key value is 'text', and the content is a short text. This model currently only supports inference on the GPU. Enter specific code examples as follows:
|
|
|
|
|
|
|
|
|
29 |
|
30 |
+
For Colab usage, you can view [this webpage](https://colab.research.google.com/drive/1uW1ZqswkQ9Z9bp5Nbo5z59cAn7I0hE6R?usp=sharing).
|
31 |
|
32 |
### Operating environment (Python Package)
|
33 |
|
34 |
```
|
35 |
+
pip install git+https://github.com/modelscope/modelscope.git
|
36 |
pip install open_clip_torch
|
37 |
pip install pytorch-lightning
|
38 |
```
|
|
|
82 |
## Training data
|
83 |
|
84 |
The training data includes [LAION5B](https://huggingface.co/datasets/laion/laion2B-en), [ImageNet](https://www.image-net.org/), [Webvid](https://m-bain.github.io/webvid-dataset/) and other public datasets. Image and video filtering is performed after pre-training such as aesthetic score, watermark score, and deduplication.
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
text2video_pytorch_model.pth
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 5645549049
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:d9609d02717b799137a97244844ab6df0d1a071568a1d24dcb62d9050f3a24a3
|
3 |
size 5645549049
|