ruohongz commited on
Commit
a4d5817
1 Parent(s): 46b9204

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +41 -1
README.md CHANGED
@@ -1,3 +1,43 @@
1
  ---
2
- license: mit
 
3
  ---
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
  ---
2
+ inference: false
3
+ license: apache-2.0
4
  ---
5
+
6
+ <br>
7
+ <br>
8
+
9
+ # LLaVA-Hound Model Card
10
+
11
+ ## Model details
12
+
13
+ **Model type:**
14
+ LLaVA-Hound is an open-source video large multimodal model, fine-tuned from video instruction following data based on large language model.
15
+
16
+ This model is the pre-trained ckpt on image and video data.
17
+
18
+ Base LLM: [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-7b-v1.5)
19
+
20
+ **Model date:**
21
+ Trained on March 15, 2024.
22
+
23
+ **Paper or resources for more information:**
24
+ https://github.com/RifleZhang/LLaVA-Hound-DPO
25
+
26
+ ## License
27
+ [lmsys/vicuna-7b-v1.5](https://huggingface.co/lmsys/vicuna-7b-v1.5) license.
28
+
29
+ **Where to send questions or comments about the model:**
30
+ https://github.com/RifleZhang/LLaVA-Hound-DPO/issues
31
+
32
+ ## Intended use
33
+ **Primary intended uses:**
34
+ Video detailed captioning
35
+
36
+ **Primary intended users:**
37
+ Researchers in artificial intelligence, large multimodal model, etc.
38
+
39
+ ## Training dataset
40
+ ShareGPTVideo dataset.
41
+
42
+ ## Evaluation
43
+ Follow https://github.com/RifleZhang/LLaVA-Hound-DPO/blob/main/README.md