gordonhubackup commited on
Commit
7f7df37
·
1 Parent(s): e1fbeb3
Files changed (1) hide show
  1. README.md +3 -3
README.md CHANGED
@@ -23,13 +23,13 @@ It composes of an EVA-CLIP vision encoder, a Q-Former, a projection layer and an
23
  LoViM_FlanT5 was trained in July 2023.
24
 
25
  **Paper or resources for more information:**
26
- https://project page
27
 
28
  **License:**
29
  BSD 3-Clause License
30
 
31
  **Where to send questions or comments about the model:**
32
- https://github.com/
33
 
34
  ## Intended use
35
  **Primary intended uses:**
@@ -48,4 +48,4 @@ For zero-shot evaluation on general image task, we selected Nocaps, Flickr30K, V
48
 
49
  For zero-shot evaluation on text-rich image OCR task, we selected ST-VQA, OCR-VQA, Text-VQA, and Doc-VQA.
50
 
51
- More detials are in our github, https://github.com/
 
23
  LoViM_FlanT5 was trained in July 2023.
24
 
25
  **Paper or resources for more information:**
26
+ https://gordonhu608.github.io/lovim/
27
 
28
  **License:**
29
  BSD 3-Clause License
30
 
31
  **Where to send questions or comments about the model:**
32
+ https://github.com/mlpc-ucsd/LoViM
33
 
34
  ## Intended use
35
  **Primary intended uses:**
 
48
 
49
  For zero-shot evaluation on text-rich image OCR task, we selected ST-VQA, OCR-VQA, Text-VQA, and Doc-VQA.
50
 
51
+ More detials are in our github, https://github.com/mlpc-ucsd/LoViM