Commit
·
7f7df37
1
Parent(s):
e1fbeb3
link
Browse files
README.md
CHANGED
@@ -23,13 +23,13 @@ It composes of an EVA-CLIP vision encoder, a Q-Former, a projection layer and an
|
|
23 |
LoViM_FlanT5 was trained in July 2023.
|
24 |
|
25 |
**Paper or resources for more information:**
|
26 |
-
https://
|
27 |
|
28 |
**License:**
|
29 |
BSD 3-Clause License
|
30 |
|
31 |
**Where to send questions or comments about the model:**
|
32 |
-
https://github.com/
|
33 |
|
34 |
## Intended use
|
35 |
**Primary intended uses:**
|
@@ -48,4 +48,4 @@ For zero-shot evaluation on general image task, we selected Nocaps, Flickr30K, V
|
|
48 |
|
49 |
For zero-shot evaluation on text-rich image OCR task, we selected ST-VQA, OCR-VQA, Text-VQA, and Doc-VQA.
|
50 |
|
51 |
-
More detials are in our github, https://github.com/
|
|
|
23 |
LoViM_FlanT5 was trained in July 2023.
|
24 |
|
25 |
**Paper or resources for more information:**
|
26 |
+
https://gordonhu608.github.io/lovim/
|
27 |
|
28 |
**License:**
|
29 |
BSD 3-Clause License
|
30 |
|
31 |
**Where to send questions or comments about the model:**
|
32 |
+
https://github.com/mlpc-ucsd/LoViM
|
33 |
|
34 |
## Intended use
|
35 |
**Primary intended uses:**
|
|
|
48 |
|
49 |
For zero-shot evaluation on text-rich image OCR task, we selected ST-VQA, OCR-VQA, Text-VQA, and Doc-VQA.
|
50 |
|
51 |
+
More detials are in our github, https://github.com/mlpc-ucsd/LoViM
|