Datasets:
tianyu-z
commited on
Commit
•
5675338
1
Parent(s):
dd76206
README.md
CHANGED
@@ -56,6 +56,32 @@ We found that OCR and text-based processing become ineffective in VCR as accurat
|
|
56 |
- **Paper:** [VCR: Visual Caption Restoration](https://arxiv.org/abs/2406.06462)
|
57 |
- **Point of Contact:** [Tianyu Zhang](mailto:tianyu.zhang@mila.quebec)
|
58 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
59 |
# Model Evaluation
|
60 |
|
61 |
## Method 1 (recommended): use the evaluation script
|
|
|
56 |
- **Paper:** [VCR: Visual Caption Restoration](https://arxiv.org/abs/2406.06462)
|
57 |
- **Point of Contact:** [Tianyu Zhang](mailto:tianyu.zhang@mila.quebec)
|
58 |
|
59 |
+
# Benchmark
|
60 |
+
EM means `"Exact Match"` and Jaccard means `"Jaccard Similarity"`. The best in closed source and open source are highlighted in **bold**. Closed source models are evaluated based on 500 test samples, while open source models are evaluated based on 5000 test samples.
|
61 |
+
| Model | Size (unknown for closed source) | En Easy EM | En Easy Jaccard | En Hard EM | En Hard Jaccard | Zh Easy EM | Zh Easy Jaccard | Zh Hard EM | Zh Hard Jaccard |
|
62 |
+
|---|---|---|---|---|---|---|---|---|---|
|
63 |
+
| Claude 3 Opus | - | 62.0 | 77.67 | 37.8 | 57.68 | 0.9 | 11.5 | 0.3 | 9.22 |
|
64 |
+
| Claude 3.5 Sonnet | - | 63.85 | 74.65 | 41.74 | 56.15 | 1.0 | 7.54 | 0.2 | 4.0 |
|
65 |
+
| GPT-4 Turbo | - | 78.74 | 88.54 | 45.15 | 65.72 | 0.2 | 8.42 | 0.0 | 8.58 |
|
66 |
+
| GPT-4V | - | 52.04 | 65.36 | 25.83 | 44.63 | - | - | - | - |
|
67 |
+
| GPT-4o | - | **91.55** | **96.44** | **73.2** | **86.17** | **14.87** | **39.05** | **2.2** | **22.72** |
|
68 |
+
| Gemini 1.5 Pro | - | 62.73 | 77.71 | 28.07 | 51.9 | 1.1 | 11.1 | 0.7 | 11.82 |
|
69 |
+
| Qwen-VL-Max | - | 76.8 | 85.71 | 41.65 | 61.18 | 6.34 | 13.45 | 0.89 | 5.4 |
|
70 |
+
| Reka Core | - | 66.46 | 84.23 | 6.71 | 25.84 | 0.0 | 3.43 | 0.0 | 3.35 |
|
71 |
+
| CogVLM2 | 19B | **83.25** | **89.75** | **37.98** | **59.99** | - | - | - | - |
|
72 |
+
| CogVLM2-Chinese | 19B | - | - | - | - | **33.24** | **57.57** | **1.34** | **17.35** |
|
73 |
+
| DeepSeek-VL | 1.3B | 23.04 | 46.84 | 0.16 | 11.89 | 0.0 | 6.56 | 0.0 | 6.46 |
|
74 |
+
| DeepSeek-VL | 7B | 38.01 | 60.02 | 1.0 | 15.9 | 0.0 | 4.08 | 0.0 | 5.11 |
|
75 |
+
| DocOwl-1.5-Omni | 8B | 0.84 | 13.34 | 0.04 | 7.76 | 0.0 | 1.14 | 0.0 | 1.37 |
|
76 |
+
| Idefics2 | 8B | 15.75 | 31.97 | 0.65 | 9.93 | - | - | - | - |
|
77 |
+
| InternLM-XComposer2-VL | 7B | 46.64 | 70.99 | 0.7 | 12.51 | 0.27 | 12.32 | 0.07 | 8.97 |
|
78 |
+
| InternVL-V1.5 | 25.5B | 14.65 | 51.42 | 1.99 | 16.73 | 4.78 | 26.43 | 0.03 | 8.46 |
|
79 |
+
| MiniCPM-V2.5 | 8B | 31.81 | 53.24 | 1.41 | 11.94 | 4.1 | 18.03 | 0.09 | 7.39 |
|
80 |
+
| Monkey | 7B | 50.66 | 67.6 | 1.96 | 14.02 | 0.62 | 8.34 | 0.12 | 6.36 |
|
81 |
+
| Qwen-VL | 7B | 49.71 | 69.94 | 2.0 | 15.04 | 0.04 | 1.5 | 0.01 | 1.17 |
|
82 |
+
| Yi-VL | 34B | 0.82 | 5.59 | 0.07 | 4.31 | 0.0 | 4.44 | 0.0 | 4.12 |
|
83 |
+
| Yi-VL | 6B | 0.75 | 5.54 | 0.06 | 4.46 | 0.00 | 4.37 | 0.00 | 4.0 |
|
84 |
+
|
85 |
# Model Evaluation
|
86 |
|
87 |
## Method 1 (recommended): use the evaluation script
|