EdwinHuang commited on
Commit
4c040ee
·
verified ·
1 Parent(s): 58a7702

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +28 -27
README.md CHANGED
@@ -4,13 +4,15 @@ base_model: Qwen/Qwen2.5-7B-Instruct
4
  library_name: transformers
5
  pipeline_tag: image-to-text
6
  tags:
7
- - multimodal
8
- - video-understanding
9
- - spatial-reasoning
10
- - vision-language
11
  datasets:
12
- - nyu-visionx/VSI-590K
13
- eval_results:
 
 
14
  - task:
15
  type: visual-question-answering
16
  name: VSI-Bench
@@ -19,12 +21,18 @@ eval_results:
19
  name: VSI-Bench
20
  metrics:
21
  - type: accuracy
22
- value: 67.5
23
  name: accuracy
24
- verified: false
25
  - task:
26
  type: visual-question-answering
27
  name: VSI-Bench-Debiased
 
 
 
 
 
 
 
28
  - task:
29
  type: visual-question-answering
30
  name: EgoSchema
@@ -33,9 +41,8 @@ eval_results:
33
  name: EgoSchema
34
  metrics:
35
  - type: accuracy
36
- value: 76.8
37
  name: accuracy
38
- verified: false
39
  - task:
40
  type: visual-question-answering
41
  name: Perception Test
@@ -44,9 +51,8 @@ eval_results:
44
  name: Perception Test
45
  metrics:
46
  - type: accuracy
47
- value: 69.9
48
  name: accuracy
49
- verified: false
50
  - task:
51
  type: visual-question-answering
52
  name: VideoMME
@@ -55,9 +61,8 @@ eval_results:
55
  name: VideoMME
56
  metrics:
57
  - type: accuracy
58
- value: 63.4
59
  name: accuracy
60
- verified: false
61
  - task:
62
  type: visual-question-answering
63
  name: MVBench
@@ -66,9 +71,8 @@ eval_results:
66
  name: MVBench
67
  metrics:
68
  - type: accuracy
69
- value: 64.5
70
  name: accuracy
71
- verified: false
72
  - task:
73
  type: visual-question-answering
74
  name: LongVideoBench
@@ -77,9 +81,8 @@ eval_results:
77
  name: LongVideoBench
78
  metrics:
79
  - type: accuracy
80
- value: 59.4
81
  name: accuracy
82
- verified: false
83
  - task:
84
  type: visual-question-answering
85
  name: VideoMMMU
@@ -88,9 +91,8 @@ eval_results:
88
  name: VideoMMMU
89
  metrics:
90
  - type: accuracy
91
- value: 38.6
92
  name: accuracy
93
- verified: false
94
  - task:
95
  type: visual-question-answering
96
  name: MMVP
@@ -99,9 +101,8 @@ eval_results:
99
  name: MMVP
100
  metrics:
101
  - type: accuracy
102
- value: 60.0
103
  name: accuracy
104
- verified: false
105
  - task:
106
  type: visual-question-answering
107
  name: 3DSR
@@ -110,9 +111,8 @@ eval_results:
110
  name: 3DSR
111
  metrics:
112
  - type: accuracy
113
- value: 54.8
114
  name: accuracy
115
- verified: false
116
  - task:
117
  type: visual-question-answering
118
  name: CV-Bench
@@ -121,9 +121,10 @@ eval_results:
121
  name: CV-Bench
122
  metrics:
123
  - type: accuracy
124
- value: 76.9
125
  name: accuracy
126
- verified: false
 
 
127
  ---
128
 
129
 
@@ -175,4 +176,4 @@ output_ids = model.generate(input_ids, images=image_tensor, image_sizes=image_si
175
  journal={arXiv preprint arXiv:2025},
176
  year={2025}
177
  }
178
- ```
 
4
  library_name: transformers
5
  pipeline_tag: image-to-text
6
  tags:
7
+ - multimodal
8
+ - video-understanding
9
+ - spatial-reasoning
10
+ - vision-language
11
  datasets:
12
+ - nyu-visionx/VSI-590K
13
+ model-index:
14
+ - name: Cambrian-S-7B
15
+ results:
16
  - task:
17
  type: visual-question-answering
18
  name: VSI-Bench
 
21
  name: VSI-Bench
22
  metrics:
23
  - type: accuracy
 
24
  name: accuracy
25
+ value: 67.5
26
  - task:
27
  type: visual-question-answering
28
  name: VSI-Bench-Debiased
29
+ dataset:
30
+ type: vsi-bench-debiased
31
+ name: VSI-Bench-Debiased
32
+ metrics:
33
+ - type: accuracy
34
+ name: accuracy
35
+ value: 59.9
36
  - task:
37
  type: visual-question-answering
38
  name: EgoSchema
 
41
  name: EgoSchema
42
  metrics:
43
  - type: accuracy
 
44
  name: accuracy
45
+ value: 76.8
46
  - task:
47
  type: visual-question-answering
48
  name: Perception Test
 
51
  name: Perception Test
52
  metrics:
53
  - type: accuracy
 
54
  name: accuracy
55
+ value: 69.9
56
  - task:
57
  type: visual-question-answering
58
  name: VideoMME
 
61
  name: VideoMME
62
  metrics:
63
  - type: accuracy
 
64
  name: accuracy
65
+ value: 63.4
66
  - task:
67
  type: visual-question-answering
68
  name: MVBench
 
71
  name: MVBench
72
  metrics:
73
  - type: accuracy
 
74
  name: accuracy
75
+ value: 64.5
76
  - task:
77
  type: visual-question-answering
78
  name: LongVideoBench
 
81
  name: LongVideoBench
82
  metrics:
83
  - type: accuracy
 
84
  name: accuracy
85
+ value: 59.4
86
  - task:
87
  type: visual-question-answering
88
  name: VideoMMMU
 
91
  name: VideoMMMU
92
  metrics:
93
  - type: accuracy
 
94
  name: accuracy
95
+ value: 38.6
96
  - task:
97
  type: visual-question-answering
98
  name: MMVP
 
101
  name: MMVP
102
  metrics:
103
  - type: accuracy
 
104
  name: accuracy
105
+ value: 60.0
106
  - task:
107
  type: visual-question-answering
108
  name: 3DSR
 
111
  name: 3DSR
112
  metrics:
113
  - type: accuracy
 
114
  name: accuracy
115
+ value: 54.8
116
  - task:
117
  type: visual-question-answering
118
  name: CV-Bench
 
121
  name: CV-Bench
122
  metrics:
123
  - type: accuracy
 
124
  name: accuracy
125
+ value: 76.9
126
+ language:
127
+ - en
128
  ---
129
 
130
 
 
176
  journal={arXiv preprint arXiv:2025},
177
  year={2025}
178
  }
179
+ ```