kcz358 commited on
Commit
91181d7
1 Parent(s): 5e399cd

Upload tokenizer

Browse files
Files changed (4) hide show
  1. README.md +63 -64
  2. added_tokens.json +1 -0
  3. tokenizer.json +10 -0
  4. tokenizer_config.json +8 -0
README.md CHANGED
@@ -1,218 +1,217 @@
1
  ---
2
- license: apache-2.0
3
  datasets:
4
  - lmms-lab/LLaVA-OneVision-Data
5
  language:
6
  - en
7
  - zh
 
 
8
  metrics:
9
  - accuracy
10
- library_name: transformers
11
  tags:
12
  - multimodal
13
-
14
  model-index:
15
  - name: llava-onevision-qwen-0.5b-si
16
  results:
17
  - task:
18
  type: multimodal
19
  dataset:
20
- type: ai2d
21
  name: AI2D
 
22
  metrics:
23
- - name: accuracy
24
- type: accuracy
25
  value: 54.2
 
26
  verified: true
27
  - task:
28
  type: multimodal
29
  dataset:
30
- type: chartqa
31
  name: ChartQA
 
32
  metrics:
33
- - name: accuracy
34
- type: accuracy
35
  value: 61.0
 
36
  verified: true
37
  - task:
38
  type: multimodal
39
  dataset:
40
- type: docvqa
41
  name: DocVQA
 
42
  metrics:
43
- - name: accuracy
44
- type: accuracy
45
  value: 75.0
 
46
  verified: true
47
  - task:
48
  type: multimodal
49
  dataset:
50
- type: infovqa
51
  name: InfoVQA
 
52
  metrics:
53
- - name: accuracy
54
- type: accuracy
55
  value: 44.8
 
56
  verified: true
57
  - task:
58
  type: multimodal
59
  dataset:
60
- type: mathverse
61
  name: MathVerse
 
62
  metrics:
63
- - name: accuracy
64
- type: accuracy
65
  value: 17.3
 
66
  verified: true
67
  - task:
68
  type: multimodal
69
  dataset:
70
- type: mathvista
71
  name: MathVista
 
72
  metrics:
73
- - name: accuracy
74
- type: accuracy
75
  value: 34.6
 
76
  verified: true
77
  - task:
78
  type: multimodal
79
  dataset:
80
- type: mmbench
81
  name: MMBench
 
82
  metrics:
83
- - name: accuracy
84
- type: accuracy
85
  value: 43.8
 
86
  verified: true
87
  - task:
88
  type: multimodal
89
  dataset:
90
- type: mme-perception
91
  name: MME-Perception
 
92
  metrics:
93
- - name: score
94
- type: score
95
  value: 1217
 
96
  verified: true
97
  - task:
98
  type: multimodal
99
  dataset:
100
- type: mme-cognition
101
  name: MME-Cognition
 
102
  metrics:
103
- - name: score
104
- type: score
105
  value: 272
106
- verified: true
 
107
  - task:
108
  type: multimodal
109
  dataset:
110
- type: mmmu
111
  name: MMMU
 
112
  metrics:
113
- - name: accuracy
114
- type: accuracy
115
  value: 31.2
 
116
  verified: true
117
  - task:
118
  type: multimodal
119
  dataset:
120
- type: mmvet
121
  name: MMVet
 
122
  metrics:
123
- - name: accuracy
124
- type: accuracy
125
  value: 26.9
 
126
  verified: true
127
  - task:
128
  type: multimodal
129
  dataset:
130
- type: mmstar
131
  name: MMStar
 
132
  metrics:
133
- - name: accuracy
134
- type: accuracy
135
  value: 36.3
 
136
  verified: true
137
  - task:
138
  type: multimodal
139
  dataset:
140
- type: seed-bench
141
  name: Seed-Bench
 
142
  metrics:
143
- - name: accuracy
144
- type: accuracy
145
  value: 63.4
 
146
  verified: true
147
  - task:
148
  type: multimodal
149
  dataset:
150
- type: science-qa
151
  name: Science-QA
 
152
  metrics:
153
- - name: accuracy
154
- type: accuracy
155
  value: 67.8
 
156
  verified: true
157
  - task:
158
  type: multimodal
159
  dataset:
160
- type: imagedc
161
  name: ImageDC
 
162
  metrics:
163
- - name: accuracy
164
- type: accuracy
165
  value: 83.0
 
166
  verified: true
167
  - task:
168
  type: multimodal
169
  dataset:
170
- type: mmlbench
171
  name: MMLBench
 
172
  metrics:
173
- - name: accuracy
174
- type: accuracy
175
  value: 43.2
 
176
  verified: true
177
  - task:
178
  type: multimodal
179
  dataset:
180
- type: realworldqa
181
  name: RealWorldQA
 
182
  metrics:
183
- - name: accuracy
184
- type: accuracy
185
  value: 53.7
 
186
  verified: true
187
  - task:
188
  type: multimodal
189
  dataset:
190
- type: vibe-eval
191
  name: Vibe-Eval
 
192
  metrics:
193
- - name: accuracy
194
- type: accuracy
195
  value: 34.9
 
196
  verified: true
197
  - task:
198
  type: multimodal
199
  dataset:
200
- type: llava-w
201
  name: LLaVA-W
 
202
  metrics:
203
- - name: accuracy
204
- type: accuracy
205
  value: 71.2
 
206
  verified: true
207
  - task:
208
  type: multimodal
209
  dataset:
210
- type: l-wilder
211
  name: L-Wilder
 
212
  metrics:
213
- - name: accuracy
214
- type: accuracy
215
  value: 51.5
 
216
  verified: true
217
  ---
218
 
 
1
  ---
 
2
  datasets:
3
  - lmms-lab/LLaVA-OneVision-Data
4
  language:
5
  - en
6
  - zh
7
+ library_name: transformers
8
+ license: apache-2.0
9
  metrics:
10
  - accuracy
 
11
  tags:
12
  - multimodal
 
13
  model-index:
14
  - name: llava-onevision-qwen-0.5b-si
15
  results:
16
  - task:
17
  type: multimodal
18
  dataset:
 
19
  name: AI2D
20
+ type: ai2d
21
  metrics:
22
+ - type: accuracy
 
23
  value: 54.2
24
+ name: accuracy
25
  verified: true
26
  - task:
27
  type: multimodal
28
  dataset:
 
29
  name: ChartQA
30
+ type: chartqa
31
  metrics:
32
+ - type: accuracy
 
33
  value: 61.0
34
+ name: accuracy
35
  verified: true
36
  - task:
37
  type: multimodal
38
  dataset:
 
39
  name: DocVQA
40
+ type: docvqa
41
  metrics:
42
+ - type: accuracy
 
43
  value: 75.0
44
+ name: accuracy
45
  verified: true
46
  - task:
47
  type: multimodal
48
  dataset:
 
49
  name: InfoVQA
50
+ type: infovqa
51
  metrics:
52
+ - type: accuracy
 
53
  value: 44.8
54
+ name: accuracy
55
  verified: true
56
  - task:
57
  type: multimodal
58
  dataset:
 
59
  name: MathVerse
60
+ type: mathverse
61
  metrics:
62
+ - type: accuracy
 
63
  value: 17.3
64
+ name: accuracy
65
  verified: true
66
  - task:
67
  type: multimodal
68
  dataset:
 
69
  name: MathVista
70
+ type: mathvista
71
  metrics:
72
+ - type: accuracy
 
73
  value: 34.6
74
+ name: accuracy
75
  verified: true
76
  - task:
77
  type: multimodal
78
  dataset:
 
79
  name: MMBench
80
+ type: mmbench
81
  metrics:
82
+ - type: accuracy
 
83
  value: 43.8
84
+ name: accuracy
85
  verified: true
86
  - task:
87
  type: multimodal
88
  dataset:
 
89
  name: MME-Perception
90
+ type: mme-perception
91
  metrics:
92
+ - type: score
 
93
  value: 1217
94
+ name: score
95
  verified: true
96
  - task:
97
  type: multimodal
98
  dataset:
 
99
  name: MME-Cognition
100
+ type: mme-cognition
101
  metrics:
102
+ - type: score
 
103
  value: 272
104
+ name: score
105
+ verified: true
106
  - task:
107
  type: multimodal
108
  dataset:
 
109
  name: MMMU
110
+ type: mmmu
111
  metrics:
112
+ - type: accuracy
 
113
  value: 31.2
114
+ name: accuracy
115
  verified: true
116
  - task:
117
  type: multimodal
118
  dataset:
 
119
  name: MMVet
120
+ type: mmvet
121
  metrics:
122
+ - type: accuracy
 
123
  value: 26.9
124
+ name: accuracy
125
  verified: true
126
  - task:
127
  type: multimodal
128
  dataset:
 
129
  name: MMStar
130
+ type: mmstar
131
  metrics:
132
+ - type: accuracy
 
133
  value: 36.3
134
+ name: accuracy
135
  verified: true
136
  - task:
137
  type: multimodal
138
  dataset:
 
139
  name: Seed-Bench
140
+ type: seed-bench
141
  metrics:
142
+ - type: accuracy
 
143
  value: 63.4
144
+ name: accuracy
145
  verified: true
146
  - task:
147
  type: multimodal
148
  dataset:
 
149
  name: Science-QA
150
+ type: science-qa
151
  metrics:
152
+ - type: accuracy
 
153
  value: 67.8
154
+ name: accuracy
155
  verified: true
156
  - task:
157
  type: multimodal
158
  dataset:
 
159
  name: ImageDC
160
+ type: imagedc
161
  metrics:
162
+ - type: accuracy
 
163
  value: 83.0
164
+ name: accuracy
165
  verified: true
166
  - task:
167
  type: multimodal
168
  dataset:
 
169
  name: MMLBench
170
+ type: mmlbench
171
  metrics:
172
+ - type: accuracy
 
173
  value: 43.2
174
+ name: accuracy
175
  verified: true
176
  - task:
177
  type: multimodal
178
  dataset:
 
179
  name: RealWorldQA
180
+ type: realworldqa
181
  metrics:
182
+ - type: accuracy
 
183
  value: 53.7
184
+ name: accuracy
185
  verified: true
186
  - task:
187
  type: multimodal
188
  dataset:
 
189
  name: Vibe-Eval
190
+ type: vibe-eval
191
  metrics:
192
+ - type: accuracy
 
193
  value: 34.9
194
+ name: accuracy
195
  verified: true
196
  - task:
197
  type: multimodal
198
  dataset:
 
199
  name: LLaVA-W
200
+ type: llava-w
201
  metrics:
202
+ - type: accuracy
 
203
  value: 71.2
204
+ name: accuracy
205
  verified: true
206
  - task:
207
  type: multimodal
208
  dataset:
 
209
  name: L-Wilder
210
+ type: l-wilder
211
  metrics:
212
+ - type: accuracy
 
213
  value: 51.5
214
+ name: accuracy
215
  verified: true
216
  ---
217
 
added_tokens.json CHANGED
@@ -1,4 +1,5 @@
1
  {
 
2
  "<|endoftext|>": 151643,
3
  "<|im_end|>": 151645,
4
  "<|im_start|>": 151644
 
1
  {
2
+ "<image>": 151646,
3
  "<|endoftext|>": 151643,
4
  "<|im_end|>": 151645,
5
  "<|im_start|>": 151644
tokenizer.json CHANGED
@@ -29,6 +29,15 @@
29
  "rstrip": false,
30
  "normalized": false,
31
  "special": true
 
 
 
 
 
 
 
 
 
32
  }
33
  ],
34
  "normalizer": {
@@ -73,6 +82,7 @@
73
  "end_of_word_suffix": "",
74
  "fuse_unk": false,
75
  "byte_fallback": false,
 
76
  "vocab": {
77
  "!": 0,
78
  "\"": 1,
 
29
  "rstrip": false,
30
  "normalized": false,
31
  "special": true
32
+ },
33
+ {
34
+ "id": 151646,
35
+ "content": "<image>",
36
+ "single_word": false,
37
+ "lstrip": false,
38
+ "rstrip": false,
39
+ "normalized": false,
40
+ "special": true
41
  }
42
  ],
43
  "normalizer": {
 
82
  "end_of_word_suffix": "",
83
  "fuse_unk": false,
84
  "byte_fallback": false,
85
+ "ignore_merges": false,
86
  "vocab": {
87
  "!": 0,
88
  "\"": 1,
tokenizer_config.json CHANGED
@@ -24,6 +24,14 @@
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
 
 
 
 
 
 
 
 
27
  }
28
  },
29
  "additional_special_tokens": [
 
24
  "rstrip": false,
25
  "single_word": false,
26
  "special": true
27
+ },
28
+ "151646": {
29
+ "content": "<image>",
30
+ "lstrip": false,
31
+ "normalized": false,
32
+ "rstrip": false,
33
+ "single_word": false,
34
+ "special": true
35
  }
36
  },
37
  "additional_special_tokens": [