prince-canuma commited on
Commit
bd295a1
1 Parent(s): 26727df

Upload folder using huggingface_hub

Browse files
README.md CHANGED
@@ -1,36 +1,36 @@
1
- ---
2
- language:
3
- - en
4
- license: apache-2.0
5
- tags:
6
- - multimodal
7
- - vision
8
- - image-text-to-text
9
- - mlx
10
- datasets:
11
- - HuggingFaceM4/OBELICS
12
- - laion/laion-coco
13
- - wikipedia
14
- - facebook/pmd
15
- - pixparse/idl-wds
16
- - pixparse/pdfa-eng-wds
17
- - wendlerc/RenderedText
18
- - HuggingFaceM4/the_cauldron
19
- - teknium/OpenHermes-2.5
20
- - GAIR/lima
21
- - databricks/databricks-dolly-15k
22
- - meta-math/MetaMathQA
23
- - TIGER-Lab/MathInstruct
24
- - microsoft/orca-math-word-problems-200k
25
- - camel-ai/math
26
- - AtlasUnified/atlas-math-sets
27
- - tiedong/goat
28
- - Lin-Chen/ShareGPT4V
29
- - jxu124/llava_conversation_58k
30
- ---
31
 
32
  # mlx-community/idefics2-8b-chatty-8bit
33
- This model was converted to MLX format from [`HuggingFaceM4/idefics2-8b-chatty`]() using mlx-vlm version **0.0.4**.
34
  Refer to the [original model card](https://huggingface.co/HuggingFaceM4/idefics2-8b-chatty) for more details on the model.
35
  ## Use with mlx
36
 
 
1
+ ---
2
+ datasets:
3
+ - HuggingFaceM4/OBELICS
4
+ - laion/laion-coco
5
+ - wikipedia
6
+ - facebook/pmd
7
+ - pixparse/idl-wds
8
+ - pixparse/pdfa-eng-wds
9
+ - wendlerc/RenderedText
10
+ - HuggingFaceM4/the_cauldron
11
+ - teknium/OpenHermes-2.5
12
+ - GAIR/lima
13
+ - databricks/databricks-dolly-15k
14
+ - meta-math/MetaMathQA
15
+ - TIGER-Lab/MathInstruct
16
+ - microsoft/orca-math-word-problems-200k
17
+ - camel-ai/math
18
+ - AtlasUnified/atlas-math-sets
19
+ - tiedong/goat
20
+ - Lin-Chen/ShareGPT4V
21
+ - jxu124/llava_conversation_58k
22
+ language:
23
+ - en
24
+ license: apache-2.0
25
+ tags:
26
+ - multimodal
27
+ - vision
28
+ - image-text-to-text
29
+ - mlx
30
+ ---
31
 
32
  # mlx-community/idefics2-8b-chatty-8bit
33
+ This model was converted to MLX format from [`HuggingFaceM4/idefics2-8b-chatty`]() using mlx-vlm version **0.1.0**.
34
  Refer to the [original model card](https://huggingface.co/HuggingFaceM4/idefics2-8b-chatty) for more details on the model.
35
  ## Use with mlx
36
 
chat_template.json ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ {
2
+ "chat_template": "{% for message in messages %}{{message['role'].capitalize()}}{% if message['content'][0]['type'] == 'image' %}{{':'}}{% else %}{{': '}}{% endif %}{% for line in message['content'] %}{% if line['type'] == 'text' %}{{line['text']}}{% elif line['type'] == 'image' %}{{ '<image>' }}{% endif %}{% endfor %}<end_of_utterance>\n{% endfor %}{% if add_generation_prompt %}{{ 'Assistant:' }}{% endif %}"
3
+ }
config.json CHANGED
@@ -25,7 +25,7 @@
25
  "vision_config": {
26
  "hidden_size": 1152,
27
  "image_size": 980,
28
- "intermediate_size": 4304,
29
  "model_type": "idefics2",
30
  "num_attention_heads": 16,
31
  "num_hidden_layers": 27,
 
25
  "vision_config": {
26
  "hidden_size": 1152,
27
  "image_size": 980,
28
+ "intermediate_size": 4352,
29
  "model_type": "idefics2",
30
  "num_attention_heads": 16,
31
  "num_hidden_layers": 27,
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:297bde4d6e362ea35ecce58162be734eaac82c8c387a018b67fefd4af3aa1efa
3
- size 5313954350
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9371e2cc90318f4a310f3ecdab5c5f430e8149a3960d82f6056cafe77c4f484b
3
+ size 5346726218
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f0b57c58ce9ec64578d6e7995ed0e14f4ddd6a9ce75d952e46074eac2857acfa
3
- size 4129248613
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:cecc2b5dee766e3b757ad4c4e3347b6683d7fc85d00766f13308515cb0752fea
3
+ size 3603348985
model.safetensors.index.json CHANGED
@@ -1,6 +1,6 @@
1
  {
2
  "metadata": {
3
- "total_size": 9443046624
4
  },
5
  "weight_map": {
6
  "connector.modality_projection.down_proj.biases": "model-00002-of-00002.safetensors",
@@ -86,6 +86,8 @@
86
  "connector.perceiver_resampler.layers.2.self_attn.v_proj.scales": "model-00002-of-00002.safetensors",
87
  "connector.perceiver_resampler.layers.2.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
88
  "connector.perceiver_resampler.norm.weight": "model-00002-of-00002.safetensors",
 
 
89
  "language_model.embed_tokens.weight": "model-00001-of-00002.safetensors",
90
  "language_model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
91
  "language_model.layers.0.mlp.down_proj.biases": "model-00001-of-00002.safetensors",
@@ -317,17 +319,17 @@
317
  "language_model.layers.17.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
318
  "language_model.layers.17.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
319
  "language_model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
320
- "language_model.layers.18.input_layernorm.weight": "model-00002-of-00002.safetensors",
321
- "language_model.layers.18.mlp.down_proj.biases": "model-00002-of-00002.safetensors",
322
- "language_model.layers.18.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
323
- "language_model.layers.18.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
324
- "language_model.layers.18.mlp.gate_proj.biases": "model-00002-of-00002.safetensors",
325
- "language_model.layers.18.mlp.gate_proj.scales": "model-00002-of-00002.safetensors",
326
- "language_model.layers.18.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
327
- "language_model.layers.18.mlp.up_proj.biases": "model-00002-of-00002.safetensors",
328
- "language_model.layers.18.mlp.up_proj.scales": "model-00002-of-00002.safetensors",
329
- "language_model.layers.18.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
330
- "language_model.layers.18.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
331
  "language_model.layers.18.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
332
  "language_model.layers.18.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
333
  "language_model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
@@ -340,29 +342,29 @@
340
  "language_model.layers.18.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
341
  "language_model.layers.18.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
342
  "language_model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
343
- "language_model.layers.19.input_layernorm.weight": "model-00002-of-00002.safetensors",
344
- "language_model.layers.19.mlp.down_proj.biases": "model-00002-of-00002.safetensors",
345
- "language_model.layers.19.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
346
- "language_model.layers.19.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
347
- "language_model.layers.19.mlp.gate_proj.biases": "model-00002-of-00002.safetensors",
348
- "language_model.layers.19.mlp.gate_proj.scales": "model-00002-of-00002.safetensors",
349
- "language_model.layers.19.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
350
- "language_model.layers.19.mlp.up_proj.biases": "model-00002-of-00002.safetensors",
351
- "language_model.layers.19.mlp.up_proj.scales": "model-00002-of-00002.safetensors",
352
- "language_model.layers.19.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
353
- "language_model.layers.19.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
354
- "language_model.layers.19.self_attn.k_proj.biases": "model-00002-of-00002.safetensors",
355
- "language_model.layers.19.self_attn.k_proj.scales": "model-00002-of-00002.safetensors",
356
- "language_model.layers.19.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
357
- "language_model.layers.19.self_attn.o_proj.biases": "model-00002-of-00002.safetensors",
358
- "language_model.layers.19.self_attn.o_proj.scales": "model-00002-of-00002.safetensors",
359
- "language_model.layers.19.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
360
- "language_model.layers.19.self_attn.q_proj.biases": "model-00002-of-00002.safetensors",
361
- "language_model.layers.19.self_attn.q_proj.scales": "model-00002-of-00002.safetensors",
362
- "language_model.layers.19.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
363
- "language_model.layers.19.self_attn.v_proj.biases": "model-00002-of-00002.safetensors",
364
- "language_model.layers.19.self_attn.v_proj.scales": "model-00002-of-00002.safetensors",
365
- "language_model.layers.19.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
366
  "language_model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
367
  "language_model.layers.2.mlp.down_proj.biases": "model-00001-of-00002.safetensors",
368
  "language_model.layers.2.mlp.down_proj.scales": "model-00001-of-00002.safetensors",
@@ -390,25 +392,25 @@
390
  "language_model.layers.20.mlp.down_proj.biases": "model-00002-of-00002.safetensors",
391
  "language_model.layers.20.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
392
  "language_model.layers.20.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
393
- "language_model.layers.20.mlp.gate_proj.biases": "model-00002-of-00002.safetensors",
394
- "language_model.layers.20.mlp.gate_proj.scales": "model-00002-of-00002.safetensors",
395
- "language_model.layers.20.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
396
  "language_model.layers.20.mlp.up_proj.biases": "model-00002-of-00002.safetensors",
397
  "language_model.layers.20.mlp.up_proj.scales": "model-00002-of-00002.safetensors",
398
  "language_model.layers.20.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
399
  "language_model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
400
- "language_model.layers.20.self_attn.k_proj.biases": "model-00002-of-00002.safetensors",
401
- "language_model.layers.20.self_attn.k_proj.scales": "model-00002-of-00002.safetensors",
402
- "language_model.layers.20.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
403
- "language_model.layers.20.self_attn.o_proj.biases": "model-00002-of-00002.safetensors",
404
- "language_model.layers.20.self_attn.o_proj.scales": "model-00002-of-00002.safetensors",
405
- "language_model.layers.20.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
406
- "language_model.layers.20.self_attn.q_proj.biases": "model-00002-of-00002.safetensors",
407
- "language_model.layers.20.self_attn.q_proj.scales": "model-00002-of-00002.safetensors",
408
- "language_model.layers.20.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
409
- "language_model.layers.20.self_attn.v_proj.biases": "model-00002-of-00002.safetensors",
410
- "language_model.layers.20.self_attn.v_proj.scales": "model-00002-of-00002.safetensors",
411
- "language_model.layers.20.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
412
  "language_model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors",
413
  "language_model.layers.21.mlp.down_proj.biases": "model-00002-of-00002.safetensors",
414
  "language_model.layers.21.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
@@ -829,438 +831,764 @@
829
  "language_model.norm.weight": "model-00002-of-00002.safetensors",
830
  "vision_model.embeddings.patch_embedding.bias": "model-00001-of-00002.safetensors",
831
  "vision_model.embeddings.patch_embedding.weight": "model-00001-of-00002.safetensors",
 
 
832
  "vision_model.embeddings.position_embedding.weight": "model-00001-of-00002.safetensors",
833
  "vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00002.safetensors",
834
  "vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00002.safetensors",
835
  "vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00002.safetensors",
836
  "vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00002.safetensors",
837
  "vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
838
  "vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00002.safetensors",
839
  "vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
840
  "vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00002.safetensors",
841
  "vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
842
  "vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
843
  "vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
844
  "vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
845
  "vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
846
  "vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
847
  "vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
848
  "vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
849
  "vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00002.safetensors",
850
  "vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00002.safetensors",
851
  "vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00002.safetensors",
852
  "vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00002.safetensors",
853
  "vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
854
  "vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00002.safetensors",
855
  "vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
856
  "vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00002.safetensors",
857
  "vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
858
  "vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
859
  "vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
860
  "vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
861
  "vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
862
  "vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
863
  "vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
864
  "vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
865
  "vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00002.safetensors",
866
  "vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00002.safetensors",
867
  "vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00002.safetensors",
868
  "vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00002.safetensors",
869
  "vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
870
  "vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00002.safetensors",
871
  "vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
872
  "vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00002.safetensors",
873
  "vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
874
  "vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
875
  "vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
876
  "vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
877
  "vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
878
  "vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
879
  "vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
880
  "vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
881
  "vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00002.safetensors",
882
  "vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00002.safetensors",
883
  "vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00002.safetensors",
884
  "vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00002.safetensors",
885
  "vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
886
  "vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00002.safetensors",
887
  "vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
888
  "vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00002.safetensors",
889
  "vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
890
  "vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
891
  "vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
892
  "vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
893
  "vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
894
  "vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
895
  "vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
896
  "vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
897
  "vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00002.safetensors",
898
  "vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00002.safetensors",
899
  "vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00002.safetensors",
900
  "vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00002.safetensors",
901
  "vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
902
  "vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00002.safetensors",
903
  "vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
904
  "vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00002.safetensors",
905
  "vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
906
  "vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
907
  "vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
908
  "vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
909
  "vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
910
  "vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
911
  "vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
912
  "vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
913
  "vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00002.safetensors",
914
  "vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00002.safetensors",
915
  "vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00002.safetensors",
916
  "vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00002.safetensors",
917
  "vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
918
  "vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00002.safetensors",
919
  "vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
920
  "vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00002.safetensors",
921
  "vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
922
  "vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
923
  "vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
924
  "vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
925
  "vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
926
  "vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
927
  "vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
928
  "vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
929
  "vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00002.safetensors",
930
  "vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00002.safetensors",
931
  "vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00002.safetensors",
932
  "vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00002.safetensors",
933
  "vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
934
  "vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00002.safetensors",
935
  "vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
936
  "vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00002.safetensors",
937
  "vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
938
  "vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
939
  "vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
940
  "vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
941
  "vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
942
  "vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
943
  "vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
944
  "vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
945
  "vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00002.safetensors",
946
  "vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00002.safetensors",
947
  "vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00002.safetensors",
948
  "vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00002.safetensors",
949
  "vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
950
  "vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00002.safetensors",
951
  "vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
952
  "vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00002.safetensors",
953
  "vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
954
  "vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
955
  "vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
956
  "vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
957
  "vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
958
  "vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
959
  "vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
960
  "vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
961
  "vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00002.safetensors",
962
  "vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00002.safetensors",
963
  "vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00002.safetensors",
964
  "vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00002.safetensors",
965
  "vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
966
  "vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00002.safetensors",
967
  "vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
968
  "vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00002.safetensors",
969
  "vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
970
  "vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
971
  "vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
972
  "vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
973
  "vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
974
  "vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
975
  "vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
976
  "vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
977
  "vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00002.safetensors",
978
  "vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00002.safetensors",
979
  "vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00002.safetensors",
980
  "vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00002.safetensors",
981
  "vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
982
  "vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00002.safetensors",
983
  "vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
984
  "vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00002.safetensors",
985
  "vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
986
  "vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
987
  "vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
988
  "vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
989
  "vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
990
  "vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
991
  "vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
992
  "vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
993
  "vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00002.safetensors",
994
  "vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00002.safetensors",
995
  "vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00002.safetensors",
996
  "vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00002.safetensors",
997
  "vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
998
  "vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00002.safetensors",
999
  "vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1000
  "vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1001
  "vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1002
  "vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1003
  "vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1004
  "vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1005
  "vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1006
  "vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1007
  "vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1008
  "vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1009
  "vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00002.safetensors",
1010
  "vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00002.safetensors",
1011
  "vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00002.safetensors",
1012
  "vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00002.safetensors",
1013
  "vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1014
  "vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1015
  "vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1016
  "vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1017
  "vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1018
  "vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1019
  "vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1020
  "vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1021
  "vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1022
  "vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1023
  "vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1024
  "vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1025
  "vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00002.safetensors",
1026
  "vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00002.safetensors",
1027
  "vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00002.safetensors",
1028
  "vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00002.safetensors",
1029
  "vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1030
  "vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1031
  "vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1032
  "vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1033
  "vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1034
  "vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1035
  "vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1036
  "vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1037
  "vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1038
  "vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1039
  "vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1040
  "vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1041
  "vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00002.safetensors",
1042
  "vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00002.safetensors",
1043
  "vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00002.safetensors",
1044
  "vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00002.safetensors",
1045
  "vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1046
  "vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1047
  "vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1048
  "vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1049
  "vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1050
  "vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1051
  "vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1052
  "vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1053
  "vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1054
  "vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1055
  "vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1056
  "vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1057
  "vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00002.safetensors",
1058
  "vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00002.safetensors",
1059
  "vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00002.safetensors",
1060
  "vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00002.safetensors",
1061
  "vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1062
  "vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1063
  "vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1064
  "vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1065
  "vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1066
  "vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1067
  "vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1068
  "vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1069
  "vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1070
  "vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1071
  "vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1072
  "vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1073
  "vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00002.safetensors",
1074
  "vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00002.safetensors",
1075
  "vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00002.safetensors",
1076
  "vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00002.safetensors",
1077
  "vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1078
  "vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1079
  "vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1080
  "vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1081
  "vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1082
  "vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1083
  "vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1084
  "vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1085
  "vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1086
  "vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1087
  "vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1088
  "vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1089
  "vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00002.safetensors",
1090
  "vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00002.safetensors",
1091
  "vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00002.safetensors",
1092
  "vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00002.safetensors",
1093
  "vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1094
  "vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1095
  "vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1096
  "vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1097
  "vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1098
  "vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1099
  "vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1100
  "vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1101
  "vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1102
  "vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1103
  "vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1104
  "vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1105
  "vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00002.safetensors",
1106
  "vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00002.safetensors",
1107
  "vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00002.safetensors",
1108
  "vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00002.safetensors",
1109
  "vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1110
  "vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1111
  "vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1112
  "vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1113
  "vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1114
  "vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1115
  "vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1116
  "vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1117
  "vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1118
  "vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1119
  "vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1120
  "vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1121
  "vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00002.safetensors",
1122
  "vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00002.safetensors",
1123
  "vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00002.safetensors",
1124
  "vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00002.safetensors",
1125
  "vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1126
  "vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1127
  "vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1128
  "vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1129
  "vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1130
  "vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1131
  "vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1132
  "vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1133
  "vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1134
  "vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1135
  "vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1136
  "vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1137
  "vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00002.safetensors",
1138
  "vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00002.safetensors",
1139
  "vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00002.safetensors",
1140
  "vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00002.safetensors",
1141
  "vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1142
  "vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1143
  "vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1144
  "vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1145
  "vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1146
  "vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1147
  "vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1148
  "vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1149
  "vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1150
  "vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1151
  "vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1152
  "vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1153
  "vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00002.safetensors",
1154
  "vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00002.safetensors",
1155
  "vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00002.safetensors",
1156
  "vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00002.safetensors",
1157
  "vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1158
  "vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1159
  "vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1160
  "vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1161
  "vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1162
  "vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1163
  "vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1164
  "vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1165
  "vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1166
  "vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1167
  "vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1168
  "vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1169
  "vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00002.safetensors",
1170
  "vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00002.safetensors",
1171
  "vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00002.safetensors",
1172
  "vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00002.safetensors",
1173
  "vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1174
  "vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1175
  "vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1176
  "vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1177
  "vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1178
  "vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1179
  "vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1180
  "vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1181
  "vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1182
  "vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1183
  "vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1184
  "vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1185
  "vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00002.safetensors",
1186
  "vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00002.safetensors",
1187
  "vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00002.safetensors",
1188
  "vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00002.safetensors",
1189
  "vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1190
  "vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1191
  "vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1192
  "vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1193
  "vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1194
  "vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1195
  "vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1196
  "vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1197
  "vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1198
  "vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1199
  "vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1200
  "vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1201
  "vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00002.safetensors",
1202
  "vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00002.safetensors",
1203
  "vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00002.safetensors",
1204
  "vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00002.safetensors",
1205
  "vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1206
  "vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1207
  "vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1208
  "vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1209
  "vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1210
  "vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1211
  "vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1212
  "vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1213
  "vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1214
  "vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1215
  "vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1216
  "vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1217
  "vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00002.safetensors",
1218
  "vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00002.safetensors",
1219
  "vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00002.safetensors",
1220
  "vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00002.safetensors",
1221
  "vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1222
  "vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1223
  "vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1224
  "vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1225
  "vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1226
  "vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1227
  "vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1228
  "vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1229
  "vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1230
  "vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1231
  "vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1232
  "vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1233
  "vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00002.safetensors",
1234
  "vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00002.safetensors",
1235
  "vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00002.safetensors",
1236
  "vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00002.safetensors",
1237
  "vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1238
  "vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1239
  "vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1240
  "vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1241
  "vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1242
  "vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1243
  "vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1244
  "vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1245
  "vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1246
  "vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1247
  "vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1248
  "vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1249
  "vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00002.safetensors",
1250
  "vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00002.safetensors",
1251
  "vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00002.safetensors",
1252
  "vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00002.safetensors",
1253
  "vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00002.safetensors",
 
 
1254
  "vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1255
  "vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00002.safetensors",
 
 
1256
  "vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1257
  "vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
 
 
1258
  "vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1259
  "vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
 
 
1260
  "vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1261
  "vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
 
 
1262
  "vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1263
  "vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
 
 
1264
  "vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1265
  "vision_model.post_layernorm.bias": "model-00001-of-00002.safetensors",
1266
  "vision_model.post_layernorm.weight": "model-00001-of-00002.safetensors"
 
1
  {
2
  "metadata": {
3
+ "total_size": 8949877536
4
  },
5
  "weight_map": {
6
  "connector.modality_projection.down_proj.biases": "model-00002-of-00002.safetensors",
 
86
  "connector.perceiver_resampler.layers.2.self_attn.v_proj.scales": "model-00002-of-00002.safetensors",
87
  "connector.perceiver_resampler.layers.2.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
88
  "connector.perceiver_resampler.norm.weight": "model-00002-of-00002.safetensors",
89
+ "language_model.embed_tokens.biases": "model-00001-of-00002.safetensors",
90
+ "language_model.embed_tokens.scales": "model-00001-of-00002.safetensors",
91
  "language_model.embed_tokens.weight": "model-00001-of-00002.safetensors",
92
  "language_model.layers.0.input_layernorm.weight": "model-00001-of-00002.safetensors",
93
  "language_model.layers.0.mlp.down_proj.biases": "model-00001-of-00002.safetensors",
 
319
  "language_model.layers.17.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
320
  "language_model.layers.17.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
321
  "language_model.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
322
+ "language_model.layers.18.input_layernorm.weight": "model-00001-of-00002.safetensors",
323
+ "language_model.layers.18.mlp.down_proj.biases": "model-00001-of-00002.safetensors",
324
+ "language_model.layers.18.mlp.down_proj.scales": "model-00001-of-00002.safetensors",
325
+ "language_model.layers.18.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
326
+ "language_model.layers.18.mlp.gate_proj.biases": "model-00001-of-00002.safetensors",
327
+ "language_model.layers.18.mlp.gate_proj.scales": "model-00001-of-00002.safetensors",
328
+ "language_model.layers.18.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
329
+ "language_model.layers.18.mlp.up_proj.biases": "model-00001-of-00002.safetensors",
330
+ "language_model.layers.18.mlp.up_proj.scales": "model-00001-of-00002.safetensors",
331
+ "language_model.layers.18.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
332
+ "language_model.layers.18.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
333
  "language_model.layers.18.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
334
  "language_model.layers.18.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
335
  "language_model.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
 
342
  "language_model.layers.18.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
343
  "language_model.layers.18.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
344
  "language_model.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
345
+ "language_model.layers.19.input_layernorm.weight": "model-00001-of-00002.safetensors",
346
+ "language_model.layers.19.mlp.down_proj.biases": "model-00001-of-00002.safetensors",
347
+ "language_model.layers.19.mlp.down_proj.scales": "model-00001-of-00002.safetensors",
348
+ "language_model.layers.19.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
349
+ "language_model.layers.19.mlp.gate_proj.biases": "model-00001-of-00002.safetensors",
350
+ "language_model.layers.19.mlp.gate_proj.scales": "model-00001-of-00002.safetensors",
351
+ "language_model.layers.19.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
352
+ "language_model.layers.19.mlp.up_proj.biases": "model-00001-of-00002.safetensors",
353
+ "language_model.layers.19.mlp.up_proj.scales": "model-00001-of-00002.safetensors",
354
+ "language_model.layers.19.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
355
+ "language_model.layers.19.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
356
+ "language_model.layers.19.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
357
+ "language_model.layers.19.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
358
+ "language_model.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
359
+ "language_model.layers.19.self_attn.o_proj.biases": "model-00001-of-00002.safetensors",
360
+ "language_model.layers.19.self_attn.o_proj.scales": "model-00001-of-00002.safetensors",
361
+ "language_model.layers.19.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
362
+ "language_model.layers.19.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
363
+ "language_model.layers.19.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
364
+ "language_model.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
365
+ "language_model.layers.19.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
366
+ "language_model.layers.19.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
367
+ "language_model.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
368
  "language_model.layers.2.input_layernorm.weight": "model-00001-of-00002.safetensors",
369
  "language_model.layers.2.mlp.down_proj.biases": "model-00001-of-00002.safetensors",
370
  "language_model.layers.2.mlp.down_proj.scales": "model-00001-of-00002.safetensors",
 
392
  "language_model.layers.20.mlp.down_proj.biases": "model-00002-of-00002.safetensors",
393
  "language_model.layers.20.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
394
  "language_model.layers.20.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
395
+ "language_model.layers.20.mlp.gate_proj.biases": "model-00001-of-00002.safetensors",
396
+ "language_model.layers.20.mlp.gate_proj.scales": "model-00001-of-00002.safetensors",
397
+ "language_model.layers.20.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
398
  "language_model.layers.20.mlp.up_proj.biases": "model-00002-of-00002.safetensors",
399
  "language_model.layers.20.mlp.up_proj.scales": "model-00002-of-00002.safetensors",
400
  "language_model.layers.20.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
401
  "language_model.layers.20.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
402
+ "language_model.layers.20.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
403
+ "language_model.layers.20.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
404
+ "language_model.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
405
+ "language_model.layers.20.self_attn.o_proj.biases": "model-00001-of-00002.safetensors",
406
+ "language_model.layers.20.self_attn.o_proj.scales": "model-00001-of-00002.safetensors",
407
+ "language_model.layers.20.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
408
+ "language_model.layers.20.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
409
+ "language_model.layers.20.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
410
+ "language_model.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
411
+ "language_model.layers.20.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
412
+ "language_model.layers.20.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
413
+ "language_model.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
414
  "language_model.layers.21.input_layernorm.weight": "model-00002-of-00002.safetensors",
415
  "language_model.layers.21.mlp.down_proj.biases": "model-00002-of-00002.safetensors",
416
  "language_model.layers.21.mlp.down_proj.scales": "model-00002-of-00002.safetensors",
 
831
  "language_model.norm.weight": "model-00002-of-00002.safetensors",
832
  "vision_model.embeddings.patch_embedding.bias": "model-00001-of-00002.safetensors",
833
  "vision_model.embeddings.patch_embedding.weight": "model-00001-of-00002.safetensors",
834
+ "vision_model.embeddings.position_embedding.biases": "model-00001-of-00002.safetensors",
835
+ "vision_model.embeddings.position_embedding.scales": "model-00001-of-00002.safetensors",
836
  "vision_model.embeddings.position_embedding.weight": "model-00001-of-00002.safetensors",
837
  "vision_model.encoder.layers.0.layer_norm1.bias": "model-00001-of-00002.safetensors",
838
  "vision_model.encoder.layers.0.layer_norm1.weight": "model-00001-of-00002.safetensors",
839
  "vision_model.encoder.layers.0.layer_norm2.bias": "model-00001-of-00002.safetensors",
840
  "vision_model.encoder.layers.0.layer_norm2.weight": "model-00001-of-00002.safetensors",
841
  "vision_model.encoder.layers.0.mlp.fc1.bias": "model-00001-of-00002.safetensors",
842
+ "vision_model.encoder.layers.0.mlp.fc1.biases": "model-00001-of-00002.safetensors",
843
+ "vision_model.encoder.layers.0.mlp.fc1.scales": "model-00001-of-00002.safetensors",
844
  "vision_model.encoder.layers.0.mlp.fc1.weight": "model-00001-of-00002.safetensors",
845
  "vision_model.encoder.layers.0.mlp.fc2.bias": "model-00001-of-00002.safetensors",
846
+ "vision_model.encoder.layers.0.mlp.fc2.biases": "model-00001-of-00002.safetensors",
847
+ "vision_model.encoder.layers.0.mlp.fc2.scales": "model-00001-of-00002.safetensors",
848
  "vision_model.encoder.layers.0.mlp.fc2.weight": "model-00001-of-00002.safetensors",
849
  "vision_model.encoder.layers.0.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
850
+ "vision_model.encoder.layers.0.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
851
+ "vision_model.encoder.layers.0.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
852
  "vision_model.encoder.layers.0.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
853
  "vision_model.encoder.layers.0.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
854
+ "vision_model.encoder.layers.0.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
855
+ "vision_model.encoder.layers.0.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
856
  "vision_model.encoder.layers.0.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
857
  "vision_model.encoder.layers.0.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
858
+ "vision_model.encoder.layers.0.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
859
+ "vision_model.encoder.layers.0.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
860
  "vision_model.encoder.layers.0.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
861
  "vision_model.encoder.layers.0.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
862
+ "vision_model.encoder.layers.0.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
863
+ "vision_model.encoder.layers.0.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
864
  "vision_model.encoder.layers.0.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
865
  "vision_model.encoder.layers.1.layer_norm1.bias": "model-00001-of-00002.safetensors",
866
  "vision_model.encoder.layers.1.layer_norm1.weight": "model-00001-of-00002.safetensors",
867
  "vision_model.encoder.layers.1.layer_norm2.bias": "model-00001-of-00002.safetensors",
868
  "vision_model.encoder.layers.1.layer_norm2.weight": "model-00001-of-00002.safetensors",
869
  "vision_model.encoder.layers.1.mlp.fc1.bias": "model-00001-of-00002.safetensors",
870
+ "vision_model.encoder.layers.1.mlp.fc1.biases": "model-00001-of-00002.safetensors",
871
+ "vision_model.encoder.layers.1.mlp.fc1.scales": "model-00001-of-00002.safetensors",
872
  "vision_model.encoder.layers.1.mlp.fc1.weight": "model-00001-of-00002.safetensors",
873
  "vision_model.encoder.layers.1.mlp.fc2.bias": "model-00001-of-00002.safetensors",
874
+ "vision_model.encoder.layers.1.mlp.fc2.biases": "model-00001-of-00002.safetensors",
875
+ "vision_model.encoder.layers.1.mlp.fc2.scales": "model-00001-of-00002.safetensors",
876
  "vision_model.encoder.layers.1.mlp.fc2.weight": "model-00001-of-00002.safetensors",
877
  "vision_model.encoder.layers.1.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
878
+ "vision_model.encoder.layers.1.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
879
+ "vision_model.encoder.layers.1.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
880
  "vision_model.encoder.layers.1.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
881
  "vision_model.encoder.layers.1.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
882
+ "vision_model.encoder.layers.1.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
883
+ "vision_model.encoder.layers.1.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
884
  "vision_model.encoder.layers.1.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
885
  "vision_model.encoder.layers.1.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
886
+ "vision_model.encoder.layers.1.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
887
+ "vision_model.encoder.layers.1.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
888
  "vision_model.encoder.layers.1.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
889
  "vision_model.encoder.layers.1.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
890
+ "vision_model.encoder.layers.1.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
891
+ "vision_model.encoder.layers.1.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
892
  "vision_model.encoder.layers.1.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
893
  "vision_model.encoder.layers.10.layer_norm1.bias": "model-00001-of-00002.safetensors",
894
  "vision_model.encoder.layers.10.layer_norm1.weight": "model-00001-of-00002.safetensors",
895
  "vision_model.encoder.layers.10.layer_norm2.bias": "model-00001-of-00002.safetensors",
896
  "vision_model.encoder.layers.10.layer_norm2.weight": "model-00001-of-00002.safetensors",
897
  "vision_model.encoder.layers.10.mlp.fc1.bias": "model-00001-of-00002.safetensors",
898
+ "vision_model.encoder.layers.10.mlp.fc1.biases": "model-00001-of-00002.safetensors",
899
+ "vision_model.encoder.layers.10.mlp.fc1.scales": "model-00001-of-00002.safetensors",
900
  "vision_model.encoder.layers.10.mlp.fc1.weight": "model-00001-of-00002.safetensors",
901
  "vision_model.encoder.layers.10.mlp.fc2.bias": "model-00001-of-00002.safetensors",
902
+ "vision_model.encoder.layers.10.mlp.fc2.biases": "model-00001-of-00002.safetensors",
903
+ "vision_model.encoder.layers.10.mlp.fc2.scales": "model-00001-of-00002.safetensors",
904
  "vision_model.encoder.layers.10.mlp.fc2.weight": "model-00001-of-00002.safetensors",
905
  "vision_model.encoder.layers.10.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
906
+ "vision_model.encoder.layers.10.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
907
+ "vision_model.encoder.layers.10.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
908
  "vision_model.encoder.layers.10.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
909
  "vision_model.encoder.layers.10.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
910
+ "vision_model.encoder.layers.10.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
911
+ "vision_model.encoder.layers.10.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
912
  "vision_model.encoder.layers.10.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
913
  "vision_model.encoder.layers.10.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
914
+ "vision_model.encoder.layers.10.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
915
+ "vision_model.encoder.layers.10.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
916
  "vision_model.encoder.layers.10.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
917
  "vision_model.encoder.layers.10.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
918
+ "vision_model.encoder.layers.10.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
919
+ "vision_model.encoder.layers.10.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
920
  "vision_model.encoder.layers.10.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
921
  "vision_model.encoder.layers.11.layer_norm1.bias": "model-00001-of-00002.safetensors",
922
  "vision_model.encoder.layers.11.layer_norm1.weight": "model-00001-of-00002.safetensors",
923
  "vision_model.encoder.layers.11.layer_norm2.bias": "model-00001-of-00002.safetensors",
924
  "vision_model.encoder.layers.11.layer_norm2.weight": "model-00001-of-00002.safetensors",
925
  "vision_model.encoder.layers.11.mlp.fc1.bias": "model-00001-of-00002.safetensors",
926
+ "vision_model.encoder.layers.11.mlp.fc1.biases": "model-00001-of-00002.safetensors",
927
+ "vision_model.encoder.layers.11.mlp.fc1.scales": "model-00001-of-00002.safetensors",
928
  "vision_model.encoder.layers.11.mlp.fc1.weight": "model-00001-of-00002.safetensors",
929
  "vision_model.encoder.layers.11.mlp.fc2.bias": "model-00001-of-00002.safetensors",
930
+ "vision_model.encoder.layers.11.mlp.fc2.biases": "model-00001-of-00002.safetensors",
931
+ "vision_model.encoder.layers.11.mlp.fc2.scales": "model-00001-of-00002.safetensors",
932
  "vision_model.encoder.layers.11.mlp.fc2.weight": "model-00001-of-00002.safetensors",
933
  "vision_model.encoder.layers.11.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
934
+ "vision_model.encoder.layers.11.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
935
+ "vision_model.encoder.layers.11.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
936
  "vision_model.encoder.layers.11.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
937
  "vision_model.encoder.layers.11.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
938
+ "vision_model.encoder.layers.11.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
939
+ "vision_model.encoder.layers.11.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
940
  "vision_model.encoder.layers.11.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
941
  "vision_model.encoder.layers.11.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
942
+ "vision_model.encoder.layers.11.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
943
+ "vision_model.encoder.layers.11.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
944
  "vision_model.encoder.layers.11.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
945
  "vision_model.encoder.layers.11.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
946
+ "vision_model.encoder.layers.11.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
947
+ "vision_model.encoder.layers.11.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
948
  "vision_model.encoder.layers.11.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
949
  "vision_model.encoder.layers.12.layer_norm1.bias": "model-00001-of-00002.safetensors",
950
  "vision_model.encoder.layers.12.layer_norm1.weight": "model-00001-of-00002.safetensors",
951
  "vision_model.encoder.layers.12.layer_norm2.bias": "model-00001-of-00002.safetensors",
952
  "vision_model.encoder.layers.12.layer_norm2.weight": "model-00001-of-00002.safetensors",
953
  "vision_model.encoder.layers.12.mlp.fc1.bias": "model-00001-of-00002.safetensors",
954
+ "vision_model.encoder.layers.12.mlp.fc1.biases": "model-00001-of-00002.safetensors",
955
+ "vision_model.encoder.layers.12.mlp.fc1.scales": "model-00001-of-00002.safetensors",
956
  "vision_model.encoder.layers.12.mlp.fc1.weight": "model-00001-of-00002.safetensors",
957
  "vision_model.encoder.layers.12.mlp.fc2.bias": "model-00001-of-00002.safetensors",
958
+ "vision_model.encoder.layers.12.mlp.fc2.biases": "model-00001-of-00002.safetensors",
959
+ "vision_model.encoder.layers.12.mlp.fc2.scales": "model-00001-of-00002.safetensors",
960
  "vision_model.encoder.layers.12.mlp.fc2.weight": "model-00001-of-00002.safetensors",
961
  "vision_model.encoder.layers.12.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
962
+ "vision_model.encoder.layers.12.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
963
+ "vision_model.encoder.layers.12.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
964
  "vision_model.encoder.layers.12.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
965
  "vision_model.encoder.layers.12.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
966
+ "vision_model.encoder.layers.12.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
967
+ "vision_model.encoder.layers.12.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
968
  "vision_model.encoder.layers.12.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
969
  "vision_model.encoder.layers.12.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
970
+ "vision_model.encoder.layers.12.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
971
+ "vision_model.encoder.layers.12.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
972
  "vision_model.encoder.layers.12.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
973
  "vision_model.encoder.layers.12.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
974
+ "vision_model.encoder.layers.12.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
975
+ "vision_model.encoder.layers.12.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
976
  "vision_model.encoder.layers.12.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
977
  "vision_model.encoder.layers.13.layer_norm1.bias": "model-00001-of-00002.safetensors",
978
  "vision_model.encoder.layers.13.layer_norm1.weight": "model-00001-of-00002.safetensors",
979
  "vision_model.encoder.layers.13.layer_norm2.bias": "model-00001-of-00002.safetensors",
980
  "vision_model.encoder.layers.13.layer_norm2.weight": "model-00001-of-00002.safetensors",
981
  "vision_model.encoder.layers.13.mlp.fc1.bias": "model-00001-of-00002.safetensors",
982
+ "vision_model.encoder.layers.13.mlp.fc1.biases": "model-00001-of-00002.safetensors",
983
+ "vision_model.encoder.layers.13.mlp.fc1.scales": "model-00001-of-00002.safetensors",
984
  "vision_model.encoder.layers.13.mlp.fc1.weight": "model-00001-of-00002.safetensors",
985
  "vision_model.encoder.layers.13.mlp.fc2.bias": "model-00001-of-00002.safetensors",
986
+ "vision_model.encoder.layers.13.mlp.fc2.biases": "model-00001-of-00002.safetensors",
987
+ "vision_model.encoder.layers.13.mlp.fc2.scales": "model-00001-of-00002.safetensors",
988
  "vision_model.encoder.layers.13.mlp.fc2.weight": "model-00001-of-00002.safetensors",
989
  "vision_model.encoder.layers.13.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
990
+ "vision_model.encoder.layers.13.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
991
+ "vision_model.encoder.layers.13.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
992
  "vision_model.encoder.layers.13.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
993
  "vision_model.encoder.layers.13.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
994
+ "vision_model.encoder.layers.13.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
995
+ "vision_model.encoder.layers.13.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
996
  "vision_model.encoder.layers.13.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
997
  "vision_model.encoder.layers.13.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
998
+ "vision_model.encoder.layers.13.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
999
+ "vision_model.encoder.layers.13.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1000
  "vision_model.encoder.layers.13.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1001
  "vision_model.encoder.layers.13.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1002
+ "vision_model.encoder.layers.13.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1003
+ "vision_model.encoder.layers.13.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1004
  "vision_model.encoder.layers.13.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1005
  "vision_model.encoder.layers.14.layer_norm1.bias": "model-00001-of-00002.safetensors",
1006
  "vision_model.encoder.layers.14.layer_norm1.weight": "model-00001-of-00002.safetensors",
1007
  "vision_model.encoder.layers.14.layer_norm2.bias": "model-00001-of-00002.safetensors",
1008
  "vision_model.encoder.layers.14.layer_norm2.weight": "model-00001-of-00002.safetensors",
1009
  "vision_model.encoder.layers.14.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1010
+ "vision_model.encoder.layers.14.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1011
+ "vision_model.encoder.layers.14.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1012
  "vision_model.encoder.layers.14.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1013
  "vision_model.encoder.layers.14.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1014
+ "vision_model.encoder.layers.14.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1015
+ "vision_model.encoder.layers.14.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1016
  "vision_model.encoder.layers.14.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1017
  "vision_model.encoder.layers.14.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1018
+ "vision_model.encoder.layers.14.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1019
+ "vision_model.encoder.layers.14.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1020
  "vision_model.encoder.layers.14.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1021
  "vision_model.encoder.layers.14.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1022
+ "vision_model.encoder.layers.14.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1023
+ "vision_model.encoder.layers.14.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1024
  "vision_model.encoder.layers.14.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1025
  "vision_model.encoder.layers.14.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1026
+ "vision_model.encoder.layers.14.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1027
+ "vision_model.encoder.layers.14.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1028
  "vision_model.encoder.layers.14.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1029
  "vision_model.encoder.layers.14.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1030
+ "vision_model.encoder.layers.14.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1031
+ "vision_model.encoder.layers.14.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1032
  "vision_model.encoder.layers.14.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1033
  "vision_model.encoder.layers.15.layer_norm1.bias": "model-00001-of-00002.safetensors",
1034
  "vision_model.encoder.layers.15.layer_norm1.weight": "model-00001-of-00002.safetensors",
1035
  "vision_model.encoder.layers.15.layer_norm2.bias": "model-00001-of-00002.safetensors",
1036
  "vision_model.encoder.layers.15.layer_norm2.weight": "model-00001-of-00002.safetensors",
1037
  "vision_model.encoder.layers.15.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1038
+ "vision_model.encoder.layers.15.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1039
+ "vision_model.encoder.layers.15.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1040
  "vision_model.encoder.layers.15.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1041
  "vision_model.encoder.layers.15.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1042
+ "vision_model.encoder.layers.15.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1043
+ "vision_model.encoder.layers.15.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1044
  "vision_model.encoder.layers.15.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1045
  "vision_model.encoder.layers.15.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1046
+ "vision_model.encoder.layers.15.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1047
+ "vision_model.encoder.layers.15.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1048
  "vision_model.encoder.layers.15.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1049
  "vision_model.encoder.layers.15.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1050
+ "vision_model.encoder.layers.15.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1051
+ "vision_model.encoder.layers.15.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1052
  "vision_model.encoder.layers.15.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1053
  "vision_model.encoder.layers.15.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1054
+ "vision_model.encoder.layers.15.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1055
+ "vision_model.encoder.layers.15.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1056
  "vision_model.encoder.layers.15.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1057
  "vision_model.encoder.layers.15.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1058
+ "vision_model.encoder.layers.15.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1059
+ "vision_model.encoder.layers.15.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1060
  "vision_model.encoder.layers.15.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1061
  "vision_model.encoder.layers.16.layer_norm1.bias": "model-00001-of-00002.safetensors",
1062
  "vision_model.encoder.layers.16.layer_norm1.weight": "model-00001-of-00002.safetensors",
1063
  "vision_model.encoder.layers.16.layer_norm2.bias": "model-00001-of-00002.safetensors",
1064
  "vision_model.encoder.layers.16.layer_norm2.weight": "model-00001-of-00002.safetensors",
1065
  "vision_model.encoder.layers.16.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1066
+ "vision_model.encoder.layers.16.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1067
+ "vision_model.encoder.layers.16.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1068
  "vision_model.encoder.layers.16.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1069
  "vision_model.encoder.layers.16.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1070
+ "vision_model.encoder.layers.16.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1071
+ "vision_model.encoder.layers.16.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1072
  "vision_model.encoder.layers.16.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1073
  "vision_model.encoder.layers.16.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1074
+ "vision_model.encoder.layers.16.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1075
+ "vision_model.encoder.layers.16.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1076
  "vision_model.encoder.layers.16.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1077
  "vision_model.encoder.layers.16.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1078
+ "vision_model.encoder.layers.16.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1079
+ "vision_model.encoder.layers.16.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1080
  "vision_model.encoder.layers.16.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1081
  "vision_model.encoder.layers.16.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1082
+ "vision_model.encoder.layers.16.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1083
+ "vision_model.encoder.layers.16.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1084
  "vision_model.encoder.layers.16.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1085
  "vision_model.encoder.layers.16.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1086
+ "vision_model.encoder.layers.16.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1087
+ "vision_model.encoder.layers.16.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1088
  "vision_model.encoder.layers.16.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1089
  "vision_model.encoder.layers.17.layer_norm1.bias": "model-00001-of-00002.safetensors",
1090
  "vision_model.encoder.layers.17.layer_norm1.weight": "model-00001-of-00002.safetensors",
1091
  "vision_model.encoder.layers.17.layer_norm2.bias": "model-00001-of-00002.safetensors",
1092
  "vision_model.encoder.layers.17.layer_norm2.weight": "model-00001-of-00002.safetensors",
1093
  "vision_model.encoder.layers.17.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1094
+ "vision_model.encoder.layers.17.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1095
+ "vision_model.encoder.layers.17.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1096
  "vision_model.encoder.layers.17.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1097
  "vision_model.encoder.layers.17.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1098
+ "vision_model.encoder.layers.17.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1099
+ "vision_model.encoder.layers.17.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1100
  "vision_model.encoder.layers.17.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1101
  "vision_model.encoder.layers.17.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1102
+ "vision_model.encoder.layers.17.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1103
+ "vision_model.encoder.layers.17.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1104
  "vision_model.encoder.layers.17.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1105
  "vision_model.encoder.layers.17.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1106
+ "vision_model.encoder.layers.17.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1107
+ "vision_model.encoder.layers.17.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1108
  "vision_model.encoder.layers.17.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1109
  "vision_model.encoder.layers.17.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1110
+ "vision_model.encoder.layers.17.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1111
+ "vision_model.encoder.layers.17.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1112
  "vision_model.encoder.layers.17.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1113
  "vision_model.encoder.layers.17.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1114
+ "vision_model.encoder.layers.17.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1115
+ "vision_model.encoder.layers.17.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1116
  "vision_model.encoder.layers.17.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1117
  "vision_model.encoder.layers.18.layer_norm1.bias": "model-00001-of-00002.safetensors",
1118
  "vision_model.encoder.layers.18.layer_norm1.weight": "model-00001-of-00002.safetensors",
1119
  "vision_model.encoder.layers.18.layer_norm2.bias": "model-00001-of-00002.safetensors",
1120
  "vision_model.encoder.layers.18.layer_norm2.weight": "model-00001-of-00002.safetensors",
1121
  "vision_model.encoder.layers.18.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1122
+ "vision_model.encoder.layers.18.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1123
+ "vision_model.encoder.layers.18.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1124
  "vision_model.encoder.layers.18.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1125
  "vision_model.encoder.layers.18.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1126
+ "vision_model.encoder.layers.18.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1127
+ "vision_model.encoder.layers.18.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1128
  "vision_model.encoder.layers.18.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1129
  "vision_model.encoder.layers.18.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1130
+ "vision_model.encoder.layers.18.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1131
+ "vision_model.encoder.layers.18.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1132
  "vision_model.encoder.layers.18.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1133
  "vision_model.encoder.layers.18.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1134
+ "vision_model.encoder.layers.18.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1135
+ "vision_model.encoder.layers.18.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1136
  "vision_model.encoder.layers.18.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1137
  "vision_model.encoder.layers.18.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1138
+ "vision_model.encoder.layers.18.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1139
+ "vision_model.encoder.layers.18.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1140
  "vision_model.encoder.layers.18.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1141
  "vision_model.encoder.layers.18.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1142
+ "vision_model.encoder.layers.18.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1143
+ "vision_model.encoder.layers.18.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1144
  "vision_model.encoder.layers.18.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1145
  "vision_model.encoder.layers.19.layer_norm1.bias": "model-00001-of-00002.safetensors",
1146
  "vision_model.encoder.layers.19.layer_norm1.weight": "model-00001-of-00002.safetensors",
1147
  "vision_model.encoder.layers.19.layer_norm2.bias": "model-00001-of-00002.safetensors",
1148
  "vision_model.encoder.layers.19.layer_norm2.weight": "model-00001-of-00002.safetensors",
1149
  "vision_model.encoder.layers.19.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1150
+ "vision_model.encoder.layers.19.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1151
+ "vision_model.encoder.layers.19.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1152
  "vision_model.encoder.layers.19.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1153
  "vision_model.encoder.layers.19.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1154
+ "vision_model.encoder.layers.19.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1155
+ "vision_model.encoder.layers.19.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1156
  "vision_model.encoder.layers.19.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1157
  "vision_model.encoder.layers.19.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1158
+ "vision_model.encoder.layers.19.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1159
+ "vision_model.encoder.layers.19.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1160
  "vision_model.encoder.layers.19.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1161
  "vision_model.encoder.layers.19.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1162
+ "vision_model.encoder.layers.19.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1163
+ "vision_model.encoder.layers.19.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1164
  "vision_model.encoder.layers.19.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1165
  "vision_model.encoder.layers.19.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1166
+ "vision_model.encoder.layers.19.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1167
+ "vision_model.encoder.layers.19.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1168
  "vision_model.encoder.layers.19.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1169
  "vision_model.encoder.layers.19.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1170
+ "vision_model.encoder.layers.19.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1171
+ "vision_model.encoder.layers.19.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1172
  "vision_model.encoder.layers.19.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1173
  "vision_model.encoder.layers.2.layer_norm1.bias": "model-00001-of-00002.safetensors",
1174
  "vision_model.encoder.layers.2.layer_norm1.weight": "model-00001-of-00002.safetensors",
1175
  "vision_model.encoder.layers.2.layer_norm2.bias": "model-00001-of-00002.safetensors",
1176
  "vision_model.encoder.layers.2.layer_norm2.weight": "model-00001-of-00002.safetensors",
1177
  "vision_model.encoder.layers.2.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1178
+ "vision_model.encoder.layers.2.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1179
+ "vision_model.encoder.layers.2.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1180
  "vision_model.encoder.layers.2.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1181
  "vision_model.encoder.layers.2.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1182
+ "vision_model.encoder.layers.2.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1183
+ "vision_model.encoder.layers.2.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1184
  "vision_model.encoder.layers.2.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1185
  "vision_model.encoder.layers.2.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1186
+ "vision_model.encoder.layers.2.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1187
+ "vision_model.encoder.layers.2.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1188
  "vision_model.encoder.layers.2.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1189
  "vision_model.encoder.layers.2.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1190
+ "vision_model.encoder.layers.2.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1191
+ "vision_model.encoder.layers.2.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1192
  "vision_model.encoder.layers.2.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1193
  "vision_model.encoder.layers.2.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1194
+ "vision_model.encoder.layers.2.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1195
+ "vision_model.encoder.layers.2.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1196
  "vision_model.encoder.layers.2.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1197
  "vision_model.encoder.layers.2.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1198
+ "vision_model.encoder.layers.2.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1199
+ "vision_model.encoder.layers.2.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1200
  "vision_model.encoder.layers.2.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1201
  "vision_model.encoder.layers.20.layer_norm1.bias": "model-00001-of-00002.safetensors",
1202
  "vision_model.encoder.layers.20.layer_norm1.weight": "model-00001-of-00002.safetensors",
1203
  "vision_model.encoder.layers.20.layer_norm2.bias": "model-00001-of-00002.safetensors",
1204
  "vision_model.encoder.layers.20.layer_norm2.weight": "model-00001-of-00002.safetensors",
1205
  "vision_model.encoder.layers.20.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1206
+ "vision_model.encoder.layers.20.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1207
+ "vision_model.encoder.layers.20.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1208
  "vision_model.encoder.layers.20.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1209
  "vision_model.encoder.layers.20.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1210
+ "vision_model.encoder.layers.20.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1211
+ "vision_model.encoder.layers.20.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1212
  "vision_model.encoder.layers.20.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1213
  "vision_model.encoder.layers.20.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1214
+ "vision_model.encoder.layers.20.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1215
+ "vision_model.encoder.layers.20.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1216
  "vision_model.encoder.layers.20.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1217
  "vision_model.encoder.layers.20.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1218
+ "vision_model.encoder.layers.20.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1219
+ "vision_model.encoder.layers.20.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1220
  "vision_model.encoder.layers.20.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1221
  "vision_model.encoder.layers.20.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1222
+ "vision_model.encoder.layers.20.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1223
+ "vision_model.encoder.layers.20.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1224
  "vision_model.encoder.layers.20.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1225
  "vision_model.encoder.layers.20.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1226
+ "vision_model.encoder.layers.20.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1227
+ "vision_model.encoder.layers.20.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1228
  "vision_model.encoder.layers.20.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1229
  "vision_model.encoder.layers.21.layer_norm1.bias": "model-00001-of-00002.safetensors",
1230
  "vision_model.encoder.layers.21.layer_norm1.weight": "model-00001-of-00002.safetensors",
1231
  "vision_model.encoder.layers.21.layer_norm2.bias": "model-00001-of-00002.safetensors",
1232
  "vision_model.encoder.layers.21.layer_norm2.weight": "model-00001-of-00002.safetensors",
1233
  "vision_model.encoder.layers.21.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1234
+ "vision_model.encoder.layers.21.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1235
+ "vision_model.encoder.layers.21.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1236
  "vision_model.encoder.layers.21.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1237
  "vision_model.encoder.layers.21.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1238
+ "vision_model.encoder.layers.21.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1239
+ "vision_model.encoder.layers.21.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1240
  "vision_model.encoder.layers.21.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1241
  "vision_model.encoder.layers.21.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1242
+ "vision_model.encoder.layers.21.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1243
+ "vision_model.encoder.layers.21.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1244
  "vision_model.encoder.layers.21.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1245
  "vision_model.encoder.layers.21.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1246
+ "vision_model.encoder.layers.21.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1247
+ "vision_model.encoder.layers.21.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1248
  "vision_model.encoder.layers.21.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1249
  "vision_model.encoder.layers.21.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1250
+ "vision_model.encoder.layers.21.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1251
+ "vision_model.encoder.layers.21.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1252
  "vision_model.encoder.layers.21.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1253
  "vision_model.encoder.layers.21.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1254
+ "vision_model.encoder.layers.21.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1255
+ "vision_model.encoder.layers.21.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1256
  "vision_model.encoder.layers.21.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1257
  "vision_model.encoder.layers.22.layer_norm1.bias": "model-00001-of-00002.safetensors",
1258
  "vision_model.encoder.layers.22.layer_norm1.weight": "model-00001-of-00002.safetensors",
1259
  "vision_model.encoder.layers.22.layer_norm2.bias": "model-00001-of-00002.safetensors",
1260
  "vision_model.encoder.layers.22.layer_norm2.weight": "model-00001-of-00002.safetensors",
1261
  "vision_model.encoder.layers.22.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1262
+ "vision_model.encoder.layers.22.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1263
+ "vision_model.encoder.layers.22.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1264
  "vision_model.encoder.layers.22.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1265
  "vision_model.encoder.layers.22.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1266
+ "vision_model.encoder.layers.22.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1267
+ "vision_model.encoder.layers.22.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1268
  "vision_model.encoder.layers.22.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1269
  "vision_model.encoder.layers.22.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1270
+ "vision_model.encoder.layers.22.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1271
+ "vision_model.encoder.layers.22.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1272
  "vision_model.encoder.layers.22.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1273
  "vision_model.encoder.layers.22.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1274
+ "vision_model.encoder.layers.22.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1275
+ "vision_model.encoder.layers.22.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1276
  "vision_model.encoder.layers.22.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1277
  "vision_model.encoder.layers.22.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1278
+ "vision_model.encoder.layers.22.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1279
+ "vision_model.encoder.layers.22.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1280
  "vision_model.encoder.layers.22.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1281
  "vision_model.encoder.layers.22.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1282
+ "vision_model.encoder.layers.22.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1283
+ "vision_model.encoder.layers.22.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1284
  "vision_model.encoder.layers.22.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1285
  "vision_model.encoder.layers.23.layer_norm1.bias": "model-00001-of-00002.safetensors",
1286
  "vision_model.encoder.layers.23.layer_norm1.weight": "model-00001-of-00002.safetensors",
1287
  "vision_model.encoder.layers.23.layer_norm2.bias": "model-00001-of-00002.safetensors",
1288
  "vision_model.encoder.layers.23.layer_norm2.weight": "model-00001-of-00002.safetensors",
1289
  "vision_model.encoder.layers.23.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1290
+ "vision_model.encoder.layers.23.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1291
+ "vision_model.encoder.layers.23.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1292
  "vision_model.encoder.layers.23.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1293
  "vision_model.encoder.layers.23.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1294
+ "vision_model.encoder.layers.23.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1295
+ "vision_model.encoder.layers.23.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1296
  "vision_model.encoder.layers.23.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1297
  "vision_model.encoder.layers.23.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1298
+ "vision_model.encoder.layers.23.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1299
+ "vision_model.encoder.layers.23.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1300
  "vision_model.encoder.layers.23.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1301
  "vision_model.encoder.layers.23.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1302
+ "vision_model.encoder.layers.23.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1303
+ "vision_model.encoder.layers.23.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1304
  "vision_model.encoder.layers.23.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1305
  "vision_model.encoder.layers.23.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1306
+ "vision_model.encoder.layers.23.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1307
+ "vision_model.encoder.layers.23.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1308
  "vision_model.encoder.layers.23.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1309
  "vision_model.encoder.layers.23.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1310
+ "vision_model.encoder.layers.23.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1311
+ "vision_model.encoder.layers.23.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1312
  "vision_model.encoder.layers.23.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1313
  "vision_model.encoder.layers.24.layer_norm1.bias": "model-00001-of-00002.safetensors",
1314
  "vision_model.encoder.layers.24.layer_norm1.weight": "model-00001-of-00002.safetensors",
1315
  "vision_model.encoder.layers.24.layer_norm2.bias": "model-00001-of-00002.safetensors",
1316
  "vision_model.encoder.layers.24.layer_norm2.weight": "model-00001-of-00002.safetensors",
1317
  "vision_model.encoder.layers.24.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1318
+ "vision_model.encoder.layers.24.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1319
+ "vision_model.encoder.layers.24.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1320
  "vision_model.encoder.layers.24.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1321
  "vision_model.encoder.layers.24.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1322
+ "vision_model.encoder.layers.24.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1323
+ "vision_model.encoder.layers.24.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1324
  "vision_model.encoder.layers.24.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1325
  "vision_model.encoder.layers.24.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1326
+ "vision_model.encoder.layers.24.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1327
+ "vision_model.encoder.layers.24.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1328
  "vision_model.encoder.layers.24.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1329
  "vision_model.encoder.layers.24.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1330
+ "vision_model.encoder.layers.24.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1331
+ "vision_model.encoder.layers.24.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1332
  "vision_model.encoder.layers.24.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1333
  "vision_model.encoder.layers.24.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1334
+ "vision_model.encoder.layers.24.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1335
+ "vision_model.encoder.layers.24.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1336
  "vision_model.encoder.layers.24.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1337
  "vision_model.encoder.layers.24.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1338
+ "vision_model.encoder.layers.24.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1339
+ "vision_model.encoder.layers.24.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1340
  "vision_model.encoder.layers.24.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1341
  "vision_model.encoder.layers.25.layer_norm1.bias": "model-00001-of-00002.safetensors",
1342
  "vision_model.encoder.layers.25.layer_norm1.weight": "model-00001-of-00002.safetensors",
1343
  "vision_model.encoder.layers.25.layer_norm2.bias": "model-00001-of-00002.safetensors",
1344
  "vision_model.encoder.layers.25.layer_norm2.weight": "model-00001-of-00002.safetensors",
1345
  "vision_model.encoder.layers.25.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1346
+ "vision_model.encoder.layers.25.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1347
+ "vision_model.encoder.layers.25.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1348
  "vision_model.encoder.layers.25.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1349
  "vision_model.encoder.layers.25.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1350
+ "vision_model.encoder.layers.25.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1351
+ "vision_model.encoder.layers.25.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1352
  "vision_model.encoder.layers.25.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1353
  "vision_model.encoder.layers.25.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1354
+ "vision_model.encoder.layers.25.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1355
+ "vision_model.encoder.layers.25.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1356
  "vision_model.encoder.layers.25.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1357
  "vision_model.encoder.layers.25.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1358
+ "vision_model.encoder.layers.25.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1359
+ "vision_model.encoder.layers.25.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1360
  "vision_model.encoder.layers.25.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1361
  "vision_model.encoder.layers.25.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1362
+ "vision_model.encoder.layers.25.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1363
+ "vision_model.encoder.layers.25.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1364
  "vision_model.encoder.layers.25.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1365
  "vision_model.encoder.layers.25.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1366
+ "vision_model.encoder.layers.25.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1367
+ "vision_model.encoder.layers.25.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1368
  "vision_model.encoder.layers.25.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1369
  "vision_model.encoder.layers.26.layer_norm1.bias": "model-00001-of-00002.safetensors",
1370
  "vision_model.encoder.layers.26.layer_norm1.weight": "model-00001-of-00002.safetensors",
1371
  "vision_model.encoder.layers.26.layer_norm2.bias": "model-00001-of-00002.safetensors",
1372
  "vision_model.encoder.layers.26.layer_norm2.weight": "model-00001-of-00002.safetensors",
1373
  "vision_model.encoder.layers.26.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1374
+ "vision_model.encoder.layers.26.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1375
+ "vision_model.encoder.layers.26.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1376
  "vision_model.encoder.layers.26.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1377
  "vision_model.encoder.layers.26.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1378
+ "vision_model.encoder.layers.26.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1379
+ "vision_model.encoder.layers.26.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1380
  "vision_model.encoder.layers.26.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1381
  "vision_model.encoder.layers.26.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1382
+ "vision_model.encoder.layers.26.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1383
+ "vision_model.encoder.layers.26.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1384
  "vision_model.encoder.layers.26.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1385
  "vision_model.encoder.layers.26.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1386
+ "vision_model.encoder.layers.26.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1387
+ "vision_model.encoder.layers.26.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1388
  "vision_model.encoder.layers.26.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1389
  "vision_model.encoder.layers.26.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1390
+ "vision_model.encoder.layers.26.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1391
+ "vision_model.encoder.layers.26.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1392
  "vision_model.encoder.layers.26.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1393
  "vision_model.encoder.layers.26.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1394
+ "vision_model.encoder.layers.26.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1395
+ "vision_model.encoder.layers.26.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1396
  "vision_model.encoder.layers.26.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1397
  "vision_model.encoder.layers.3.layer_norm1.bias": "model-00001-of-00002.safetensors",
1398
  "vision_model.encoder.layers.3.layer_norm1.weight": "model-00001-of-00002.safetensors",
1399
  "vision_model.encoder.layers.3.layer_norm2.bias": "model-00001-of-00002.safetensors",
1400
  "vision_model.encoder.layers.3.layer_norm2.weight": "model-00001-of-00002.safetensors",
1401
  "vision_model.encoder.layers.3.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1402
+ "vision_model.encoder.layers.3.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1403
+ "vision_model.encoder.layers.3.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1404
  "vision_model.encoder.layers.3.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1405
  "vision_model.encoder.layers.3.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1406
+ "vision_model.encoder.layers.3.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1407
+ "vision_model.encoder.layers.3.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1408
  "vision_model.encoder.layers.3.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1409
  "vision_model.encoder.layers.3.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1410
+ "vision_model.encoder.layers.3.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1411
+ "vision_model.encoder.layers.3.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1412
  "vision_model.encoder.layers.3.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1413
  "vision_model.encoder.layers.3.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1414
+ "vision_model.encoder.layers.3.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1415
+ "vision_model.encoder.layers.3.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1416
  "vision_model.encoder.layers.3.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1417
  "vision_model.encoder.layers.3.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1418
+ "vision_model.encoder.layers.3.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1419
+ "vision_model.encoder.layers.3.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1420
  "vision_model.encoder.layers.3.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1421
  "vision_model.encoder.layers.3.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1422
+ "vision_model.encoder.layers.3.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1423
+ "vision_model.encoder.layers.3.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1424
  "vision_model.encoder.layers.3.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1425
  "vision_model.encoder.layers.4.layer_norm1.bias": "model-00001-of-00002.safetensors",
1426
  "vision_model.encoder.layers.4.layer_norm1.weight": "model-00001-of-00002.safetensors",
1427
  "vision_model.encoder.layers.4.layer_norm2.bias": "model-00001-of-00002.safetensors",
1428
  "vision_model.encoder.layers.4.layer_norm2.weight": "model-00001-of-00002.safetensors",
1429
  "vision_model.encoder.layers.4.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1430
+ "vision_model.encoder.layers.4.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1431
+ "vision_model.encoder.layers.4.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1432
  "vision_model.encoder.layers.4.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1433
  "vision_model.encoder.layers.4.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1434
+ "vision_model.encoder.layers.4.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1435
+ "vision_model.encoder.layers.4.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1436
  "vision_model.encoder.layers.4.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1437
  "vision_model.encoder.layers.4.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1438
+ "vision_model.encoder.layers.4.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1439
+ "vision_model.encoder.layers.4.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1440
  "vision_model.encoder.layers.4.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1441
  "vision_model.encoder.layers.4.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1442
+ "vision_model.encoder.layers.4.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1443
+ "vision_model.encoder.layers.4.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1444
  "vision_model.encoder.layers.4.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1445
  "vision_model.encoder.layers.4.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1446
+ "vision_model.encoder.layers.4.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1447
+ "vision_model.encoder.layers.4.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1448
  "vision_model.encoder.layers.4.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1449
  "vision_model.encoder.layers.4.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1450
+ "vision_model.encoder.layers.4.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1451
+ "vision_model.encoder.layers.4.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1452
  "vision_model.encoder.layers.4.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1453
  "vision_model.encoder.layers.5.layer_norm1.bias": "model-00001-of-00002.safetensors",
1454
  "vision_model.encoder.layers.5.layer_norm1.weight": "model-00001-of-00002.safetensors",
1455
  "vision_model.encoder.layers.5.layer_norm2.bias": "model-00001-of-00002.safetensors",
1456
  "vision_model.encoder.layers.5.layer_norm2.weight": "model-00001-of-00002.safetensors",
1457
  "vision_model.encoder.layers.5.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1458
+ "vision_model.encoder.layers.5.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1459
+ "vision_model.encoder.layers.5.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1460
  "vision_model.encoder.layers.5.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1461
  "vision_model.encoder.layers.5.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1462
+ "vision_model.encoder.layers.5.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1463
+ "vision_model.encoder.layers.5.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1464
  "vision_model.encoder.layers.5.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1465
  "vision_model.encoder.layers.5.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1466
+ "vision_model.encoder.layers.5.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1467
+ "vision_model.encoder.layers.5.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1468
  "vision_model.encoder.layers.5.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1469
  "vision_model.encoder.layers.5.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1470
+ "vision_model.encoder.layers.5.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1471
+ "vision_model.encoder.layers.5.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1472
  "vision_model.encoder.layers.5.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1473
  "vision_model.encoder.layers.5.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1474
+ "vision_model.encoder.layers.5.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1475
+ "vision_model.encoder.layers.5.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1476
  "vision_model.encoder.layers.5.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1477
  "vision_model.encoder.layers.5.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1478
+ "vision_model.encoder.layers.5.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1479
+ "vision_model.encoder.layers.5.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1480
  "vision_model.encoder.layers.5.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1481
  "vision_model.encoder.layers.6.layer_norm1.bias": "model-00001-of-00002.safetensors",
1482
  "vision_model.encoder.layers.6.layer_norm1.weight": "model-00001-of-00002.safetensors",
1483
  "vision_model.encoder.layers.6.layer_norm2.bias": "model-00001-of-00002.safetensors",
1484
  "vision_model.encoder.layers.6.layer_norm2.weight": "model-00001-of-00002.safetensors",
1485
  "vision_model.encoder.layers.6.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1486
+ "vision_model.encoder.layers.6.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1487
+ "vision_model.encoder.layers.6.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1488
  "vision_model.encoder.layers.6.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1489
  "vision_model.encoder.layers.6.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1490
+ "vision_model.encoder.layers.6.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1491
+ "vision_model.encoder.layers.6.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1492
  "vision_model.encoder.layers.6.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1493
  "vision_model.encoder.layers.6.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1494
+ "vision_model.encoder.layers.6.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1495
+ "vision_model.encoder.layers.6.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1496
  "vision_model.encoder.layers.6.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1497
  "vision_model.encoder.layers.6.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1498
+ "vision_model.encoder.layers.6.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1499
+ "vision_model.encoder.layers.6.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1500
  "vision_model.encoder.layers.6.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1501
  "vision_model.encoder.layers.6.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1502
+ "vision_model.encoder.layers.6.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1503
+ "vision_model.encoder.layers.6.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1504
  "vision_model.encoder.layers.6.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1505
  "vision_model.encoder.layers.6.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1506
+ "vision_model.encoder.layers.6.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1507
+ "vision_model.encoder.layers.6.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1508
  "vision_model.encoder.layers.6.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1509
  "vision_model.encoder.layers.7.layer_norm1.bias": "model-00001-of-00002.safetensors",
1510
  "vision_model.encoder.layers.7.layer_norm1.weight": "model-00001-of-00002.safetensors",
1511
  "vision_model.encoder.layers.7.layer_norm2.bias": "model-00001-of-00002.safetensors",
1512
  "vision_model.encoder.layers.7.layer_norm2.weight": "model-00001-of-00002.safetensors",
1513
  "vision_model.encoder.layers.7.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1514
+ "vision_model.encoder.layers.7.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1515
+ "vision_model.encoder.layers.7.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1516
  "vision_model.encoder.layers.7.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1517
  "vision_model.encoder.layers.7.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1518
+ "vision_model.encoder.layers.7.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1519
+ "vision_model.encoder.layers.7.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1520
  "vision_model.encoder.layers.7.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1521
  "vision_model.encoder.layers.7.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1522
+ "vision_model.encoder.layers.7.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1523
+ "vision_model.encoder.layers.7.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1524
  "vision_model.encoder.layers.7.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1525
  "vision_model.encoder.layers.7.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1526
+ "vision_model.encoder.layers.7.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1527
+ "vision_model.encoder.layers.7.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1528
  "vision_model.encoder.layers.7.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1529
  "vision_model.encoder.layers.7.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1530
+ "vision_model.encoder.layers.7.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1531
+ "vision_model.encoder.layers.7.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1532
  "vision_model.encoder.layers.7.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1533
  "vision_model.encoder.layers.7.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1534
+ "vision_model.encoder.layers.7.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1535
+ "vision_model.encoder.layers.7.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1536
  "vision_model.encoder.layers.7.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1537
  "vision_model.encoder.layers.8.layer_norm1.bias": "model-00001-of-00002.safetensors",
1538
  "vision_model.encoder.layers.8.layer_norm1.weight": "model-00001-of-00002.safetensors",
1539
  "vision_model.encoder.layers.8.layer_norm2.bias": "model-00001-of-00002.safetensors",
1540
  "vision_model.encoder.layers.8.layer_norm2.weight": "model-00001-of-00002.safetensors",
1541
  "vision_model.encoder.layers.8.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1542
+ "vision_model.encoder.layers.8.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1543
+ "vision_model.encoder.layers.8.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1544
  "vision_model.encoder.layers.8.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1545
  "vision_model.encoder.layers.8.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1546
+ "vision_model.encoder.layers.8.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1547
+ "vision_model.encoder.layers.8.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1548
  "vision_model.encoder.layers.8.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1549
  "vision_model.encoder.layers.8.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1550
+ "vision_model.encoder.layers.8.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1551
+ "vision_model.encoder.layers.8.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1552
  "vision_model.encoder.layers.8.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1553
  "vision_model.encoder.layers.8.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1554
+ "vision_model.encoder.layers.8.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1555
+ "vision_model.encoder.layers.8.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1556
  "vision_model.encoder.layers.8.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1557
  "vision_model.encoder.layers.8.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1558
+ "vision_model.encoder.layers.8.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1559
+ "vision_model.encoder.layers.8.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1560
  "vision_model.encoder.layers.8.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1561
  "vision_model.encoder.layers.8.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1562
+ "vision_model.encoder.layers.8.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1563
+ "vision_model.encoder.layers.8.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1564
  "vision_model.encoder.layers.8.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1565
  "vision_model.encoder.layers.9.layer_norm1.bias": "model-00001-of-00002.safetensors",
1566
  "vision_model.encoder.layers.9.layer_norm1.weight": "model-00001-of-00002.safetensors",
1567
  "vision_model.encoder.layers.9.layer_norm2.bias": "model-00001-of-00002.safetensors",
1568
  "vision_model.encoder.layers.9.layer_norm2.weight": "model-00001-of-00002.safetensors",
1569
  "vision_model.encoder.layers.9.mlp.fc1.bias": "model-00001-of-00002.safetensors",
1570
+ "vision_model.encoder.layers.9.mlp.fc1.biases": "model-00001-of-00002.safetensors",
1571
+ "vision_model.encoder.layers.9.mlp.fc1.scales": "model-00001-of-00002.safetensors",
1572
  "vision_model.encoder.layers.9.mlp.fc1.weight": "model-00001-of-00002.safetensors",
1573
  "vision_model.encoder.layers.9.mlp.fc2.bias": "model-00001-of-00002.safetensors",
1574
+ "vision_model.encoder.layers.9.mlp.fc2.biases": "model-00001-of-00002.safetensors",
1575
+ "vision_model.encoder.layers.9.mlp.fc2.scales": "model-00001-of-00002.safetensors",
1576
  "vision_model.encoder.layers.9.mlp.fc2.weight": "model-00001-of-00002.safetensors",
1577
  "vision_model.encoder.layers.9.self_attn.k_proj.bias": "model-00001-of-00002.safetensors",
1578
+ "vision_model.encoder.layers.9.self_attn.k_proj.biases": "model-00001-of-00002.safetensors",
1579
+ "vision_model.encoder.layers.9.self_attn.k_proj.scales": "model-00001-of-00002.safetensors",
1580
  "vision_model.encoder.layers.9.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1581
  "vision_model.encoder.layers.9.self_attn.out_proj.bias": "model-00001-of-00002.safetensors",
1582
+ "vision_model.encoder.layers.9.self_attn.out_proj.biases": "model-00001-of-00002.safetensors",
1583
+ "vision_model.encoder.layers.9.self_attn.out_proj.scales": "model-00001-of-00002.safetensors",
1584
  "vision_model.encoder.layers.9.self_attn.out_proj.weight": "model-00001-of-00002.safetensors",
1585
  "vision_model.encoder.layers.9.self_attn.q_proj.bias": "model-00001-of-00002.safetensors",
1586
+ "vision_model.encoder.layers.9.self_attn.q_proj.biases": "model-00001-of-00002.safetensors",
1587
+ "vision_model.encoder.layers.9.self_attn.q_proj.scales": "model-00001-of-00002.safetensors",
1588
  "vision_model.encoder.layers.9.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1589
  "vision_model.encoder.layers.9.self_attn.v_proj.bias": "model-00001-of-00002.safetensors",
1590
+ "vision_model.encoder.layers.9.self_attn.v_proj.biases": "model-00001-of-00002.safetensors",
1591
+ "vision_model.encoder.layers.9.self_attn.v_proj.scales": "model-00001-of-00002.safetensors",
1592
  "vision_model.encoder.layers.9.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1593
  "vision_model.post_layernorm.bias": "model-00001-of-00002.safetensors",
1594
  "vision_model.post_layernorm.weight": "model-00001-of-00002.safetensors"
tokenizer.json CHANGED
The diff for this file is too large to render. See raw diff
 
tokenizer_config.json CHANGED
@@ -1,6 +1,7 @@
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
 
4
  "added_tokens_decoder": {
5
  "0": {
6
  "content": "<unk>",
@@ -59,6 +60,7 @@
59
  "bos_token": "<s>",
60
  "clean_up_tokenization_spaces": false,
61
  "eos_token": "</s>",
 
62
  "model_max_length": 1000000000000000019884624838656,
63
  "pad_token": "<unk>",
64
  "processor_class": "Idefics2Processor",
 
1
  {
2
  "add_bos_token": true,
3
  "add_eos_token": false,
4
+ "add_prefix_space": null,
5
  "added_tokens_decoder": {
6
  "0": {
7
  "content": "<unk>",
 
60
  "bos_token": "<s>",
61
  "clean_up_tokenization_spaces": false,
62
  "eos_token": "</s>",
63
+ "legacy": false,
64
  "model_max_length": 1000000000000000019884624838656,
65
  "pad_token": "<unk>",
66
  "processor_class": "Idefics2Processor",