albertvillanova HF Staff commited on
Commit
a64ea3e
·
verified ·
1 Parent(s): b0b2701

Upload Qwen2_5_VLForConditionalGeneration

Browse files
Files changed (3) hide show
  1. config.json +5 -38
  2. generation_config.json +1 -1
  3. model.safetensors +1 -1
config.json CHANGED
@@ -2,42 +2,12 @@
2
  "architectures": [
3
  "Qwen2_5_VLForConditionalGeneration"
4
  ],
5
- "attention_dropout": 0.0,
6
- "bos_token_id": 151643,
7
  "dtype": "bfloat16",
8
- "eos_token_id": 151645,
9
- "hidden_act": "silu",
10
- "hidden_size": 2048,
11
  "image_token_id": 151655,
12
- "initializer_range": 0.02,
13
- "intermediate_size": 11008,
14
- "max_position_embeddings": 128000,
15
- "max_window_layers": 70,
16
  "model_type": "qwen2_5_vl",
17
- "num_attention_heads": 16,
18
- "num_hidden_layers": 36,
19
- "num_key_value_heads": 2,
20
- "rms_norm_eps": 1e-06,
21
- "rope_scaling": {
22
- "mrope_section": [
23
- 16,
24
- 24,
25
- 24
26
- ],
27
- "rope_type": "default",
28
- "type": "default"
29
- },
30
- "rope_theta": 1000000.0,
31
- "sliding_window": 32768,
32
  "text_config": {
33
- "_name_or_path": "Qwen/Qwen2.5-VL-3B-Instruct",
34
- "architectures": [
35
- "Qwen2_5_VLForConditionalGeneration"
36
- ],
37
  "attention_dropout": 0.0,
38
- "bos_token_id": 151643,
39
  "dtype": "bfloat16",
40
- "eos_token_id": 151645,
41
  "hidden_act": "silu",
42
  "hidden_size": 16,
43
  "initializer_range": 0.02,
@@ -53,25 +23,23 @@
53
  "num_hidden_layers": 2,
54
  "num_key_value_heads": 2,
55
  "rms_norm_eps": 1e-06,
56
- "rope_scaling": {
57
  "mrope_section": [
58
  1,
59
  1
60
  ],
 
61
  "rope_type": "default",
62
  "type": "default"
63
  },
64
- "rope_theta": 1000000.0,
65
  "sliding_window": null,
66
  "tie_word_embeddings": true,
67
  "use_cache": true,
68
  "use_sliding_window": false,
69
- "vision_token_id": 151654,
70
  "vocab_size": 151936
71
  },
72
- "transformers_version": "4.57.1",
73
- "use_cache": true,
74
- "use_sliding_window": false,
75
  "video_token_id": 151656,
76
  "vision_config": {
77
  "depth": 32,
@@ -103,6 +71,5 @@
103
  },
104
  "vision_end_token_id": 151653,
105
  "vision_start_token_id": 151652,
106
- "vision_token_id": 151654,
107
- "vocab_size": 151936
108
  }
 
2
  "architectures": [
3
  "Qwen2_5_VLForConditionalGeneration"
4
  ],
 
 
5
  "dtype": "bfloat16",
 
 
 
6
  "image_token_id": 151655,
 
 
 
 
7
  "model_type": "qwen2_5_vl",
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8
  "text_config": {
 
 
 
 
9
  "attention_dropout": 0.0,
 
10
  "dtype": "bfloat16",
 
11
  "hidden_act": "silu",
12
  "hidden_size": 16,
13
  "initializer_range": 0.02,
 
23
  "num_hidden_layers": 2,
24
  "num_key_value_heads": 2,
25
  "rms_norm_eps": 1e-06,
26
+ "rope_parameters": {
27
  "mrope_section": [
28
  1,
29
  1
30
  ],
31
+ "rope_theta": 1000000.0,
32
  "rope_type": "default",
33
  "type": "default"
34
  },
 
35
  "sliding_window": null,
36
  "tie_word_embeddings": true,
37
  "use_cache": true,
38
  "use_sliding_window": false,
 
39
  "vocab_size": 151936
40
  },
41
+ "tie_word_embeddings": true,
42
+ "transformers_version": "5.0.0.dev0",
 
43
  "video_token_id": 151656,
44
  "vision_config": {
45
  "depth": 32,
 
71
  },
72
  "vision_end_token_id": 151653,
73
  "vision_start_token_id": 151652,
74
+ "vision_token_id": 151654
 
75
  }
generation_config.json CHANGED
@@ -2,5 +2,5 @@
2
  "_from_model_config": true,
3
  "bos_token_id": 151643,
4
  "eos_token_id": 151645,
5
- "transformers_version": "4.57.1"
6
  }
 
2
  "_from_model_config": true,
3
  "bos_token_id": 151643,
4
  "eos_token_id": 151645,
5
+ "transformers_version": "5.0.0.dev0"
6
  }
model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:e9ee6de012cbabb328744133306ae85256f9d1d994145c50a0a341918c7af35c
3
  size 18086192
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0d65a042afb5cad045b9e8d426c926dec7e479d4f11453cd6ce30779074c0097
3
  size 18086192