danielhanchen commited on
Commit
5993019
1 Parent(s): 5c53ccd

Upload MllamaForConditionalGeneration

Browse files
config.json CHANGED
@@ -21,6 +21,7 @@
21
  "quant_method": "bitsandbytes"
22
  },
23
  "text_config": {
 
24
  "_name_or_path": "",
25
  "add_cross_attention": false,
26
  "architectures": null,
@@ -114,8 +115,10 @@
114
  "vocab_size": 128256
115
  },
116
  "torch_dtype": "bfloat16",
117
- "transformers_version": "4.45.0",
 
118
  "vision_config": {
 
119
  "_name_or_path": "",
120
  "add_cross_attention": false,
121
  "architectures": null,
 
21
  "quant_method": "bitsandbytes"
22
  },
23
  "text_config": {
24
+ "_attn_implementation_autoset": false,
25
  "_name_or_path": "",
26
  "add_cross_attention": false,
27
  "architectures": null,
 
115
  "vocab_size": 128256
116
  },
117
  "torch_dtype": "bfloat16",
118
+ "transformers_version": "4.46.3",
119
+ "unsloth_fixed": true,
120
  "vision_config": {
121
+ "_attn_implementation_autoset": false,
122
  "_name_or_path": "",
123
  "add_cross_attention": false,
124
  "architectures": null,
generation_config.json CHANGED
@@ -3,5 +3,5 @@
3
  "bos_token_id": 128000,
4
  "eos_token_id": 128001,
5
  "pad_token_id": 128004,
6
- "transformers_version": "4.45.0"
7
  }
 
3
  "bos_token_id": 128000,
4
  "eos_token_id": 128001,
5
  "pad_token_id": 128004,
6
+ "transformers_version": "4.46.3"
7
  }
model-00001-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:2e69f8fcdfd5863e6bfa5b57101d06d00ff078b11e372ca71cbb4bde23882a42
3
- size 4984953659
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:449b885c3695adb34f8184e1b499b3f0244de9eceaa67b3055cded34683557f8
3
+ size 6058466370
model-00002-of-00002.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:48975aae8e6a2efcb65673e54d170db7958a59aedcecb4b03bb6b0e5e33342f8
3
- size 2140423619
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:49625c07f1a5518f9bfd693bc8256a579a91ed6c4617abf2075f44074fb9abf3
3
+ size 1066911340
model.safetensors.index.json CHANGED
@@ -1081,26 +1081,26 @@
1081
  "language_model.model.layers.3.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1082
  "language_model.model.layers.3.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1083
  "language_model.model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1084
- "language_model.model.layers.30.input_layernorm.weight": "model-00002-of-00002.safetensors",
1085
- "language_model.model.layers.30.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1086
- "language_model.model.layers.30.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1087
- "language_model.model.layers.30.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1088
- "language_model.model.layers.30.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1089
- "language_model.model.layers.30.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1090
- "language_model.model.layers.30.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1091
  "language_model.model.layers.30.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1092
  "language_model.model.layers.30.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1093
  "language_model.model.layers.30.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1094
  "language_model.model.layers.30.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1095
  "language_model.model.layers.30.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1096
  "language_model.model.layers.30.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1097
- "language_model.model.layers.30.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1098
- "language_model.model.layers.30.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1099
- "language_model.model.layers.30.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1100
- "language_model.model.layers.30.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1101
- "language_model.model.layers.30.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1102
- "language_model.model.layers.30.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1103
- "language_model.model.layers.30.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1104
  "language_model.model.layers.30.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1105
  "language_model.model.layers.30.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1106
  "language_model.model.layers.30.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
@@ -1125,410 +1125,410 @@
1125
  "language_model.model.layers.30.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1126
  "language_model.model.layers.30.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1127
  "language_model.model.layers.30.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1128
- "language_model.model.layers.31.input_layernorm.weight": "model-00002-of-00002.safetensors",
1129
- "language_model.model.layers.31.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1130
- "language_model.model.layers.31.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1131
- "language_model.model.layers.31.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1132
- "language_model.model.layers.31.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1133
- "language_model.model.layers.31.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1134
- "language_model.model.layers.31.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1135
- "language_model.model.layers.31.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1136
- "language_model.model.layers.31.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1137
- "language_model.model.layers.31.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1138
- "language_model.model.layers.31.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1139
- "language_model.model.layers.31.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1140
- "language_model.model.layers.31.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1141
- "language_model.model.layers.31.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1142
- "language_model.model.layers.31.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1143
- "language_model.model.layers.31.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1144
- "language_model.model.layers.31.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1145
- "language_model.model.layers.31.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1146
- "language_model.model.layers.31.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1147
- "language_model.model.layers.31.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1148
- "language_model.model.layers.31.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1149
- "language_model.model.layers.31.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1150
- "language_model.model.layers.31.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1151
- "language_model.model.layers.31.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1152
- "language_model.model.layers.31.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1153
- "language_model.model.layers.31.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1154
- "language_model.model.layers.31.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1155
- "language_model.model.layers.31.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1156
- "language_model.model.layers.31.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1157
- "language_model.model.layers.31.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1158
- "language_model.model.layers.31.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1159
- "language_model.model.layers.31.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1160
- "language_model.model.layers.31.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1161
- "language_model.model.layers.31.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1162
- "language_model.model.layers.31.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1163
- "language_model.model.layers.31.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1164
- "language_model.model.layers.31.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1165
- "language_model.model.layers.31.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1166
- "language_model.model.layers.31.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1167
- "language_model.model.layers.31.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1168
- "language_model.model.layers.31.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1169
- "language_model.model.layers.31.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1170
- "language_model.model.layers.31.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1171
- "language_model.model.layers.31.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1172
- "language_model.model.layers.32.input_layernorm.weight": "model-00002-of-00002.safetensors",
1173
- "language_model.model.layers.32.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1174
- "language_model.model.layers.32.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1175
- "language_model.model.layers.32.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1176
- "language_model.model.layers.32.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1177
- "language_model.model.layers.32.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1178
- "language_model.model.layers.32.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1179
- "language_model.model.layers.32.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1180
- "language_model.model.layers.32.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1181
- "language_model.model.layers.32.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1182
- "language_model.model.layers.32.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1183
- "language_model.model.layers.32.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1184
- "language_model.model.layers.32.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1185
- "language_model.model.layers.32.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1186
- "language_model.model.layers.32.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1187
- "language_model.model.layers.32.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1188
- "language_model.model.layers.32.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1189
- "language_model.model.layers.32.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1190
- "language_model.model.layers.32.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1191
- "language_model.model.layers.32.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1192
- "language_model.model.layers.32.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1193
- "language_model.model.layers.32.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1194
- "language_model.model.layers.32.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1195
- "language_model.model.layers.32.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1196
- "language_model.model.layers.32.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1197
- "language_model.model.layers.32.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1198
- "language_model.model.layers.32.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1199
- "language_model.model.layers.32.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1200
- "language_model.model.layers.32.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1201
- "language_model.model.layers.32.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1202
- "language_model.model.layers.32.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1203
- "language_model.model.layers.32.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1204
- "language_model.model.layers.32.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1205
- "language_model.model.layers.32.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1206
- "language_model.model.layers.32.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1207
- "language_model.model.layers.32.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1208
- "language_model.model.layers.32.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1209
- "language_model.model.layers.32.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1210
- "language_model.model.layers.32.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1211
- "language_model.model.layers.32.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1212
- "language_model.model.layers.32.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1213
- "language_model.model.layers.32.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1214
- "language_model.model.layers.32.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1215
- "language_model.model.layers.32.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1216
- "language_model.model.layers.33.cross_attn.k_norm.weight": "model-00002-of-00002.safetensors",
1217
- "language_model.model.layers.33.cross_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1218
- "language_model.model.layers.33.cross_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1219
- "language_model.model.layers.33.cross_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1220
- "language_model.model.layers.33.cross_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1221
- "language_model.model.layers.33.cross_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1222
- "language_model.model.layers.33.cross_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1223
- "language_model.model.layers.33.cross_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1224
- "language_model.model.layers.33.cross_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1225
- "language_model.model.layers.33.cross_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1226
- "language_model.model.layers.33.cross_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1227
- "language_model.model.layers.33.cross_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1228
- "language_model.model.layers.33.cross_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1229
- "language_model.model.layers.33.cross_attn.q_norm.weight": "model-00002-of-00002.safetensors",
1230
- "language_model.model.layers.33.cross_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1231
- "language_model.model.layers.33.cross_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1232
- "language_model.model.layers.33.cross_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1233
- "language_model.model.layers.33.cross_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1234
- "language_model.model.layers.33.cross_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1235
- "language_model.model.layers.33.cross_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1236
- "language_model.model.layers.33.cross_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1237
- "language_model.model.layers.33.cross_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1238
- "language_model.model.layers.33.cross_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1239
- "language_model.model.layers.33.cross_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1240
- "language_model.model.layers.33.cross_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1241
- "language_model.model.layers.33.cross_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1242
- "language_model.model.layers.33.cross_attn_attn_gate": "model-00002-of-00002.safetensors",
1243
- "language_model.model.layers.33.cross_attn_mlp_gate": "model-00002-of-00002.safetensors",
1244
- "language_model.model.layers.33.input_layernorm.weight": "model-00002-of-00002.safetensors",
1245
- "language_model.model.layers.33.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1246
- "language_model.model.layers.33.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1247
- "language_model.model.layers.33.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1248
- "language_model.model.layers.33.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1249
- "language_model.model.layers.33.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1250
- "language_model.model.layers.33.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1251
- "language_model.model.layers.33.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1252
- "language_model.model.layers.33.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1253
- "language_model.model.layers.33.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1254
- "language_model.model.layers.33.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1255
- "language_model.model.layers.33.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1256
- "language_model.model.layers.33.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1257
- "language_model.model.layers.33.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1258
- "language_model.model.layers.33.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1259
- "language_model.model.layers.33.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1260
- "language_model.model.layers.33.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1261
- "language_model.model.layers.33.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1262
- "language_model.model.layers.33.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1263
- "language_model.model.layers.33.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1264
- "language_model.model.layers.34.input_layernorm.weight": "model-00002-of-00002.safetensors",
1265
- "language_model.model.layers.34.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1266
- "language_model.model.layers.34.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1267
- "language_model.model.layers.34.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1268
- "language_model.model.layers.34.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1269
- "language_model.model.layers.34.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1270
- "language_model.model.layers.34.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1271
- "language_model.model.layers.34.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1272
- "language_model.model.layers.34.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1273
- "language_model.model.layers.34.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1274
- "language_model.model.layers.34.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1275
- "language_model.model.layers.34.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1276
- "language_model.model.layers.34.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1277
- "language_model.model.layers.34.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1278
- "language_model.model.layers.34.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1279
- "language_model.model.layers.34.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1280
- "language_model.model.layers.34.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1281
- "language_model.model.layers.34.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1282
- "language_model.model.layers.34.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1283
- "language_model.model.layers.34.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1284
- "language_model.model.layers.34.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1285
- "language_model.model.layers.34.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1286
- "language_model.model.layers.34.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1287
- "language_model.model.layers.34.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1288
- "language_model.model.layers.34.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1289
- "language_model.model.layers.34.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1290
- "language_model.model.layers.34.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1291
- "language_model.model.layers.34.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1292
- "language_model.model.layers.34.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1293
- "language_model.model.layers.34.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1294
- "language_model.model.layers.34.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1295
- "language_model.model.layers.34.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1296
- "language_model.model.layers.34.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1297
- "language_model.model.layers.34.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1298
- "language_model.model.layers.34.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1299
- "language_model.model.layers.34.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1300
- "language_model.model.layers.34.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1301
- "language_model.model.layers.34.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1302
- "language_model.model.layers.34.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1303
- "language_model.model.layers.34.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1304
- "language_model.model.layers.34.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1305
- "language_model.model.layers.34.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1306
- "language_model.model.layers.34.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1307
- "language_model.model.layers.34.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1308
- "language_model.model.layers.35.input_layernorm.weight": "model-00002-of-00002.safetensors",
1309
- "language_model.model.layers.35.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1310
- "language_model.model.layers.35.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1311
- "language_model.model.layers.35.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1312
- "language_model.model.layers.35.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1313
- "language_model.model.layers.35.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1314
- "language_model.model.layers.35.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1315
- "language_model.model.layers.35.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1316
- "language_model.model.layers.35.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1317
- "language_model.model.layers.35.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1318
- "language_model.model.layers.35.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1319
- "language_model.model.layers.35.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1320
- "language_model.model.layers.35.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1321
- "language_model.model.layers.35.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1322
- "language_model.model.layers.35.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1323
- "language_model.model.layers.35.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1324
- "language_model.model.layers.35.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1325
- "language_model.model.layers.35.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1326
- "language_model.model.layers.35.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1327
- "language_model.model.layers.35.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1328
- "language_model.model.layers.35.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1329
- "language_model.model.layers.35.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1330
- "language_model.model.layers.35.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1331
- "language_model.model.layers.35.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1332
- "language_model.model.layers.35.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1333
- "language_model.model.layers.35.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1334
- "language_model.model.layers.35.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1335
- "language_model.model.layers.35.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1336
- "language_model.model.layers.35.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1337
- "language_model.model.layers.35.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1338
- "language_model.model.layers.35.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1339
- "language_model.model.layers.35.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1340
- "language_model.model.layers.35.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1341
- "language_model.model.layers.35.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1342
- "language_model.model.layers.35.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1343
- "language_model.model.layers.35.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1344
- "language_model.model.layers.35.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1345
- "language_model.model.layers.35.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1346
- "language_model.model.layers.35.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1347
- "language_model.model.layers.35.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1348
- "language_model.model.layers.35.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1349
- "language_model.model.layers.35.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1350
- "language_model.model.layers.35.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1351
- "language_model.model.layers.35.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1352
- "language_model.model.layers.36.input_layernorm.weight": "model-00002-of-00002.safetensors",
1353
- "language_model.model.layers.36.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1354
- "language_model.model.layers.36.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1355
- "language_model.model.layers.36.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1356
- "language_model.model.layers.36.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1357
- "language_model.model.layers.36.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1358
- "language_model.model.layers.36.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1359
- "language_model.model.layers.36.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1360
- "language_model.model.layers.36.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1361
- "language_model.model.layers.36.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1362
- "language_model.model.layers.36.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1363
- "language_model.model.layers.36.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1364
- "language_model.model.layers.36.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1365
- "language_model.model.layers.36.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1366
- "language_model.model.layers.36.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1367
- "language_model.model.layers.36.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1368
- "language_model.model.layers.36.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1369
- "language_model.model.layers.36.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1370
- "language_model.model.layers.36.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1371
- "language_model.model.layers.36.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1372
- "language_model.model.layers.36.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1373
- "language_model.model.layers.36.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1374
- "language_model.model.layers.36.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1375
- "language_model.model.layers.36.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1376
- "language_model.model.layers.36.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1377
- "language_model.model.layers.36.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1378
- "language_model.model.layers.36.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1379
- "language_model.model.layers.36.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1380
- "language_model.model.layers.36.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1381
- "language_model.model.layers.36.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1382
- "language_model.model.layers.36.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1383
- "language_model.model.layers.36.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1384
- "language_model.model.layers.36.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1385
- "language_model.model.layers.36.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1386
- "language_model.model.layers.36.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1387
- "language_model.model.layers.36.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1388
- "language_model.model.layers.36.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1389
- "language_model.model.layers.36.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1390
- "language_model.model.layers.36.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1391
- "language_model.model.layers.36.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1392
- "language_model.model.layers.36.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1393
- "language_model.model.layers.36.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1394
- "language_model.model.layers.36.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1395
- "language_model.model.layers.36.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1396
- "language_model.model.layers.37.input_layernorm.weight": "model-00002-of-00002.safetensors",
1397
- "language_model.model.layers.37.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1398
- "language_model.model.layers.37.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1399
- "language_model.model.layers.37.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1400
- "language_model.model.layers.37.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1401
- "language_model.model.layers.37.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1402
- "language_model.model.layers.37.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1403
- "language_model.model.layers.37.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1404
- "language_model.model.layers.37.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1405
- "language_model.model.layers.37.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1406
- "language_model.model.layers.37.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1407
- "language_model.model.layers.37.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1408
- "language_model.model.layers.37.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1409
- "language_model.model.layers.37.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1410
- "language_model.model.layers.37.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1411
- "language_model.model.layers.37.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1412
- "language_model.model.layers.37.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1413
- "language_model.model.layers.37.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1414
- "language_model.model.layers.37.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1415
- "language_model.model.layers.37.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1416
- "language_model.model.layers.37.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1417
- "language_model.model.layers.37.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1418
- "language_model.model.layers.37.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1419
- "language_model.model.layers.37.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1420
- "language_model.model.layers.37.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1421
- "language_model.model.layers.37.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1422
- "language_model.model.layers.37.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1423
- "language_model.model.layers.37.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1424
- "language_model.model.layers.37.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1425
- "language_model.model.layers.37.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1426
- "language_model.model.layers.37.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1427
- "language_model.model.layers.37.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1428
- "language_model.model.layers.37.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1429
- "language_model.model.layers.37.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1430
- "language_model.model.layers.37.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1431
- "language_model.model.layers.37.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1432
- "language_model.model.layers.37.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1433
- "language_model.model.layers.37.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1434
- "language_model.model.layers.37.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1435
- "language_model.model.layers.37.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1436
- "language_model.model.layers.37.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1437
- "language_model.model.layers.37.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1438
- "language_model.model.layers.37.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1439
- "language_model.model.layers.37.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1440
- "language_model.model.layers.38.cross_attn.k_norm.weight": "model-00002-of-00002.safetensors",
1441
- "language_model.model.layers.38.cross_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1442
- "language_model.model.layers.38.cross_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1443
- "language_model.model.layers.38.cross_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1444
- "language_model.model.layers.38.cross_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1445
- "language_model.model.layers.38.cross_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1446
- "language_model.model.layers.38.cross_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1447
- "language_model.model.layers.38.cross_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1448
- "language_model.model.layers.38.cross_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1449
- "language_model.model.layers.38.cross_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1450
- "language_model.model.layers.38.cross_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1451
- "language_model.model.layers.38.cross_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1452
- "language_model.model.layers.38.cross_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1453
- "language_model.model.layers.38.cross_attn.q_norm.weight": "model-00002-of-00002.safetensors",
1454
- "language_model.model.layers.38.cross_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1455
- "language_model.model.layers.38.cross_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1456
- "language_model.model.layers.38.cross_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1457
- "language_model.model.layers.38.cross_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1458
- "language_model.model.layers.38.cross_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1459
- "language_model.model.layers.38.cross_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1460
- "language_model.model.layers.38.cross_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1461
- "language_model.model.layers.38.cross_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1462
- "language_model.model.layers.38.cross_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1463
- "language_model.model.layers.38.cross_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1464
- "language_model.model.layers.38.cross_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1465
- "language_model.model.layers.38.cross_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1466
- "language_model.model.layers.38.cross_attn_attn_gate": "model-00002-of-00002.safetensors",
1467
- "language_model.model.layers.38.cross_attn_mlp_gate": "model-00002-of-00002.safetensors",
1468
- "language_model.model.layers.38.input_layernorm.weight": "model-00002-of-00002.safetensors",
1469
- "language_model.model.layers.38.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1470
- "language_model.model.layers.38.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1471
- "language_model.model.layers.38.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1472
- "language_model.model.layers.38.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1473
- "language_model.model.layers.38.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1474
- "language_model.model.layers.38.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1475
- "language_model.model.layers.38.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1476
- "language_model.model.layers.38.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1477
- "language_model.model.layers.38.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1478
- "language_model.model.layers.38.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1479
- "language_model.model.layers.38.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1480
- "language_model.model.layers.38.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1481
- "language_model.model.layers.38.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1482
- "language_model.model.layers.38.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1483
- "language_model.model.layers.38.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1484
- "language_model.model.layers.38.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1485
- "language_model.model.layers.38.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1486
- "language_model.model.layers.38.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1487
- "language_model.model.layers.38.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1488
- "language_model.model.layers.39.input_layernorm.weight": "model-00002-of-00002.safetensors",
1489
- "language_model.model.layers.39.mlp.down_proj.weight": "model-00002-of-00002.safetensors",
1490
- "language_model.model.layers.39.mlp.down_proj.weight.absmax": "model-00002-of-00002.safetensors",
1491
- "language_model.model.layers.39.mlp.down_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1492
- "language_model.model.layers.39.mlp.down_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1493
- "language_model.model.layers.39.mlp.down_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1494
- "language_model.model.layers.39.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1495
- "language_model.model.layers.39.mlp.gate_proj.weight": "model-00002-of-00002.safetensors",
1496
- "language_model.model.layers.39.mlp.gate_proj.weight.absmax": "model-00002-of-00002.safetensors",
1497
- "language_model.model.layers.39.mlp.gate_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1498
- "language_model.model.layers.39.mlp.gate_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1499
- "language_model.model.layers.39.mlp.gate_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1500
- "language_model.model.layers.39.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1501
- "language_model.model.layers.39.mlp.up_proj.weight": "model-00002-of-00002.safetensors",
1502
- "language_model.model.layers.39.mlp.up_proj.weight.absmax": "model-00002-of-00002.safetensors",
1503
- "language_model.model.layers.39.mlp.up_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1504
- "language_model.model.layers.39.mlp.up_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1505
- "language_model.model.layers.39.mlp.up_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1506
- "language_model.model.layers.39.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1507
- "language_model.model.layers.39.post_attention_layernorm.weight": "model-00002-of-00002.safetensors",
1508
- "language_model.model.layers.39.self_attn.k_proj.weight": "model-00002-of-00002.safetensors",
1509
- "language_model.model.layers.39.self_attn.k_proj.weight.absmax": "model-00002-of-00002.safetensors",
1510
- "language_model.model.layers.39.self_attn.k_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1511
- "language_model.model.layers.39.self_attn.k_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1512
- "language_model.model.layers.39.self_attn.k_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1513
- "language_model.model.layers.39.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1514
- "language_model.model.layers.39.self_attn.o_proj.weight": "model-00002-of-00002.safetensors",
1515
- "language_model.model.layers.39.self_attn.o_proj.weight.absmax": "model-00002-of-00002.safetensors",
1516
- "language_model.model.layers.39.self_attn.o_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1517
- "language_model.model.layers.39.self_attn.o_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1518
- "language_model.model.layers.39.self_attn.o_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1519
- "language_model.model.layers.39.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1520
- "language_model.model.layers.39.self_attn.q_proj.weight": "model-00002-of-00002.safetensors",
1521
- "language_model.model.layers.39.self_attn.q_proj.weight.absmax": "model-00002-of-00002.safetensors",
1522
- "language_model.model.layers.39.self_attn.q_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1523
- "language_model.model.layers.39.self_attn.q_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1524
- "language_model.model.layers.39.self_attn.q_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1525
- "language_model.model.layers.39.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1526
- "language_model.model.layers.39.self_attn.v_proj.weight": "model-00002-of-00002.safetensors",
1527
- "language_model.model.layers.39.self_attn.v_proj.weight.absmax": "model-00002-of-00002.safetensors",
1528
- "language_model.model.layers.39.self_attn.v_proj.weight.nested_absmax": "model-00002-of-00002.safetensors",
1529
- "language_model.model.layers.39.self_attn.v_proj.weight.nested_quant_map": "model-00002-of-00002.safetensors",
1530
- "language_model.model.layers.39.self_attn.v_proj.weight.quant_map": "model-00002-of-00002.safetensors",
1531
- "language_model.model.layers.39.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00002-of-00002.safetensors",
1532
  "language_model.model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
1533
  "language_model.model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1534
  "language_model.model.layers.4.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
@@ -1797,7 +1797,7 @@
1797
  "language_model.model.layers.9.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1798
  "language_model.model.layers.9.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1799
  "language_model.model.layers.9.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1800
- "language_model.model.norm.weight": "model-00002-of-00002.safetensors",
1801
  "multi_modal_projector.bias": "model-00002-of-00002.safetensors",
1802
  "multi_modal_projector.weight": "model-00002-of-00002.safetensors",
1803
  "multi_modal_projector.weight.absmax": "model-00002-of-00002.safetensors",
 
1081
  "language_model.model.layers.3.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1082
  "language_model.model.layers.3.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1083
  "language_model.model.layers.3.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1084
+ "language_model.model.layers.30.input_layernorm.weight": "model-00001-of-00002.safetensors",
1085
+ "language_model.model.layers.30.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1086
+ "language_model.model.layers.30.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1087
+ "language_model.model.layers.30.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1088
+ "language_model.model.layers.30.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1089
+ "language_model.model.layers.30.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1090
+ "language_model.model.layers.30.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1091
  "language_model.model.layers.30.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1092
  "language_model.model.layers.30.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1093
  "language_model.model.layers.30.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1094
  "language_model.model.layers.30.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1095
  "language_model.model.layers.30.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1096
  "language_model.model.layers.30.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1097
+ "language_model.model.layers.30.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1098
+ "language_model.model.layers.30.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1099
+ "language_model.model.layers.30.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1100
+ "language_model.model.layers.30.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1101
+ "language_model.model.layers.30.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1102
+ "language_model.model.layers.30.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1103
+ "language_model.model.layers.30.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1104
  "language_model.model.layers.30.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1105
  "language_model.model.layers.30.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1106
  "language_model.model.layers.30.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
 
1125
  "language_model.model.layers.30.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1126
  "language_model.model.layers.30.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1127
  "language_model.model.layers.30.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1128
+ "language_model.model.layers.31.input_layernorm.weight": "model-00001-of-00002.safetensors",
1129
+ "language_model.model.layers.31.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1130
+ "language_model.model.layers.31.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1131
+ "language_model.model.layers.31.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1132
+ "language_model.model.layers.31.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1133
+ "language_model.model.layers.31.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1134
+ "language_model.model.layers.31.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1135
+ "language_model.model.layers.31.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1136
+ "language_model.model.layers.31.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1137
+ "language_model.model.layers.31.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1138
+ "language_model.model.layers.31.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1139
+ "language_model.model.layers.31.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1140
+ "language_model.model.layers.31.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1141
+ "language_model.model.layers.31.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1142
+ "language_model.model.layers.31.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1143
+ "language_model.model.layers.31.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1144
+ "language_model.model.layers.31.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1145
+ "language_model.model.layers.31.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1146
+ "language_model.model.layers.31.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1147
+ "language_model.model.layers.31.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1148
+ "language_model.model.layers.31.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1149
+ "language_model.model.layers.31.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1150
+ "language_model.model.layers.31.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1151
+ "language_model.model.layers.31.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1152
+ "language_model.model.layers.31.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1153
+ "language_model.model.layers.31.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1154
+ "language_model.model.layers.31.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1155
+ "language_model.model.layers.31.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1156
+ "language_model.model.layers.31.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1157
+ "language_model.model.layers.31.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1158
+ "language_model.model.layers.31.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1159
+ "language_model.model.layers.31.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1160
+ "language_model.model.layers.31.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1161
+ "language_model.model.layers.31.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1162
+ "language_model.model.layers.31.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1163
+ "language_model.model.layers.31.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1164
+ "language_model.model.layers.31.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1165
+ "language_model.model.layers.31.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1166
+ "language_model.model.layers.31.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1167
+ "language_model.model.layers.31.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1168
+ "language_model.model.layers.31.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1169
+ "language_model.model.layers.31.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1170
+ "language_model.model.layers.31.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1171
+ "language_model.model.layers.31.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1172
+ "language_model.model.layers.32.input_layernorm.weight": "model-00001-of-00002.safetensors",
1173
+ "language_model.model.layers.32.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1174
+ "language_model.model.layers.32.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1175
+ "language_model.model.layers.32.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1176
+ "language_model.model.layers.32.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1177
+ "language_model.model.layers.32.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1178
+ "language_model.model.layers.32.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1179
+ "language_model.model.layers.32.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1180
+ "language_model.model.layers.32.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1181
+ "language_model.model.layers.32.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1182
+ "language_model.model.layers.32.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1183
+ "language_model.model.layers.32.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1184
+ "language_model.model.layers.32.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1185
+ "language_model.model.layers.32.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1186
+ "language_model.model.layers.32.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1187
+ "language_model.model.layers.32.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1188
+ "language_model.model.layers.32.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1189
+ "language_model.model.layers.32.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1190
+ "language_model.model.layers.32.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1191
+ "language_model.model.layers.32.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1192
+ "language_model.model.layers.32.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1193
+ "language_model.model.layers.32.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1194
+ "language_model.model.layers.32.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1195
+ "language_model.model.layers.32.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1196
+ "language_model.model.layers.32.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1197
+ "language_model.model.layers.32.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1198
+ "language_model.model.layers.32.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1199
+ "language_model.model.layers.32.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1200
+ "language_model.model.layers.32.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1201
+ "language_model.model.layers.32.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1202
+ "language_model.model.layers.32.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1203
+ "language_model.model.layers.32.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1204
+ "language_model.model.layers.32.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1205
+ "language_model.model.layers.32.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1206
+ "language_model.model.layers.32.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1207
+ "language_model.model.layers.32.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1208
+ "language_model.model.layers.32.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1209
+ "language_model.model.layers.32.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1210
+ "language_model.model.layers.32.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1211
+ "language_model.model.layers.32.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1212
+ "language_model.model.layers.32.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1213
+ "language_model.model.layers.32.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1214
+ "language_model.model.layers.32.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1215
+ "language_model.model.layers.32.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1216
+ "language_model.model.layers.33.cross_attn.k_norm.weight": "model-00001-of-00002.safetensors",
1217
+ "language_model.model.layers.33.cross_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1218
+ "language_model.model.layers.33.cross_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1219
+ "language_model.model.layers.33.cross_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1220
+ "language_model.model.layers.33.cross_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1221
+ "language_model.model.layers.33.cross_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1222
+ "language_model.model.layers.33.cross_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1223
+ "language_model.model.layers.33.cross_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1224
+ "language_model.model.layers.33.cross_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1225
+ "language_model.model.layers.33.cross_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1226
+ "language_model.model.layers.33.cross_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1227
+ "language_model.model.layers.33.cross_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1228
+ "language_model.model.layers.33.cross_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1229
+ "language_model.model.layers.33.cross_attn.q_norm.weight": "model-00001-of-00002.safetensors",
1230
+ "language_model.model.layers.33.cross_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1231
+ "language_model.model.layers.33.cross_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1232
+ "language_model.model.layers.33.cross_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1233
+ "language_model.model.layers.33.cross_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1234
+ "language_model.model.layers.33.cross_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1235
+ "language_model.model.layers.33.cross_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1236
+ "language_model.model.layers.33.cross_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1237
+ "language_model.model.layers.33.cross_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1238
+ "language_model.model.layers.33.cross_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1239
+ "language_model.model.layers.33.cross_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1240
+ "language_model.model.layers.33.cross_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1241
+ "language_model.model.layers.33.cross_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1242
+ "language_model.model.layers.33.cross_attn_attn_gate": "model-00001-of-00002.safetensors",
1243
+ "language_model.model.layers.33.cross_attn_mlp_gate": "model-00001-of-00002.safetensors",
1244
+ "language_model.model.layers.33.input_layernorm.weight": "model-00001-of-00002.safetensors",
1245
+ "language_model.model.layers.33.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1246
+ "language_model.model.layers.33.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1247
+ "language_model.model.layers.33.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1248
+ "language_model.model.layers.33.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1249
+ "language_model.model.layers.33.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1250
+ "language_model.model.layers.33.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1251
+ "language_model.model.layers.33.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1252
+ "language_model.model.layers.33.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1253
+ "language_model.model.layers.33.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1254
+ "language_model.model.layers.33.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1255
+ "language_model.model.layers.33.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1256
+ "language_model.model.layers.33.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1257
+ "language_model.model.layers.33.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1258
+ "language_model.model.layers.33.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1259
+ "language_model.model.layers.33.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1260
+ "language_model.model.layers.33.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1261
+ "language_model.model.layers.33.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1262
+ "language_model.model.layers.33.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1263
+ "language_model.model.layers.33.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1264
+ "language_model.model.layers.34.input_layernorm.weight": "model-00001-of-00002.safetensors",
1265
+ "language_model.model.layers.34.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1266
+ "language_model.model.layers.34.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1267
+ "language_model.model.layers.34.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1268
+ "language_model.model.layers.34.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1269
+ "language_model.model.layers.34.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1270
+ "language_model.model.layers.34.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1271
+ "language_model.model.layers.34.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1272
+ "language_model.model.layers.34.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1273
+ "language_model.model.layers.34.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1274
+ "language_model.model.layers.34.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1275
+ "language_model.model.layers.34.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1276
+ "language_model.model.layers.34.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1277
+ "language_model.model.layers.34.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1278
+ "language_model.model.layers.34.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1279
+ "language_model.model.layers.34.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1280
+ "language_model.model.layers.34.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1281
+ "language_model.model.layers.34.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1282
+ "language_model.model.layers.34.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1283
+ "language_model.model.layers.34.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1284
+ "language_model.model.layers.34.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1285
+ "language_model.model.layers.34.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1286
+ "language_model.model.layers.34.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1287
+ "language_model.model.layers.34.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1288
+ "language_model.model.layers.34.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1289
+ "language_model.model.layers.34.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1290
+ "language_model.model.layers.34.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1291
+ "language_model.model.layers.34.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1292
+ "language_model.model.layers.34.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1293
+ "language_model.model.layers.34.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1294
+ "language_model.model.layers.34.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1295
+ "language_model.model.layers.34.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1296
+ "language_model.model.layers.34.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1297
+ "language_model.model.layers.34.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1298
+ "language_model.model.layers.34.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1299
+ "language_model.model.layers.34.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1300
+ "language_model.model.layers.34.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1301
+ "language_model.model.layers.34.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1302
+ "language_model.model.layers.34.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1303
+ "language_model.model.layers.34.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1304
+ "language_model.model.layers.34.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1305
+ "language_model.model.layers.34.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1306
+ "language_model.model.layers.34.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1307
+ "language_model.model.layers.34.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1308
+ "language_model.model.layers.35.input_layernorm.weight": "model-00001-of-00002.safetensors",
1309
+ "language_model.model.layers.35.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1310
+ "language_model.model.layers.35.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1311
+ "language_model.model.layers.35.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1312
+ "language_model.model.layers.35.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1313
+ "language_model.model.layers.35.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1314
+ "language_model.model.layers.35.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1315
+ "language_model.model.layers.35.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1316
+ "language_model.model.layers.35.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1317
+ "language_model.model.layers.35.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1318
+ "language_model.model.layers.35.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1319
+ "language_model.model.layers.35.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1320
+ "language_model.model.layers.35.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1321
+ "language_model.model.layers.35.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1322
+ "language_model.model.layers.35.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1323
+ "language_model.model.layers.35.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1324
+ "language_model.model.layers.35.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1325
+ "language_model.model.layers.35.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1326
+ "language_model.model.layers.35.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1327
+ "language_model.model.layers.35.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1328
+ "language_model.model.layers.35.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1329
+ "language_model.model.layers.35.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1330
+ "language_model.model.layers.35.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1331
+ "language_model.model.layers.35.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1332
+ "language_model.model.layers.35.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1333
+ "language_model.model.layers.35.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1334
+ "language_model.model.layers.35.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1335
+ "language_model.model.layers.35.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1336
+ "language_model.model.layers.35.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1337
+ "language_model.model.layers.35.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1338
+ "language_model.model.layers.35.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1339
+ "language_model.model.layers.35.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1340
+ "language_model.model.layers.35.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1341
+ "language_model.model.layers.35.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1342
+ "language_model.model.layers.35.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1343
+ "language_model.model.layers.35.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1344
+ "language_model.model.layers.35.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1345
+ "language_model.model.layers.35.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1346
+ "language_model.model.layers.35.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1347
+ "language_model.model.layers.35.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1348
+ "language_model.model.layers.35.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1349
+ "language_model.model.layers.35.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1350
+ "language_model.model.layers.35.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1351
+ "language_model.model.layers.35.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1352
+ "language_model.model.layers.36.input_layernorm.weight": "model-00001-of-00002.safetensors",
1353
+ "language_model.model.layers.36.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1354
+ "language_model.model.layers.36.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1355
+ "language_model.model.layers.36.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1356
+ "language_model.model.layers.36.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1357
+ "language_model.model.layers.36.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1358
+ "language_model.model.layers.36.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1359
+ "language_model.model.layers.36.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1360
+ "language_model.model.layers.36.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1361
+ "language_model.model.layers.36.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1362
+ "language_model.model.layers.36.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1363
+ "language_model.model.layers.36.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1364
+ "language_model.model.layers.36.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1365
+ "language_model.model.layers.36.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1366
+ "language_model.model.layers.36.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1367
+ "language_model.model.layers.36.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1368
+ "language_model.model.layers.36.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1369
+ "language_model.model.layers.36.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1370
+ "language_model.model.layers.36.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1371
+ "language_model.model.layers.36.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1372
+ "language_model.model.layers.36.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1373
+ "language_model.model.layers.36.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1374
+ "language_model.model.layers.36.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1375
+ "language_model.model.layers.36.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1376
+ "language_model.model.layers.36.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1377
+ "language_model.model.layers.36.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1378
+ "language_model.model.layers.36.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1379
+ "language_model.model.layers.36.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1380
+ "language_model.model.layers.36.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1381
+ "language_model.model.layers.36.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1382
+ "language_model.model.layers.36.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1383
+ "language_model.model.layers.36.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1384
+ "language_model.model.layers.36.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1385
+ "language_model.model.layers.36.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1386
+ "language_model.model.layers.36.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1387
+ "language_model.model.layers.36.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1388
+ "language_model.model.layers.36.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1389
+ "language_model.model.layers.36.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1390
+ "language_model.model.layers.36.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1391
+ "language_model.model.layers.36.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1392
+ "language_model.model.layers.36.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1393
+ "language_model.model.layers.36.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1394
+ "language_model.model.layers.36.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1395
+ "language_model.model.layers.36.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1396
+ "language_model.model.layers.37.input_layernorm.weight": "model-00001-of-00002.safetensors",
1397
+ "language_model.model.layers.37.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1398
+ "language_model.model.layers.37.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1399
+ "language_model.model.layers.37.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1400
+ "language_model.model.layers.37.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1401
+ "language_model.model.layers.37.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1402
+ "language_model.model.layers.37.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1403
+ "language_model.model.layers.37.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1404
+ "language_model.model.layers.37.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1405
+ "language_model.model.layers.37.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1406
+ "language_model.model.layers.37.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1407
+ "language_model.model.layers.37.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1408
+ "language_model.model.layers.37.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1409
+ "language_model.model.layers.37.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1410
+ "language_model.model.layers.37.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1411
+ "language_model.model.layers.37.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1412
+ "language_model.model.layers.37.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1413
+ "language_model.model.layers.37.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1414
+ "language_model.model.layers.37.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1415
+ "language_model.model.layers.37.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1416
+ "language_model.model.layers.37.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1417
+ "language_model.model.layers.37.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1418
+ "language_model.model.layers.37.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1419
+ "language_model.model.layers.37.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1420
+ "language_model.model.layers.37.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1421
+ "language_model.model.layers.37.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1422
+ "language_model.model.layers.37.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1423
+ "language_model.model.layers.37.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1424
+ "language_model.model.layers.37.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1425
+ "language_model.model.layers.37.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1426
+ "language_model.model.layers.37.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1427
+ "language_model.model.layers.37.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1428
+ "language_model.model.layers.37.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1429
+ "language_model.model.layers.37.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1430
+ "language_model.model.layers.37.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1431
+ "language_model.model.layers.37.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1432
+ "language_model.model.layers.37.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1433
+ "language_model.model.layers.37.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1434
+ "language_model.model.layers.37.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1435
+ "language_model.model.layers.37.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1436
+ "language_model.model.layers.37.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1437
+ "language_model.model.layers.37.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1438
+ "language_model.model.layers.37.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1439
+ "language_model.model.layers.37.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1440
+ "language_model.model.layers.38.cross_attn.k_norm.weight": "model-00001-of-00002.safetensors",
1441
+ "language_model.model.layers.38.cross_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1442
+ "language_model.model.layers.38.cross_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1443
+ "language_model.model.layers.38.cross_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1444
+ "language_model.model.layers.38.cross_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1445
+ "language_model.model.layers.38.cross_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1446
+ "language_model.model.layers.38.cross_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1447
+ "language_model.model.layers.38.cross_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1448
+ "language_model.model.layers.38.cross_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1449
+ "language_model.model.layers.38.cross_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1450
+ "language_model.model.layers.38.cross_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1451
+ "language_model.model.layers.38.cross_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1452
+ "language_model.model.layers.38.cross_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1453
+ "language_model.model.layers.38.cross_attn.q_norm.weight": "model-00001-of-00002.safetensors",
1454
+ "language_model.model.layers.38.cross_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1455
+ "language_model.model.layers.38.cross_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1456
+ "language_model.model.layers.38.cross_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1457
+ "language_model.model.layers.38.cross_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1458
+ "language_model.model.layers.38.cross_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1459
+ "language_model.model.layers.38.cross_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1460
+ "language_model.model.layers.38.cross_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1461
+ "language_model.model.layers.38.cross_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1462
+ "language_model.model.layers.38.cross_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1463
+ "language_model.model.layers.38.cross_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1464
+ "language_model.model.layers.38.cross_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1465
+ "language_model.model.layers.38.cross_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1466
+ "language_model.model.layers.38.cross_attn_attn_gate": "model-00001-of-00002.safetensors",
1467
+ "language_model.model.layers.38.cross_attn_mlp_gate": "model-00001-of-00002.safetensors",
1468
+ "language_model.model.layers.38.input_layernorm.weight": "model-00001-of-00002.safetensors",
1469
+ "language_model.model.layers.38.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1470
+ "language_model.model.layers.38.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1471
+ "language_model.model.layers.38.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1472
+ "language_model.model.layers.38.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1473
+ "language_model.model.layers.38.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1474
+ "language_model.model.layers.38.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1475
+ "language_model.model.layers.38.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1476
+ "language_model.model.layers.38.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1477
+ "language_model.model.layers.38.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1478
+ "language_model.model.layers.38.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1479
+ "language_model.model.layers.38.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1480
+ "language_model.model.layers.38.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1481
+ "language_model.model.layers.38.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1482
+ "language_model.model.layers.38.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1483
+ "language_model.model.layers.38.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1484
+ "language_model.model.layers.38.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1485
+ "language_model.model.layers.38.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1486
+ "language_model.model.layers.38.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1487
+ "language_model.model.layers.38.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1488
+ "language_model.model.layers.39.input_layernorm.weight": "model-00001-of-00002.safetensors",
1489
+ "language_model.model.layers.39.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1490
+ "language_model.model.layers.39.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
1491
+ "language_model.model.layers.39.mlp.down_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1492
+ "language_model.model.layers.39.mlp.down_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1493
+ "language_model.model.layers.39.mlp.down_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1494
+ "language_model.model.layers.39.mlp.down_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1495
+ "language_model.model.layers.39.mlp.gate_proj.weight": "model-00001-of-00002.safetensors",
1496
+ "language_model.model.layers.39.mlp.gate_proj.weight.absmax": "model-00001-of-00002.safetensors",
1497
+ "language_model.model.layers.39.mlp.gate_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1498
+ "language_model.model.layers.39.mlp.gate_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1499
+ "language_model.model.layers.39.mlp.gate_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1500
+ "language_model.model.layers.39.mlp.gate_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1501
+ "language_model.model.layers.39.mlp.up_proj.weight": "model-00001-of-00002.safetensors",
1502
+ "language_model.model.layers.39.mlp.up_proj.weight.absmax": "model-00001-of-00002.safetensors",
1503
+ "language_model.model.layers.39.mlp.up_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1504
+ "language_model.model.layers.39.mlp.up_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1505
+ "language_model.model.layers.39.mlp.up_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1506
+ "language_model.model.layers.39.mlp.up_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1507
+ "language_model.model.layers.39.post_attention_layernorm.weight": "model-00001-of-00002.safetensors",
1508
+ "language_model.model.layers.39.self_attn.k_proj.weight": "model-00001-of-00002.safetensors",
1509
+ "language_model.model.layers.39.self_attn.k_proj.weight.absmax": "model-00001-of-00002.safetensors",
1510
+ "language_model.model.layers.39.self_attn.k_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1511
+ "language_model.model.layers.39.self_attn.k_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1512
+ "language_model.model.layers.39.self_attn.k_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1513
+ "language_model.model.layers.39.self_attn.k_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1514
+ "language_model.model.layers.39.self_attn.o_proj.weight": "model-00001-of-00002.safetensors",
1515
+ "language_model.model.layers.39.self_attn.o_proj.weight.absmax": "model-00001-of-00002.safetensors",
1516
+ "language_model.model.layers.39.self_attn.o_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1517
+ "language_model.model.layers.39.self_attn.o_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1518
+ "language_model.model.layers.39.self_attn.o_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1519
+ "language_model.model.layers.39.self_attn.o_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1520
+ "language_model.model.layers.39.self_attn.q_proj.weight": "model-00001-of-00002.safetensors",
1521
+ "language_model.model.layers.39.self_attn.q_proj.weight.absmax": "model-00001-of-00002.safetensors",
1522
+ "language_model.model.layers.39.self_attn.q_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1523
+ "language_model.model.layers.39.self_attn.q_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1524
+ "language_model.model.layers.39.self_attn.q_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1525
+ "language_model.model.layers.39.self_attn.q_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1526
+ "language_model.model.layers.39.self_attn.v_proj.weight": "model-00001-of-00002.safetensors",
1527
+ "language_model.model.layers.39.self_attn.v_proj.weight.absmax": "model-00001-of-00002.safetensors",
1528
+ "language_model.model.layers.39.self_attn.v_proj.weight.nested_absmax": "model-00001-of-00002.safetensors",
1529
+ "language_model.model.layers.39.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1530
+ "language_model.model.layers.39.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1531
+ "language_model.model.layers.39.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1532
  "language_model.model.layers.4.input_layernorm.weight": "model-00001-of-00002.safetensors",
1533
  "language_model.model.layers.4.mlp.down_proj.weight": "model-00001-of-00002.safetensors",
1534
  "language_model.model.layers.4.mlp.down_proj.weight.absmax": "model-00001-of-00002.safetensors",
 
1797
  "language_model.model.layers.9.self_attn.v_proj.weight.nested_quant_map": "model-00001-of-00002.safetensors",
1798
  "language_model.model.layers.9.self_attn.v_proj.weight.quant_map": "model-00001-of-00002.safetensors",
1799
  "language_model.model.layers.9.self_attn.v_proj.weight.quant_state.bitsandbytes__nf4": "model-00001-of-00002.safetensors",
1800
+ "language_model.model.norm.weight": "model-00001-of-00002.safetensors",
1801
  "multi_modal_projector.bias": "model-00002-of-00002.safetensors",
1802
  "multi_modal_projector.weight": "model-00002-of-00002.safetensors",
1803
  "multi_modal_projector.weight.absmax": "model-00002-of-00002.safetensors",