diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..1daabf81f8f1aa095be25abdcc57e56ecd1fd5a6 --- /dev/null +++ b/config.json @@ -0,0 +1,28 @@ +{ + "_name_or_path": "./megameditron", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 4096, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 140, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.37.2", + "use_cache": true, + "vocab_size": 32017 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..69b7806611a4865cd48c3e991dbd7d8312e0c5d3 --- /dev/null +++ b/generation_config.json @@ -0,0 +1,6 @@ +{ + "_from_model_config": true, + "bos_token_id": 1, + "eos_token_id": 2, + "transformers_version": "4.37.2" +} diff --git a/model-00001-of-00106.safetensors b/model-00001-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..33166b0c91237e68a8544fc1e8b67997c0fb38cf --- /dev/null +++ b/model-00001-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e44f30dc0a92bc6574428303e23a6c0964174ee94b1e701fa42f119eddf2110c +size 4807296496 diff --git a/model-00002-of-00106.safetensors b/model-00002-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b510f7099720d7036402d438b0447d8623f5580d --- /dev/null +++ b/model-00002-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f371c9797b49a1818be31f13189b21894be17abc77eb68a6c7dacb4f8e9704b2 +size 4630578440 diff --git a/model-00003-of-00106.safetensors b/model-00003-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..50e0eb58551956bf348807b6be46f4eaf513a985 --- /dev/null +++ b/model-00003-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8f5f1edd281d9775b494a113540cb48d6841569c03a96cb17e253e8a8d1f1593 +size 4362142864 diff --git a/model-00004-of-00106.safetensors b/model-00004-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bcf4195654cd184f90efa725e4690623542be811 --- /dev/null +++ b/model-00004-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2132e1459c61cd88cd2b4c312e1cb875bf5243465b8a4efa6b9cc3b3bfc02cec +size 4966188864 diff --git a/model-00005-of-00106.safetensors b/model-00005-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ce94951bbcca1f06492694fa6f76fe52932e990 --- /dev/null +++ b/model-00005-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb6e2477e9b0b7132ab5cd83a99a3704a27e23ca754ae0be5df32b6ddbfcf441 +size 4362142864 diff --git a/model-00006-of-00106.safetensors b/model-00006-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..20b7fe7943a90745aa6afa48f1e2d498d4d52b1d --- /dev/null +++ b/model-00006-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:275616391af8a2683efbe3869bad28bb5a7e5c64b018f1f1914112d7769903a9 +size 4362142864 diff --git a/model-00007-of-00106.safetensors b/model-00007-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d63f809f9752cd95db396dc7c71f7736cf1be7d7 --- /dev/null +++ b/model-00007-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d57593a90e8ab7b3773462e08a2d03a20a6368e8424738f61b2bdf4f64037753 +size 4966188864 diff --git a/model-00008-of-00106.safetensors b/model-00008-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ea207535b339b64c13ed658bd0b3fc40feb3544c --- /dev/null +++ b/model-00008-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01c176dd13d24cce03d915ba907fdfe32af02054f7675e130610a4a50f057dc5 +size 4362142880 diff --git a/model-00009-of-00106.safetensors b/model-00009-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0e7ff137d14c7be0ef9993495c703b4097cd4a2f --- /dev/null +++ b/model-00009-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28db210f73f2188bab3828c467e1ba8d141109443378c3a29d1ec140cee24d9d +size 4362142872 diff --git a/model-00010-of-00106.safetensors b/model-00010-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..504b108e144f2bfa3d12e455f0a5319dbdd7da2f --- /dev/null +++ b/model-00010-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2d12869ec39c315347caa7afbca5586788d4b95f32585925ba0be7a96adaf2bd +size 4966188880 diff --git a/model-00011-of-00106.safetensors b/model-00011-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..16b650de386d5461262bdf8a79c9bc6b38f63128 --- /dev/null +++ b/model-00011-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1487a550365f076689e3659b67466dbc2b322b477957f37abc6b62b2486a42c7 +size 4362142872 diff --git a/model-00012-of-00106.safetensors b/model-00012-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..61a613c9324a64b23baf435765342ad48ff31119 --- /dev/null +++ b/model-00012-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1804d2848dc8898194ef2d25056ddffc6a86bff2c210cbeb78ffc2b2128dcd78 +size 4362142872 diff --git a/model-00013-of-00106.safetensors b/model-00013-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d203b40c3738329d323228460567f999437baed6 --- /dev/null +++ b/model-00013-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf7f8ac8e8a0a32ecff1124bf0513e3a3b4bfd680b4c7d5a5f27c75f6aedc267 +size 4966188880 diff --git a/model-00014-of-00106.safetensors b/model-00014-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..acd1f1d1368f23f064e8bb4de796adb2bc437204 --- /dev/null +++ b/model-00014-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5ef8c4db30c35951ae16df352286cc47594d12b8cf0b4afc72989ac8bfaf2b50 +size 4362142872 diff --git a/model-00015-of-00106.safetensors b/model-00015-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..464194a6fe4fc623c080b9ebfafc99fd4f65c23e --- /dev/null +++ b/model-00015-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cbf5c961404fd9b659d6fd7ed2c3fe31b745e1fb1c2dfd2627789fbaf7066271 +size 4362142872 diff --git a/model-00016-of-00106.safetensors b/model-00016-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7db51d4e4eeab763d383d2bbccdb8ecf3d47b57f --- /dev/null +++ b/model-00016-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3e1e018e77c90d7315f9ca5adc896d042c624d9d7cdd4545e77b119c960955a1 +size 4966188880 diff --git a/model-00017-of-00106.safetensors b/model-00017-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ddf3dccdaf51b969b46d1d147e72db86754eebae --- /dev/null +++ b/model-00017-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:66f17b39749a12195242e6fc76058c2ce681560c31d2b61cdf01d618440e60b8 +size 4362142872 diff --git a/model-00018-of-00106.safetensors b/model-00018-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..307973dc2d567c704bb449cc2ab1f72afbf62852 --- /dev/null +++ b/model-00018-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6e7727d00bd5306de377d8d697e119ea62c2a53770f6fb54baf791fcdd6457f +size 4362142872 diff --git a/model-00019-of-00106.safetensors b/model-00019-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..11763c6a2b1f0cc1463fe4c496212388f550553b --- /dev/null +++ b/model-00019-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3f980f2724b5efef11aba2d3c2e21f376360e68a4f41c85a3a8f14c57b899a98 +size 4966188880 diff --git a/model-00020-of-00106.safetensors b/model-00020-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a278ded880f4c1958ceb44df55add8d36d9b8376 --- /dev/null +++ b/model-00020-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df6d28fe1b87f35581a393ed23a9a7960137c586363a43fe134a6f802f3e9919 +size 4362142872 diff --git a/model-00021-of-00106.safetensors b/model-00021-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7613268634848d68f11e5268000c6ebea59da51f --- /dev/null +++ b/model-00021-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:85371cddd82e6459d7b1cbba1237e0b414539dfe3522882179e5af0abd64ccd6 +size 4362142872 diff --git a/model-00022-of-00106.safetensors b/model-00022-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fb06be44ab8bba4cf7fdf9f72ec74b09b5ca92bd --- /dev/null +++ b/model-00022-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:959127991902a94a91669e30363c90a67810a52e3142ce5f7a2f908964c73bf0 +size 4966188880 diff --git a/model-00023-of-00106.safetensors b/model-00023-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f4bb4e381c91aa498b25b69be5256a32d07f4b58 --- /dev/null +++ b/model-00023-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:278ae88f1c61df761593d4a3cce286b8be4cce0e90a37e5ab3c78b016ac402a4 +size 4362142872 diff --git a/model-00024-of-00106.safetensors b/model-00024-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3b4798db55766b5e912fe9de6c76665589f02ad3 --- /dev/null +++ b/model-00024-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fc90c062ac66b626383f8eb13655054e6d8197771c7063cab579babd82a484c4 +size 4362142872 diff --git a/model-00025-of-00106.safetensors b/model-00025-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..095c81c44573d75e3c683a30e5cc6870b9f6e36d --- /dev/null +++ b/model-00025-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b2235386f1e48e03ed8a8c5d920c04931419bd4b765a31a1fa05e002a03c556 +size 4966188880 diff --git a/model-00026-of-00106.safetensors b/model-00026-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3d603a980fb3b4dc7c2591375641e3da2a865877 --- /dev/null +++ b/model-00026-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:569fb6ca087dc124de13c19d5f5027f4aacf1104e330ba9796307741b26925c3 +size 4362142872 diff --git a/model-00027-of-00106.safetensors b/model-00027-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3fbfaadb8e504a265bf7610a2ce66953d707ea6a --- /dev/null +++ b/model-00027-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e50faf290f8fde5b4a4b04f78228894a4f59b91372c37160fe04d0d157b324f +size 4362142872 diff --git a/model-00028-of-00106.safetensors b/model-00028-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4ba7b8a37acb9592d4c7741f00f14e004ce97f6a --- /dev/null +++ b/model-00028-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bd5554cb2e1b2c20b57ef323d03f8503182973e5bf73687b3f6ed81f88b259e2 +size 4966188880 diff --git a/model-00029-of-00106.safetensors b/model-00029-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eda62ab1a838b046fc4d644797a3123b7d846d88 --- /dev/null +++ b/model-00029-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6a712f513bed6dddd8cde9b88485c5db25ef4ee35600867d851d92e76df86c0 +size 4362142872 diff --git a/model-00030-of-00106.safetensors b/model-00030-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1d8c036515d9ba727939465cdf32c216750a7842 --- /dev/null +++ b/model-00030-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:166c4281d86ec9683fc339a01bf537d42ef12a7b186ac14a5e0f7de798f3fdde +size 4362142872 diff --git a/model-00031-of-00106.safetensors b/model-00031-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c8bddafd38dcc897de7d8e219282108463a7279c --- /dev/null +++ b/model-00031-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d36da9926dd8868f30b9584b870fa068c4a174d7809f77cc5809aae56896390a +size 4966188880 diff --git a/model-00032-of-00106.safetensors b/model-00032-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a415d256f0e5f9124bd9f64989a191d35a16897 --- /dev/null +++ b/model-00032-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:70ed077b14fff9ecb57102bb6f7ea32b248b452670bf31456fde62a5886458ff +size 4362142872 diff --git a/model-00033-of-00106.safetensors b/model-00033-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4104b7e3e5865ab41bc7cd9298c6bbdf02954155 --- /dev/null +++ b/model-00033-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:01dcf19ef6fbb286fe2d868ef7b738c796982e7d3ba2163e71b26e8814a8a823 +size 4362142872 diff --git a/model-00034-of-00106.safetensors b/model-00034-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..40623bb40c33272a59ad17df0290dcc8573752bc --- /dev/null +++ b/model-00034-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1533a69d5910e3ec69df3349b8c3703703367af01fdff78b433be0df812f840 +size 4966188880 diff --git a/model-00035-of-00106.safetensors b/model-00035-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe8aff6aafc8e9e77e02b9008b7c2abfd866680a --- /dev/null +++ b/model-00035-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ae7bd608da1edf12dac3e684ad504632192543cd1f5b2b3a99c0fb291dbb241a +size 4362142872 diff --git a/model-00036-of-00106.safetensors b/model-00036-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..28340adc16d203b02b1a890234e107413bf2ff7f --- /dev/null +++ b/model-00036-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:37666359397e186f8905bdf4634ee01625b4f3e4a652b612441654164bba140d +size 4362142872 diff --git a/model-00037-of-00106.safetensors b/model-00037-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5369c75fc5bb6e625cbb44edc4d8339d3cb8a573 --- /dev/null +++ b/model-00037-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:308dc865f397c805a29813436bc87ab62182749640cba6b8f71deb5ab2ab91e7 +size 4966188880 diff --git a/model-00038-of-00106.safetensors b/model-00038-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..446da034436265df736da9ea2c1bd6095a68d6de --- /dev/null +++ b/model-00038-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8e82ecfee42184e7b57cd8a1d4e7ca676309e9d48c89aa15f2b809cc454f6972 +size 4362142872 diff --git a/model-00039-of-00106.safetensors b/model-00039-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c1fac7052e814b1b62fcf1faa5a050068bb1d8cd --- /dev/null +++ b/model-00039-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6281dede905574ce3856ad6b1d4d5d30cbd0fefe2ec4a09fb7e14f51c979a7b9 +size 4362142872 diff --git a/model-00040-of-00106.safetensors b/model-00040-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7d7e1ea2af6bda2462e8af187c9492937e227add --- /dev/null +++ b/model-00040-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11bc1ce583318a4cf9cf647386786f818ff6092fbeda1df269c11310f50917c4 +size 4966188880 diff --git a/model-00041-of-00106.safetensors b/model-00041-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9064ddca3870852846d2d27fc07b58536f24d409 --- /dev/null +++ b/model-00041-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b3c78e972b00a18357f9e5c3fddd0045bb1dff28120581a5c278d087c346499 +size 4362142872 diff --git a/model-00042-of-00106.safetensors b/model-00042-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1b159924bd87057289092fd1adb0c816980614ce --- /dev/null +++ b/model-00042-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28e71dd95cf65563c3d0be1870cf8c7d61cfda48a8b82d878046c385b3983e96 +size 4362142872 diff --git a/model-00043-of-00106.safetensors b/model-00043-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cc6d3241b139e61a56042dbbca1ae8e9fc77db0c --- /dev/null +++ b/model-00043-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c6e0acbfd05b56f2b2c2196f4189c86065ff759ef47e0484527ea1c660990b9e +size 4966188880 diff --git a/model-00044-of-00106.safetensors b/model-00044-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..566dfd6750ce89e91802db42fafe71b312675a3d --- /dev/null +++ b/model-00044-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d3e45974ce0bc071c085268af0a2b9e00745b53f7b7cebaed7e09a10f095ac6 +size 4362142872 diff --git a/model-00045-of-00106.safetensors b/model-00045-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..427accd9da33b71750016e9b723ddca46dd6743d --- /dev/null +++ b/model-00045-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:18a3021cf61db1e8aa5910886a52b7faae9b2563511778369bda21f63412ceff +size 4362142872 diff --git a/model-00046-of-00106.safetensors b/model-00046-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8120aebc9d18c1c4a24fb2d6cf3d87932d840172 --- /dev/null +++ b/model-00046-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8eef6c91dc333e5afd238bd0cb6f1bf3b75c4daab1dd5d3a6febc03b51b0f653 +size 4966188880 diff --git a/model-00047-of-00106.safetensors b/model-00047-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..72a935cab367dd8cc291daf82b46970385bcca7c --- /dev/null +++ b/model-00047-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:098845e4ee41cc8deb0f3ece1770c4a18dd7296dfde2e8deef668eb72853ba8e +size 4362142872 diff --git a/model-00048-of-00106.safetensors b/model-00048-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b6d466fdc056814d8831a71fee008f36e92cd6f7 --- /dev/null +++ b/model-00048-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ecafa9aa5a28f2b0f681c84025326c4ca6a11eb67a141fec86fc90665f843779 +size 4362142872 diff --git a/model-00049-of-00106.safetensors b/model-00049-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b045730363544e6c87c2822ed86db4d8c6567d88 --- /dev/null +++ b/model-00049-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2de18de7b84c98cc5a88b42ee0083e1ed5ac331a7e3cfb4ec7c2a67fd106101e +size 4966188880 diff --git a/model-00050-of-00106.safetensors b/model-00050-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..32b3db560ef69e744dc243b121203f01161242f9 --- /dev/null +++ b/model-00050-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a424dd5d20164b62090102009417f14af50b0fbf0f088bec04e31b5a70eda3eb +size 4362142872 diff --git a/model-00051-of-00106.safetensors b/model-00051-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..91eedb660e302cc38e508efeeb80be1f32a8eba8 --- /dev/null +++ b/model-00051-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e614cfe577a9d577f382dc32fef44ccaaf6c47f6db072d64be64702e30675cc +size 4362142872 diff --git a/model-00052-of-00106.safetensors b/model-00052-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..734d6ec06ccbc311c6cb4ae39974fa875129e2c0 --- /dev/null +++ b/model-00052-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08812c4d717df1e9605c807a126bc82025932fc0413736fad8915591c47fc38b +size 4966188880 diff --git a/model-00053-of-00106.safetensors b/model-00053-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8673f942350aa5eecb30c027e908421161d52d09 --- /dev/null +++ b/model-00053-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8c1544b3ea3c9eb3138c832c128127eb88da63f0ada7e18c27141c9062c6916 +size 4362142872 diff --git a/model-00054-of-00106.safetensors b/model-00054-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7fdd3bc6f1101969081d10ca814b27c7a0611c97 --- /dev/null +++ b/model-00054-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0d8c71d183a835d95aa5a4da78120c423be9e69553226fc403581b3e39b66f9 +size 4362142872 diff --git a/model-00055-of-00106.safetensors b/model-00055-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..58bde545f543d0358c8b7311af40703ca926e99a --- /dev/null +++ b/model-00055-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:41eff8a9f4f45426a9ab5ee4c566f26d0652fb562658e277a99a598d4d6489fb +size 4966188880 diff --git a/model-00056-of-00106.safetensors b/model-00056-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ed5e1b819ca07b2028ce9bf5839d86c13b8c232 --- /dev/null +++ b/model-00056-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8ee9fb375f5dce92a2f33d1c4ca498d7a222946f2a7d7f89a50cb9c127d80a20 +size 4362142872 diff --git a/model-00057-of-00106.safetensors b/model-00057-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4b7aa8e0ead5b929a85928d44cb929a1e9b8c42 --- /dev/null +++ b/model-00057-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:83b758deceac0e0673f096f7cc6239d6fe80234838162020ed8ceeadfd5255c3 +size 4362142872 diff --git a/model-00058-of-00106.safetensors b/model-00058-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..43b3da653add34e34a8280b5b362d7d6e87a77bc --- /dev/null +++ b/model-00058-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52ced676b61fb4ec6360b4fcd6de9b88831aa5fb9f3454560e79ecb69bbd7192 +size 4966188880 diff --git a/model-00059-of-00106.safetensors b/model-00059-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..02fc9b0756f3eeb6c4233e5ea95dd4b877134ad8 --- /dev/null +++ b/model-00059-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a8f35275375af18220bf137a33cc361900d9c0701f602d288209ab4101b36c23 +size 4362142872 diff --git a/model-00060-of-00106.safetensors b/model-00060-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8dcd60371f0f04a4ba361a5bc59d5d5ac2a1e19d --- /dev/null +++ b/model-00060-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:229ab1d252983df0a1987cb057b3bc0304fc470282f89abf747d2aa72a2347c7 +size 4362142872 diff --git a/model-00061-of-00106.safetensors b/model-00061-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6df27faa2b5ce0700dddad11c31ae3dbcb752ac6 --- /dev/null +++ b/model-00061-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e136dfe309067bc704d7eb9295b5843875732d8a76edf429fec64205c63b00f6 +size 4966188880 diff --git a/model-00062-of-00106.safetensors b/model-00062-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7981733b9f4a6b919475b87b3b7d7197349ecea3 --- /dev/null +++ b/model-00062-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:701629a21571832fdd58d90794655f5ff672159e5095b6504fb27a0deecf568b +size 4362142872 diff --git a/model-00063-of-00106.safetensors b/model-00063-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fd5a5dc3fa0f116ef58de2dffc70a4b273980293 --- /dev/null +++ b/model-00063-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bb56694f13308f5c61738ce855e2ca56a6b27ce4a71a60c568860932bc2d6fc +size 4362142872 diff --git a/model-00064-of-00106.safetensors b/model-00064-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5b38788c4d90a230ee6651fb1ebf947bc368fcf1 --- /dev/null +++ b/model-00064-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c33106480d46c5b5b60ab4c12a930b98a1b3b071de30202889dfc35cda4a36b1 +size 4966188880 diff --git a/model-00065-of-00106.safetensors b/model-00065-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..653ae6a08ddd6f5d7b408b718a098752051fbf9d --- /dev/null +++ b/model-00065-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:94f572ece4f3a12023f786f7906c1e0c554c2b4a498bb92d01e1a9f719b1de06 +size 4362142872 diff --git a/model-00066-of-00106.safetensors b/model-00066-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c7cc182d3a7ab09e638a44aebb789a053e9d487 --- /dev/null +++ b/model-00066-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:664a976ce8697bfbdfac9a7d50d2750291ccbcc587277792c1ec47a7fb88edf7 +size 4362142872 diff --git a/model-00067-of-00106.safetensors b/model-00067-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5150b4e4111bb74ee6616b8342b58080525cecfb --- /dev/null +++ b/model-00067-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:373799b869397bd7f04bf4bdab8c1a2c3383ebf480f53da868e64c0bb822e391 +size 4966188880 diff --git a/model-00068-of-00106.safetensors b/model-00068-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..73bc5615da0bf4baf05590e810fc0bc69bf72df8 --- /dev/null +++ b/model-00068-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:434c7b0fcea46ea41fc98d828cda880a20548e5d4be1d46d24ac1090e6e0b13b +size 4362142872 diff --git a/model-00069-of-00106.safetensors b/model-00069-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ce9caf3b11ec907194f0cc81347003549f562ad --- /dev/null +++ b/model-00069-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:408f6f18b0c801e4eb5e598e915c38ab269eb3cca43055e1180f3695e8eb17a4 +size 4362142872 diff --git a/model-00070-of-00106.safetensors b/model-00070-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c58ed717e8780e8bb47da5bd60bff8610856668f --- /dev/null +++ b/model-00070-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a4fb0eea6af180fdc633012353285bfef70c51dca74832d2bde05e9ab6870c56 +size 4966188880 diff --git a/model-00071-of-00106.safetensors b/model-00071-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9ce88bce132680a8ef7171c7f30027c579b3130e --- /dev/null +++ b/model-00071-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2e618159d7b8a8d29c4211630122d1571de5b50e9e6fa38215bf01be2368fdc9 +size 4362142872 diff --git a/model-00072-of-00106.safetensors b/model-00072-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4dc3dfa8c6386bbfdf95cc5ce6a0af584bd5f0ec --- /dev/null +++ b/model-00072-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:714a862aa969c010bcbd18a1ea6cfaa866abbea626b01337105ec32c14bf5cfa +size 4362142872 diff --git a/model-00073-of-00106.safetensors b/model-00073-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ff0395134c079dfb8fc3f8b9cc63b976702a421b --- /dev/null +++ b/model-00073-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4c9d3f4853381770123a6d475083189778aae48ef4c4aedb053cb233ed98b67 +size 4966188880 diff --git a/model-00074-of-00106.safetensors b/model-00074-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e69ebbc31adb40706b9d95910ab44c70655de5db --- /dev/null +++ b/model-00074-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d95c68d44ef9d28bc49eef120148a398083d891c257c0b66a2fdf3b196f6347a +size 4362142872 diff --git a/model-00075-of-00106.safetensors b/model-00075-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6b6a0684ae246250b7a7dfe49aaec639202703c4 --- /dev/null +++ b/model-00075-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a82be2810f5d2f71b2fdf6e4d4b259fbf8ccd423a9a90df55e9ced95543872d9 +size 4362142872 diff --git a/model-00076-of-00106.safetensors b/model-00076-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1fb559f74d7cf559b2a3c4d249a6f44d81d1f370 --- /dev/null +++ b/model-00076-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2006c89bd3f29645470ebf9bda09a7a0c71ee82b98257faa0b11e4fa1a98bc24 +size 4966188896 diff --git a/model-00077-of-00106.safetensors b/model-00077-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e3d0da585161fe7634e9513b9d7b69e65eb5dc85 --- /dev/null +++ b/model-00077-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:17c5cbc7105348c35dcc4585859425cbba32d1606ed27a80816c4d19c18058aa +size 4362142888 diff --git a/model-00078-of-00106.safetensors b/model-00078-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dbb9d154339c8722d7c57679ae090bec3702ab76 --- /dev/null +++ b/model-00078-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f333a9127d2c4bf5fe43c21850468887431a0cc5a89588954bec2b528bf73063 +size 4362142880 diff --git a/model-00079-of-00106.safetensors b/model-00079-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..055f2ed5b55f5480a5b202d3f4bfc2d7d4061065 --- /dev/null +++ b/model-00079-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:81925a68710b6c2d01a651834f875d32945af30879f8fc5958dbb387afc45162 +size 4966188896 diff --git a/model-00080-of-00106.safetensors b/model-00080-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ee774879cd3347f393c08481cda1838c210eb97 --- /dev/null +++ b/model-00080-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f290152b299b751f8f998c3972c3cbc937361dc5c885147ea3be42d43465a18 +size 4362142888 diff --git a/model-00081-of-00106.safetensors b/model-00081-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0fb2432ee4465b1bcc234cb8fe4ed5fccaa3863f --- /dev/null +++ b/model-00081-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f95a4572c12729f9a61d2c8834cd093728f75bd40903e83a9de9630f4feca761 +size 4362142880 diff --git a/model-00082-of-00106.safetensors b/model-00082-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cc9e6d469281aa74b2447dc87ecc09935e48590a --- /dev/null +++ b/model-00082-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:78b05a3f165eafce8faf5bb425badacd86e60d9f9212ef0b0ac6ea53a7a4c559 +size 4966188896 diff --git a/model-00083-of-00106.safetensors b/model-00083-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fe2e34002783daf126551db1786d6e5d9105b41a --- /dev/null +++ b/model-00083-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e3e5516f6a4fd3de816334a40579c24ccbc77ed28a3bfeff642de70e636596ed +size 4362142888 diff --git a/model-00084-of-00106.safetensors b/model-00084-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..419a0e9e2d4067367853e83749fb131a4036905b --- /dev/null +++ b/model-00084-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b4b915ef61a3934deb9201ca7aff381a9273b8c95d65fa422e772214af6c427c +size 4362142880 diff --git a/model-00085-of-00106.safetensors b/model-00085-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b4251611565c6817fb8c706d0058e24cd03f6aaf --- /dev/null +++ b/model-00085-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1103ae73e205cc57bb9d929aaf5c91f3d0e52c65bae3618022e53946cd42bbc7 +size 4966188896 diff --git a/model-00086-of-00106.safetensors b/model-00086-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..64921b28b7ec7890862faf0428477fda9e346fe8 --- /dev/null +++ b/model-00086-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec9689a044e7ce3008baa51bc7939fbc6ea821f7bacc464d2a32ec42666cc93a +size 4362142888 diff --git a/model-00087-of-00106.safetensors b/model-00087-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1d811f714ce690678827e0cb618baa52cdb4e9c6 --- /dev/null +++ b/model-00087-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b9fe1123e5e5e0feee4978a3a945d2ed95c1622b9a5d3fe9c4f1e03b4f2fbf2 +size 4362142880 diff --git a/model-00088-of-00106.safetensors b/model-00088-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e592a61c242d59970eaad718423b0a3a09315919 --- /dev/null +++ b/model-00088-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:31f3cbe92230c3faf09f76ba2c331b5acf5e20ea85143d9cdc54f3be7db28dd4 +size 4966188896 diff --git a/model-00089-of-00106.safetensors b/model-00089-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1292239ca6b0a18f6d7f6d57ea68121fb568e5f0 --- /dev/null +++ b/model-00089-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e08e8c329a19735fc2d63e4beec413d74c2509cbd3633a787bfeecfcdf07878b +size 4362142888 diff --git a/model-00090-of-00106.safetensors b/model-00090-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cfcf4c6150ec8c2df172e29792a731473c2f602a --- /dev/null +++ b/model-00090-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2709b32ab7b5dc00d17a0981556582a3aa78df09b0e7891d3868b90772d9366b +size 4362142880 diff --git a/model-00091-of-00106.safetensors b/model-00091-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cdf62abf98961c91ed7d8a8d97b6e115bc230cc2 --- /dev/null +++ b/model-00091-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:25941d4a3d86117895df623ecdc8efddf3e35c903b991efb8111a29315606b59 +size 4966188896 diff --git a/model-00092-of-00106.safetensors b/model-00092-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb7b59ebe0ad8a9ed2b0d78041724e49732f1e7a --- /dev/null +++ b/model-00092-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c1d3deaca9c0037fba87bd5f9cec816799cae009b224c64f3acafa5596db1a9e +size 4362142888 diff --git a/model-00093-of-00106.safetensors b/model-00093-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2928a0815e2e2498248593393aa0a205d4417770 --- /dev/null +++ b/model-00093-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73c31dcd8e3aee5e5be7a24b4933010422501c16b0950bc59af8cec1750c3ccc +size 4362142880 diff --git a/model-00094-of-00106.safetensors b/model-00094-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fc971ecc148d156fc82172094ce575b9e95e6a7e --- /dev/null +++ b/model-00094-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:890d5375c1b6e3cb4ab03a3566e8901dff91fb74ead136fbd527115a6419df4b +size 4966188896 diff --git a/model-00095-of-00106.safetensors b/model-00095-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cc173c30623d8e85a1cf597384541d1fabdb61e0 --- /dev/null +++ b/model-00095-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:3364e9ad68c056580d192785439d7caf45265aed9d570536fca9340b54fa7115 +size 4362142888 diff --git a/model-00096-of-00106.safetensors b/model-00096-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4339ac981b54803adccb3b0fae00154d3373b4f9 --- /dev/null +++ b/model-00096-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cf0143c10df725a6e5767796101b080a73b1c3c858f048a7ba1f648b1098ec0c +size 4362142880 diff --git a/model-00097-of-00106.safetensors b/model-00097-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2dd34dd96d0f367eb1910e74184333af4c48497e --- /dev/null +++ b/model-00097-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b027ebfe5ff46d916b788d3a7c9c159f28e6a9e82015336f189c502c2b526c3c +size 4966188896 diff --git a/model-00098-of-00106.safetensors b/model-00098-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8d95fcd84ccda32629e8be4af13b2e2802b8c878 --- /dev/null +++ b/model-00098-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:21155f7aa9179bbc7fd0f225b232d69fa1330ea8211ee61c838cf541ef7cd64b +size 4362142888 diff --git a/model-00099-of-00106.safetensors b/model-00099-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..48702eada1acf6728cd763900e21e9eb9f6ac5c8 --- /dev/null +++ b/model-00099-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99c07af902980cb8d554472e5281237cb5553e71957f092fe4d7e2b23ec18476 +size 4362142880 diff --git a/model-00100-of-00106.safetensors b/model-00100-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4e3d77270e22d653e7b5146487b83cd50a9174cd --- /dev/null +++ b/model-00100-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6a2ae4e0f89f5de413c72342b00d18906c4c6a1639b46a2adc1ffc3fd8f98bb9 +size 4966188896 diff --git a/model-00101-of-00106.safetensors b/model-00101-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a3bd0c382067ea400b406ef2dd557434c535f03a --- /dev/null +++ b/model-00101-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dd9de26f56f053717c89672c67fcc442c08b1c63b9ed6509df049978bd92a096 +size 4362142888 diff --git a/model-00102-of-00106.safetensors b/model-00102-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2af841593c88c072e364ac6dbb664e0707eeffd2 --- /dev/null +++ b/model-00102-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:16543c3d18de15116bf7a6b8aef8a1f7cc0335932407224c1b538c1de3502170 +size 4362142880 diff --git a/model-00103-of-00106.safetensors b/model-00103-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..094fa94158c0756e05c804ceab33606c1ee7c141 --- /dev/null +++ b/model-00103-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:de5dcc91d12791a7e76316c2a8d429d44f7a8eb3479a1888d98b37f343f5f071 +size 4966188896 diff --git a/model-00104-of-00106.safetensors b/model-00104-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..11d1c57a1e71a126106b2d6ba3de5297bbbff048 --- /dev/null +++ b/model-00104-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7c68f29b6523448c64bf55c99dadaba42d12fac7c75d605b51c64a2fbc0b3fa6 +size 4362142888 diff --git a/model-00105-of-00106.safetensors b/model-00105-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..05f4d903546aa9d2f01fa651a71b9f130d661eca --- /dev/null +++ b/model-00105-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e290ad3f7970ae6df0245ed8976c630921cb88b7744b0ef1cb7c4ff79be74c1 +size 4362142880 diff --git a/model-00106-of-00106.safetensors b/model-00106-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bddc5745b91384abb6352fa2c9782ba2e479d60f --- /dev/null +++ b/model-00106-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cfe9cd3dee566025937757a04120b1adba997596a87eafc671f6a220b2a52512 +size 1988756024 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..1a56da2660c0ec777e487395f13e2e8dc0e25267 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,1270 @@ +{ + "metadata": { + "total_size": 481264762880 + }, + "weight_map": { + "lm_head.weight": "model-00106-of-00106.safetensors", + "model.embed_tokens.weight": "model-00001-of-00106.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.100.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.gate_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.up_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.input_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.down_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.up_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.post_attention_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.102.input_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.down_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.post_attention_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.self_attn.k_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.o_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.q_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.v_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.103.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.gate_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.self_attn.k_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.o_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.q_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.v_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.104.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.gate_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.up_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.input_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.down_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.up_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.post_attention_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.106.input_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.down_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.post_attention_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.self_attn.k_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.o_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.q_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.v_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.107.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.gate_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.self_attn.k_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.o_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.q_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.v_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.108.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.gate_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.up_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.input_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.down_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.up_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.post_attention_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.110.input_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.down_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.post_attention_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.self_attn.k_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.o_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.q_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.v_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.111.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.gate_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.self_attn.k_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.o_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.q_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.v_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.112.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.gate_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.up_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.input_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.down_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.up_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.post_attention_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.114.input_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.down_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.post_attention_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.self_attn.k_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.o_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.q_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.v_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.115.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.gate_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.self_attn.k_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.o_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.q_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.v_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.116.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.gate_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.up_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.input_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.down_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.up_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.post_attention_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.118.input_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.down_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.post_attention_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.self_attn.k_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.o_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.q_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.v_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.119.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.gate_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.self_attn.k_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.o_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.q_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.v_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.120.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.gate_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.up_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.input_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.down_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.up_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.post_attention_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.122.input_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.down_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.post_attention_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.self_attn.k_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.o_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.q_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.v_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.123.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.gate_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.self_attn.k_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.o_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.q_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.v_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.124.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.gate_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.up_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.input_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.down_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.up_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.post_attention_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.126.input_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.down_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.post_attention_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.self_attn.k_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.o_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.q_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.v_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.127.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.gate_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.self_attn.k_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.o_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.q_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.v_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.128.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.gate_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.up_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.input_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.down_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.up_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.post_attention_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.130.input_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.down_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.post_attention_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.self_attn.k_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.o_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.q_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.v_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.131.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.gate_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.self_attn.k_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.o_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.q_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.v_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.132.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.gate_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.up_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.input_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.down_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.up_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.post_attention_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.134.input_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.down_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.post_attention_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.self_attn.k_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.o_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.q_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.v_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.135.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.gate_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.self_attn.k_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.o_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.q_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.v_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.136.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.gate_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.up_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.input_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.down_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.up_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.post_attention_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.138.input_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.down_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.post_attention_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.self_attn.k_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.o_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.q_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.v_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.139.input_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.down_proj.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.gate_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.post_attention_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.self_attn.k_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.o_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.q_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.v_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.62.input_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.64.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.input_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.66.input_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.68.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.input_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.70.input_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.72.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.input_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.74.input_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.76.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.input_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.78.input_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.80.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.input_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.82.input_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.83.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.84.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.input_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.86.input_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.87.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.88.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.gate_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.up_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.input_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.down_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.up_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.post_attention_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.90.input_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.down_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.post_attention_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.self_attn.k_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.o_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.q_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.v_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.91.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.gate_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.self_attn.k_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.o_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.q_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.v_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.92.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.gate_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.up_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.input_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.down_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.up_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.post_attention_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.94.input_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.down_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.post_attention_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.self_attn.k_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.o_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.q_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.v_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.95.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.gate_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.self_attn.k_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.o_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.q_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.v_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.96.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.gate_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.up_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.input_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.down_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.up_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.post_attention_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.98.input_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.down_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.post_attention_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.self_attn.k_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.o_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.q_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.v_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.99.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.gate_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.self_attn.k_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.o_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.q_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.v_proj.weight": "model-00075-of-00106.safetensors", + "model.norm.weight": "model-00106-of-00106.safetensors" + } +}