diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..8ce7bdf46c25517e6ea638ebc768c64fc086eb18 --- /dev/null +++ b/config.json @@ -0,0 +1,29 @@ +{ + "_name_or_path": "./megacarcoroni", + "architectures": [ + "LlamaForCausalLM" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 1, + "eos_token_id": 2, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 4096, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 140, + "num_key_value_heads": 8, + "pad_token_id": 0, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": null, + "rope_theta": 10000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.37.2", + "use_cache": false, + "vocab_size": 32000 +} diff --git a/generation_config.json b/generation_config.json new file mode 100644 index 0000000000000000000000000000000000000000..c20e2cf5ab6e5f02fa5086c642cc0c346e3c4ced --- /dev/null +++ b/generation_config.json @@ -0,0 +1,8 @@ +{ + "_from_model_config": true, + "bos_token_id": 1, + "eos_token_id": 2, + "pad_token_id": 0, + "transformers_version": "4.37.2", + "use_cache": false +} diff --git a/model-00001-of-00106.safetensors b/model-00001-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ce5e7bd43e55ce396ad946926c355feccf291afb --- /dev/null +++ b/model-00001-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b21385d70b303324eff225a6b6a8705975b805cb23ebe1fd18e75405d2846de4 +size 4806739440 diff --git a/model-00002-of-00106.safetensors b/model-00002-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2cb2e62655d0eb1f7f4224cbbe1062d089e97f09 --- /dev/null +++ b/model-00002-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9d2c8dcf3a734d0d7e73dae05b975197664be5229202875018d1f8ab23474aa8 +size 4630578440 diff --git a/model-00003-of-00106.safetensors b/model-00003-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..dcca3d6202b0574d0a96137781ff5dbff26d2871 --- /dev/null +++ b/model-00003-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f000c6dd16e76bcf456cdc84853be3c6d01fa131c6d7a7399183296be28ec90f +size 4362142864 diff --git a/model-00004-of-00106.safetensors b/model-00004-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b3b66ef63f1e55b77f2c9ab57e5879a26cf6a831 --- /dev/null +++ b/model-00004-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:67386c6a282a1defe3f3c9636f6b165b14e81abc96665db9e4de6d68a04eab5c +size 4966188864 diff --git a/model-00005-of-00106.safetensors b/model-00005-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6c31a501d025ac30801fb2110393cfcd50d42a52 --- /dev/null +++ b/model-00005-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a96516b9e430fa970a82fa415bedae34bd2c61f45d9c20218bb8064066d44add +size 4362142864 diff --git a/model-00006-of-00106.safetensors b/model-00006-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..69101b8bb1fb22f556b3d8a3a9a4c3887d624eb4 --- /dev/null +++ b/model-00006-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a1716a29f0aa0e9e8cce505d0f1fd06aff03e090d04ebe402d37531cd49a8f13 +size 4362142864 diff --git a/model-00007-of-00106.safetensors b/model-00007-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9303aa4b4200526c69c09094b80f886de5796e17 --- /dev/null +++ b/model-00007-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c9f305229739d3eef00d6df105adfd795cd6381307e887e20c1087fbcbbedc72 +size 4966188864 diff --git a/model-00008-of-00106.safetensors b/model-00008-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..913aed4f0ee9ea4d1e87e03caaf9adb4dd3227d5 --- /dev/null +++ b/model-00008-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:02ff2b92df7629fa797a2afc729d2bdedf1fd627c2a1eeadb6dc2475c8a5a4f0 +size 4362142880 diff --git a/model-00009-of-00106.safetensors b/model-00009-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..764b1517b7e46b66c586e436afd64f9cd4ff2965 --- /dev/null +++ b/model-00009-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c5d017e765d91e0e8d927504e7d75cb746a47f5fabe8c2e51d7f48af110a9220 +size 4362142872 diff --git a/model-00010-of-00106.safetensors b/model-00010-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c639976082fc5dad4ab5826fde14e8a4ddd776a4 --- /dev/null +++ b/model-00010-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d24b98baef4e5bb8b97646113b3595e3736b09bdae48200635fb0466c93e19ae +size 4966188880 diff --git a/model-00011-of-00106.safetensors b/model-00011-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..98c7173adaba658d63042a3dd214a066fa19680f --- /dev/null +++ b/model-00011-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a417e07496ddc82043773c786e11e1e5d5955d542b59f9ad66d0ee7a66e3c1e0 +size 4362142872 diff --git a/model-00012-of-00106.safetensors b/model-00012-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b4c98031e5d5ea64ca17773a529fdb9e0fa1fc47 --- /dev/null +++ b/model-00012-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a5ffac5372b71c252999d3f419ee635a6b7f962d8819dfe53e66ec4c398041f8 +size 4362142872 diff --git a/model-00013-of-00106.safetensors b/model-00013-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..eb6079355de6a9f19d52a9d690f3bac664ac1f23 --- /dev/null +++ b/model-00013-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:587adef026b7eec7f0553dfbd34fff62737f6e220e729b16325262b595149530 +size 4966188880 diff --git a/model-00014-of-00106.safetensors b/model-00014-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f4f58021b3b776e3c5b226e0d4e42bd97dba3b30 --- /dev/null +++ b/model-00014-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2ef65978f45eb6f4bf9e9d30cc0909c327110980b183efd4594411b09bd7f210 +size 4362142872 diff --git a/model-00015-of-00106.safetensors b/model-00015-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b4400d42176b856b651553bfd5627b5e56591f18 --- /dev/null +++ b/model-00015-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:531cbe73ae058d9a9941332853a6f35c95fcdce481328e48c333305a180cd143 +size 4362142872 diff --git a/model-00016-of-00106.safetensors b/model-00016-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6c17e0f2b41b87838de612a14cd1380e94566eb5 --- /dev/null +++ b/model-00016-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5c71c6f14cc4cda9cb27811748a811e4b7d24ce9195a9e9d4e83b247319b56fa +size 4966188880 diff --git a/model-00017-of-00106.safetensors b/model-00017-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3e45d7cdbbdf4b563c1e1c4409bedd4814d9c250 --- /dev/null +++ b/model-00017-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1a5c74e24d8967dc7ea1cf810b29cb8cd81e7ccbf1ad2785b063df776f84a0cd +size 4362142872 diff --git a/model-00018-of-00106.safetensors b/model-00018-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d6e88a0a1c15fb1973dd1f13e8d44832d18fa6bd --- /dev/null +++ b/model-00018-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e769b9d5c25540f063059451621d5dca8ff3ad2ff6be1034844f876ea49f89d1 +size 4362142872 diff --git a/model-00019-of-00106.safetensors b/model-00019-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0ba26e6786d161dfad5415f3b2f238d8c78f8f15 --- /dev/null +++ b/model-00019-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e2635f058a77897f09ca9b13b11b13dc4ac04c006c3f8f940b67b98f963387f3 +size 4966188880 diff --git a/model-00020-of-00106.safetensors b/model-00020-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a209fa43fa12fde7122c9b3fd48cba38288805a3 --- /dev/null +++ b/model-00020-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d822ebbdc83097720e2a7941218c31696c3f27921e198e63fd53b6e7042a16e1 +size 4362142872 diff --git a/model-00021-of-00106.safetensors b/model-00021-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..00046c198e2a7ee5515ca8a040897c0deca38b07 --- /dev/null +++ b/model-00021-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:51c54f792ee91def281219b3976d3f58009bb282e18537fd3127045ffdfe5499 +size 4362142872 diff --git a/model-00022-of-00106.safetensors b/model-00022-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..77621c3435b9899642b21866f816ec374466811a --- /dev/null +++ b/model-00022-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d4303fed0f736eca7917344106b2f14ea587ee74663b151feb5c3fec0d1fa246 +size 4966188880 diff --git a/model-00023-of-00106.safetensors b/model-00023-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c5d960b310be26d3f6738fb47f0ee457dc4cee7f --- /dev/null +++ b/model-00023-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6306da4afe04018ba971f2ac1fed9efa26d86ddd4acadb820a7e4db021dd48bb +size 4362142872 diff --git a/model-00024-of-00106.safetensors b/model-00024-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0279b543e8c7cc9a6330978e49551e00aa2e68f2 --- /dev/null +++ b/model-00024-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7a97dbd0a8416b96016f774259a422e0ef15f1484ff3ae4a1569aa6f350f6242 +size 4362142872 diff --git a/model-00025-of-00106.safetensors b/model-00025-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..839b9bf568d37f4c41d7d25ed34366801b74711d --- /dev/null +++ b/model-00025-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2fb5e4f97dd0d43e428e9e9bb6cbba50609ac33061d7532c7b3ec853b7efa31e +size 4966188880 diff --git a/model-00026-of-00106.safetensors b/model-00026-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..58e300aaa3fb07c5846efa96d1267a8fd2f1525b --- /dev/null +++ b/model-00026-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8094b9c97244819865d4624b59415d8ffc3769f38399b8cf9da0c283da49c62a +size 4362142872 diff --git a/model-00027-of-00106.safetensors b/model-00027-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..61f79151df1b04242a4459b8e112f6d756bf2458 --- /dev/null +++ b/model-00027-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ed2bca02e0193a2d52acb710cf40fbce0f410e8fadb0bb4187561a131948e6ed +size 4362142872 diff --git a/model-00028-of-00106.safetensors b/model-00028-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..05c1c08ce9f2aee3eb480221f32652a1c3e342ff --- /dev/null +++ b/model-00028-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:95948f02180a8778c585a79ced8b06894b65df27541f364581566a5b2a3e360b +size 4966188880 diff --git a/model-00029-of-00106.safetensors b/model-00029-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3554b136e8d6a4233ff64eb3f2417e2b0b5df2aa --- /dev/null +++ b/model-00029-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99c3bcf80561fad4f2eae54f728c57b5b35d09d44439a82adb53c650f9bdc9b1 +size 4362142872 diff --git a/model-00030-of-00106.safetensors b/model-00030-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b6a570aa676eda4811b0ca0255bd36a5d7971d25 --- /dev/null +++ b/model-00030-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f81828f963b9482f607b1ffa8f2d39acabd29a24b460f7a842e1f0d0672a8db +size 4362142872 diff --git a/model-00031-of-00106.safetensors b/model-00031-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a26c3148a0d74dc73447e7822d520b0aadd2a40e --- /dev/null +++ b/model-00031-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:030e175150d0cb29de1cc1c9e24a91e58f35b76f3883b7195f22c504e5098b55 +size 4966188880 diff --git a/model-00032-of-00106.safetensors b/model-00032-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b755640b647007c5fbf423f883439c2856bb40dd --- /dev/null +++ b/model-00032-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:722b12e8fd3378b8925a00c0f1e0b5b6cdc18820dfda24e4dd55a0a32c2bea70 +size 4362142872 diff --git a/model-00033-of-00106.safetensors b/model-00033-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9bfd87d0b35dac1f4f4660f512b44daed4504736 --- /dev/null +++ b/model-00033-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75a2ef7a447d5e03996ca2d014d535297fd40aeabaecb31ffacfe677c9f10d7c +size 4362142872 diff --git a/model-00034-of-00106.safetensors b/model-00034-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2e6cf5ab6839ab1da6ed309d49442183d3a828d4 --- /dev/null +++ b/model-00034-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:34eb22eb6a2e98e8a7faaf2a65cf3a536beb5eeaaf53a8f55d71ba980c43abb6 +size 4966188880 diff --git a/model-00035-of-00106.safetensors b/model-00035-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b35569ba85edd720aa27801912d1ce0c679dc85b --- /dev/null +++ b/model-00035-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb818d41c370749dbb292b56ef05e5e546b46a17ac183fe7e1b71a42aee46fdd +size 4362142872 diff --git a/model-00036-of-00106.safetensors b/model-00036-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..254a8f36cb469cda1a4229036af2b9269803ef0d --- /dev/null +++ b/model-00036-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7f21134cc180b8dfb351f0e9efb05e425d60bfcd1e0225ffe3f5ea2287c3e336 +size 4362142872 diff --git a/model-00037-of-00106.safetensors b/model-00037-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3a9f313b8157c0d2ea735fdef79a3c020e082b24 --- /dev/null +++ b/model-00037-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8dd7c663b2f4418e4f19661796186026035eb514cc05361668abab99ba8c6b9 +size 4966188880 diff --git a/model-00038-of-00106.safetensors b/model-00038-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..99140e8cf0ada8471703779b582cdd6cf8378089 --- /dev/null +++ b/model-00038-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e7624ee696b22aa2ef62cc74b3a24a35a5a5cf4941cd10ae925fb83493750bd6 +size 4362142872 diff --git a/model-00039-of-00106.safetensors b/model-00039-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..29dc334a7c0c854871516de566e4942d05fdde24 --- /dev/null +++ b/model-00039-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f8996cf449c3096af63066a0dae3cd5b2636e74afcfd07c510259de7ffc35184 +size 4362142872 diff --git a/model-00040-of-00106.safetensors b/model-00040-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..29b8e79aaf0b56bd9c02d32dd0df310a3092f318 --- /dev/null +++ b/model-00040-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4176c897ae65fb2d632b00048e7fe30e2d60e5e1284360f67af449d0d15d7269 +size 4966188880 diff --git a/model-00041-of-00106.safetensors b/model-00041-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0dd584f816737dc513aafe8faba95de3b9ba5a0d --- /dev/null +++ b/model-00041-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5b52e28388227b546e024693c9622dc29f10fbf344002f3f1c7602d6cbe79a66 +size 4362142872 diff --git a/model-00042-of-00106.safetensors b/model-00042-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5262051d1e10858e96b9188d899ac6c1ab62487f --- /dev/null +++ b/model-00042-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eb1087ec8075176256dd34381fbcd66aae8bdf13eb1b825a9c24c864ac5a8793 +size 4362142872 diff --git a/model-00043-of-00106.safetensors b/model-00043-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d48a156d881fe240c6fcba010e51ccb0d1926c1 --- /dev/null +++ b/model-00043-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:11c94da85047d84a480ecf5888646189a1c0322c05c3e028be12111603eb1344 +size 4966188880 diff --git a/model-00044-of-00106.safetensors b/model-00044-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0b6334f8bcccb7f89cf60cfec69d3d0c6c888c1c --- /dev/null +++ b/model-00044-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e6d579b80f3459bab63df71cb5a8c8f0349164f44e582e57d0a8f738870025a6 +size 4362142872 diff --git a/model-00045-of-00106.safetensors b/model-00045-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f0284ad379a7aec8ff6674377c9bd97e8956eb8a --- /dev/null +++ b/model-00045-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4bda6b524e00581f7a24ccf7a2a94260a6107ba82dba4273a161e663f47370cf +size 4362142872 diff --git a/model-00046-of-00106.safetensors b/model-00046-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f6e39b143c4a945c2d9bebfaffd9cde855a716ff --- /dev/null +++ b/model-00046-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6dab034260ae166d7c17e4dcdf1d535ba50efbfd4fc55706342ce6995c06ffc7 +size 4966188880 diff --git a/model-00047-of-00106.safetensors b/model-00047-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..74618eb054683ffbaa9585c26b22f617405ffff2 --- /dev/null +++ b/model-00047-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c655c806521fc6bbf5992ebcc0c7d31894af558ab29eae0dc08e40087cbd4e2 +size 4362142872 diff --git a/model-00048-of-00106.safetensors b/model-00048-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..138aab0002f2b86b191188fc3f2c5bd3102d8413 --- /dev/null +++ b/model-00048-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:366b0303c30c1059ad42c585e7ba5b3e1e37ddf6e0ca2635a77f6b8a73ad3b6b +size 4362142872 diff --git a/model-00049-of-00106.safetensors b/model-00049-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f2eecbb2ad0272315be9257b398b3f8b36288fbd --- /dev/null +++ b/model-00049-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:26e441eeadbb392be715ae6cc74b3fe6824cc3c7331e25e87cb8dfec9ca61562 +size 4966188880 diff --git a/model-00050-of-00106.safetensors b/model-00050-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b97fbdb224640a92b43f13d5858c7aa641245775 --- /dev/null +++ b/model-00050-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:57a4e0098466b7de09ed3668d40495262e6f7873e2af7e9df1328f9405494d0f +size 4362142872 diff --git a/model-00051-of-00106.safetensors b/model-00051-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ef7d2663be5674736b19423a016dda423ea0c36f --- /dev/null +++ b/model-00051-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e8f6f94fe465bd6a28b2e79ce820b34982b93cb88e0bc0f9e4f2ed0e89cc5d8d +size 4362142872 diff --git a/model-00052-of-00106.safetensors b/model-00052-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1cee8daae5e4f6cb1e51f0c8aa7f1ebd7219a738 --- /dev/null +++ b/model-00052-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:486fbcf8c625e2d3c8615be346ac8af3c4e04ad200f0df2bd6d3c0fd315a51eb +size 4966188880 diff --git a/model-00053-of-00106.safetensors b/model-00053-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..197eacb42632c297fd77acbad24f842db13e63c6 --- /dev/null +++ b/model-00053-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4f188407942416c38d202943b4e72cafa6f46d9d71f9a5bd9231fb965648a0da +size 4362142872 diff --git a/model-00054-of-00106.safetensors b/model-00054-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a6eef0b5d5ca2adae06e7fc3d8ddeed572a429d7 --- /dev/null +++ b/model-00054-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:88d82fb1690dbde699dd1a88cbe519055b3fc95a848707363f010701161affe4 +size 4362142872 diff --git a/model-00055-of-00106.safetensors b/model-00055-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..11fe72b97291b18c0df73a1cf8594ae06c428a31 --- /dev/null +++ b/model-00055-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:753ad5189b72f7da202bab4f59369c45ef2b10868817f79200dd548ec009de1c +size 4966188880 diff --git a/model-00056-of-00106.safetensors b/model-00056-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d1ab983c32dcc6bcf58066283a2ba18f14601ac0 --- /dev/null +++ b/model-00056-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9316bae5df31ed48fd38593c0c4396198aeafa9668b77cac652d3a36f7c2cafe +size 4362142872 diff --git a/model-00057-of-00106.safetensors b/model-00057-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..66e159daadc4a0e1665822c69f921aff51006c84 --- /dev/null +++ b/model-00057-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2606cb9a14ec3bd53d3e85317d6202a5a2af352ab2e336e5d1f405d8e2e4f5f +size 4362142872 diff --git a/model-00058-of-00106.safetensors b/model-00058-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..829d4c42900882c41071b3b3761fc49802980be5 --- /dev/null +++ b/model-00058-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0c5194dccdfc5af4542d1500c3efc560aadc4d17fc39cbabf0d89593ef266342 +size 4966188880 diff --git a/model-00059-of-00106.safetensors b/model-00059-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2de3345af504ec1eedf0c30725a86a222b5a3139 --- /dev/null +++ b/model-00059-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:116eaa1b1748b9518f2898fc069051069db690e1e83a0fcc3fb494b043e7a147 +size 4362142872 diff --git a/model-00060-of-00106.safetensors b/model-00060-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..92cfb775cb828fb9cd9868f048da3c1982accc1b --- /dev/null +++ b/model-00060-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fb1304a7cf5cfd2d2472b6941a9f91830ce719c43ee52c884fd360b8593fd951 +size 4362142872 diff --git a/model-00061-of-00106.safetensors b/model-00061-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1bf9df08754b3e3cc36f93aa2b80cefa29fdf591 --- /dev/null +++ b/model-00061-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c42d35fb37a70be7e9966a97b948943f8b240a1855d7df802f816fa20f963c44 +size 4966188880 diff --git a/model-00062-of-00106.safetensors b/model-00062-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b2428a4572bf9c164face8bc6a0af15048e6e2d7 --- /dev/null +++ b/model-00062-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b709e372b476e12928d457222f8a3b8059ccb233870d493febc258dff51c9644 +size 4362142872 diff --git a/model-00063-of-00106.safetensors b/model-00063-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7fea40cc64ee5e38addca61bf19a9beb94c5e01e --- /dev/null +++ b/model-00063-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:82f695843658f3d39cb864efcdfb6a48b2ff2f9c97c3c52607993feefcc36f8e +size 4362142872 diff --git a/model-00064-of-00106.safetensors b/model-00064-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..43edb78411e7fbff7e9724f1617d8cd16ff59fba --- /dev/null +++ b/model-00064-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe5b144d3a148d200c363eef2b52a15659640ab30096ba007916d854a67466c6 +size 4966188880 diff --git a/model-00065-of-00106.safetensors b/model-00065-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0adef4f05bc795c56a71814ebbace04b602f7706 --- /dev/null +++ b/model-00065-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a124412002c75c495d3cd19499854aac7880c1d58269e4708d7fab5048aa6f8a +size 4362142872 diff --git a/model-00066-of-00106.safetensors b/model-00066-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d5d56fb520f90f48787aaaedc406fad9662909e --- /dev/null +++ b/model-00066-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b17249ad41e25a8ca519929df3861bb64106450967367ae8fa6004d396fd1e7d +size 4362142872 diff --git a/model-00067-of-00106.safetensors b/model-00067-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1814c0280fb946295dac2e643f3f394b6e6b7591 --- /dev/null +++ b/model-00067-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec73b2b593f43edcff50d7de0f060e9bbd6453f7c83c45ef6751f7d5f8e05be2 +size 4966188880 diff --git a/model-00068-of-00106.safetensors b/model-00068-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1aae4261805ca034c526c2fa321027d7a81d6b70 --- /dev/null +++ b/model-00068-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:64dcbf039886b885ca430e53642ad5a63db8e3854ab6451cc903ab1245dab63e +size 4362142872 diff --git a/model-00069-of-00106.safetensors b/model-00069-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d7676db16de00d080937cb0b25bf978fa3fe0aac --- /dev/null +++ b/model-00069-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d2c25d11d0d1c55079a314cd443894a08212e0942ea0ea1e22ab459869801c62 +size 4362142872 diff --git a/model-00070-of-00106.safetensors b/model-00070-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..075ff9b20b2e27e247a2aa2ab5802ea74767e131 --- /dev/null +++ b/model-00070-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69f09b277fedb99e67e5bf1008526fb0ea6d59670710d499ddc4d4c230285438 +size 4966188880 diff --git a/model-00071-of-00106.safetensors b/model-00071-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a99dfd2aae480dc661cfb4ef6c9cfdf44551df07 --- /dev/null +++ b/model-00071-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a07d23a5677a0a59428395f5b557197a5e3d1bebd6b167ba44453dc58ee45361 +size 4362142872 diff --git a/model-00072-of-00106.safetensors b/model-00072-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c011ffe98714c2e3014bb51ade5d0fb9ff6d85b8 --- /dev/null +++ b/model-00072-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e47abf113e74c8b144d9b27f21c9c7176f2e0240b3c9a789c258bade2c569756 +size 4362142872 diff --git a/model-00073-of-00106.safetensors b/model-00073-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..217fe41e5a8bcbe503593f9d421cd2b40affd244 --- /dev/null +++ b/model-00073-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7806df2549eb4de5582e0aa2506b9b2600a794b5407afc8a0c1b54374569c162 +size 4966188880 diff --git a/model-00074-of-00106.safetensors b/model-00074-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cdb7613cecac571ac12f0e8e920c512f6a6a67f7 --- /dev/null +++ b/model-00074-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f5942622f1bdbfa62ed578a54f9d6df10ddf84e0fe71a71fa3b9f60f444bf30d +size 4362142872 diff --git a/model-00075-of-00106.safetensors b/model-00075-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bea6aa3828ff1c7bb477a5978fddfa151295c622 --- /dev/null +++ b/model-00075-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ffe80449cac7ee4a7041945bf823c11e1a72b120a9d18c0db1a2e09669495527 +size 4362142872 diff --git a/model-00076-of-00106.safetensors b/model-00076-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0cdacaa16f729a6b7d6d58a19590a76111155e62 --- /dev/null +++ b/model-00076-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c983f08781a6a1062aea3a65539c7d5024a147ba8a80af75cfe64413b1fd9add +size 4966188896 diff --git a/model-00077-of-00106.safetensors b/model-00077-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a542e61b0300d6689f9844827f1f43ef09672b53 --- /dev/null +++ b/model-00077-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:43d524970c5265206dce94a7646ada0304d6615b88859fb160e816e8706efff7 +size 4362142888 diff --git a/model-00078-of-00106.safetensors b/model-00078-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6c1c2409660fe20b8fe675ee01da740caa428d64 --- /dev/null +++ b/model-00078-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fe40f63a92dade659e074a01717f45212eebf6ba403085d62b2599f23775d65d +size 4362142880 diff --git a/model-00079-of-00106.safetensors b/model-00079-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b8e925f028d3461be2ce4cbbf5ff0314bf37d0db --- /dev/null +++ b/model-00079-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5e7ee69573d9721f4a3f59bb376cbfe0983e155135931b8813dd212adb08d3b2 +size 4966188896 diff --git a/model-00080-of-00106.safetensors b/model-00080-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0d9b13040fb0fb0799f99695e9dd67f6f9479a1c --- /dev/null +++ b/model-00080-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b1e69275960e1acaafb771c2f39aeb6d16ab3cff1714c0e2e7d6aff5b296e88 +size 4362142888 diff --git a/model-00081-of-00106.safetensors b/model-00081-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2d427aa59fdebc4af92f4554e6ebb419aee87d76 --- /dev/null +++ b/model-00081-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75a645fd39451d36cacb212ae443f6cca91143bbcc2aedd50c48f8ae4b59accf +size 4362142880 diff --git a/model-00082-of-00106.safetensors b/model-00082-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5a9b70749ad555be83edeb69446af5a7c1d3d70c --- /dev/null +++ b/model-00082-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69e25258a61246d726f55766f08a5c872b3596c8a291a38b5c9aa51fc3d773bf +size 4966188896 diff --git a/model-00083-of-00106.safetensors b/model-00083-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ed633b24a33e7a91459e086bb2363bce5c3fed49 --- /dev/null +++ b/model-00083-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5a0fbe38bcb998e0d43b39d0bc89669859a2e67a4107d814ab270648e1197f9f +size 4362142888 diff --git a/model-00084-of-00106.safetensors b/model-00084-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..663ede4b1ad29f128bea2e6f829dfe0bcbadc615 --- /dev/null +++ b/model-00084-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9a27b7744c7496e61642361804725ffd3d3baf011a0bf2ee490b81b7a62514e7 +size 4362142880 diff --git a/model-00085-of-00106.safetensors b/model-00085-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0fe02434092554af6cc698b715837fbf577df550 --- /dev/null +++ b/model-00085-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6fd1a3d801c171cf8723e2b77a485863b9221fa454c65af788c8ce6a5b2d778b +size 4966188896 diff --git a/model-00086-of-00106.safetensors b/model-00086-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8568348db394c3b50ca70b327638f2fd544d4e19 --- /dev/null +++ b/model-00086-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4e74890d5ffc82a5d3546216e24ffdb04a467c2379da52bb68558468d23b0d8d +size 4362142888 diff --git a/model-00087-of-00106.safetensors b/model-00087-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e71a61fa9d45adce905ea7f1562579e62c81ef2a --- /dev/null +++ b/model-00087-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7cf4ab80120ab69576eca02e4f0abb776629343f42d6f49087f7c354808c70d2 +size 4362142880 diff --git a/model-00088-of-00106.safetensors b/model-00088-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..934c6a5e5e78d232ab2fbbfc279057c671cdae2a --- /dev/null +++ b/model-00088-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:75c303804774e6a5411641b809cd40127576bb35d8e76173acb2d9d95858b9b7 +size 4966188896 diff --git a/model-00089-of-00106.safetensors b/model-00089-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7c86ed1e6b902b56d2ee31e08e880f6012b87134 --- /dev/null +++ b/model-00089-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ad9c50f5470a94bfddaf0d06e234b8e2c121fedcdfad340ea0f41a85ae14a27e +size 4362142888 diff --git a/model-00090-of-00106.safetensors b/model-00090-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a67f89a7267faf186689c12a7a1ebc2ba18a0f88 --- /dev/null +++ b/model-00090-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:86e0114b314166c030b80ee3510977c88db1168e7168f63bf7d3c917e5c7a14b +size 4362142880 diff --git a/model-00091-of-00106.safetensors b/model-00091-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9de365ddab0c2f84e8da9d5bf127ca754fdc39ed --- /dev/null +++ b/model-00091-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c85e2fbd2092abb693863e6672ba14f06b230e0ed2a2d646553f877eb85abecc +size 4966188896 diff --git a/model-00092-of-00106.safetensors b/model-00092-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5c6c767afa137a62907421db70ad5f841f5eb9e1 --- /dev/null +++ b/model-00092-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7605966691436e89919f31b650c683da7fe5d808d0676be87b4eaf713e6efac2 +size 4362142888 diff --git a/model-00093-of-00106.safetensors b/model-00093-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..064aa81b5289a91d9ee2c5e1dc67bc2419fa109c --- /dev/null +++ b/model-00093-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:350ee64e3f138e8ba90ce668a27f3efba16e08d74d5eb74f885601db384988aa +size 4362142880 diff --git a/model-00094-of-00106.safetensors b/model-00094-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a591894d8b9033c2d38cbd67ceba594b90c528f --- /dev/null +++ b/model-00094-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:edce6e3bf1660ad1ce1b4f6ab2bfa4b19b2a3f02d391c763c74ddf0e4cba1353 +size 4966188896 diff --git a/model-00095-of-00106.safetensors b/model-00095-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..60f4eedfb977629f4f5fdd5fe652a70f78612f75 --- /dev/null +++ b/model-00095-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:99454219d7f50d1a400956c97e86da411ddb176f7fc5dbe2bde33f9ec1c40bff +size 4362142888 diff --git a/model-00096-of-00106.safetensors b/model-00096-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..265a5a5a4e37b7a7960b36cc12608462e05f7464 --- /dev/null +++ b/model-00096-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5efe2a3e596f69d7b8f7156434bf864b56a204680e5b840ba70d917d7e0be124 +size 4362142880 diff --git a/model-00097-of-00106.safetensors b/model-00097-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c439dcd3b6ecd6bcc5b3a99614e33a9a35d7284b --- /dev/null +++ b/model-00097-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e83b4c2234def1bb2afb411ecbb90e6738a8cf4ee7e97514385cc1fc5e8c2dad +size 4966188896 diff --git a/model-00098-of-00106.safetensors b/model-00098-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..38a21c0a3209115f49aae7b48e23ad7f7af6bb54 --- /dev/null +++ b/model-00098-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4390fb9094470b41f90342a30b2baeed5f3011c6b2a46aed80c0fb30de10ba53 +size 4362142888 diff --git a/model-00099-of-00106.safetensors b/model-00099-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bd893d9f4d0be296b2bb64894f5ce486ac57c511 --- /dev/null +++ b/model-00099-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:48598d0a60d5cc9d9817385060f139d3527862c977b550610cf43eb90cd54189 +size 4362142880 diff --git a/model-00100-of-00106.safetensors b/model-00100-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e050b46ef380d543c9cfa347d213c8faa8b3eff3 --- /dev/null +++ b/model-00100-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cca5317a06acbdf2c31f450f1cb9d0ff172f30a7a176a96944658ba9b4b1c15e +size 4966188896 diff --git a/model-00101-of-00106.safetensors b/model-00101-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bbad03465b7e699e4e86db7644824bcfef323600 --- /dev/null +++ b/model-00101-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04f10b31266d09536711e1ef7a2e8a044ab0f94fab32f87f75bf810da9dd055a +size 4362142888 diff --git a/model-00102-of-00106.safetensors b/model-00102-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2a86de4ba5e73997622806d6f84afc35663890fc --- /dev/null +++ b/model-00102-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:111ce0bce2656a5fda168bf621b4b99c1a881eaf50f17ffc00c06c13d86a4432 +size 4362142880 diff --git a/model-00103-of-00106.safetensors b/model-00103-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..90c3efed6ee9a5d1cb3d05dfcf809843336833bc --- /dev/null +++ b/model-00103-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a7623d12f67f55c490d96583ef81b0f872e1f8a0de657dc82b7aa8f7b9786af6 +size 4966188896 diff --git a/model-00104-of-00106.safetensors b/model-00104-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..0c8557752f3561f7d87a384042f7dadad5b8d1d8 --- /dev/null +++ b/model-00104-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9f6d0d3c0b8e3389a73cb33faee892120e2dea3a9b21ef1f97f4de825dd906ef +size 4362142888 diff --git a/model-00105-of-00106.safetensors b/model-00105-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fba13c465b43f602e33f74e5bb2f46b78852025a --- /dev/null +++ b/model-00105-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7b54297a573a59a7fdc72ce04324e969a29697fee6dc57ba0336cfa717f8c446 +size 4362142880 diff --git a/model-00106-of-00106.safetensors b/model-00106-of-00106.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9d5335fa7fe3930e296af30158aa350619c78122 --- /dev/null +++ b/model-00106-of-00106.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6ff2097d710c553bb61843c02f08a3075e28ac9b8b44ca25c54cda150b7c77da +size 1988198968 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..4205262bd427465e4a13b0284bd583c3db9c93cb --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,1270 @@ +{ + "metadata": { + "total_size": 481263648768 + }, + "weight_map": { + "lm_head.weight": "model-00106-of-00106.safetensors", + "model.embed_tokens.weight": "model-00001-of-00106.safetensors", + "model.layers.0.input_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.down_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.gate_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.mlp.up_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.post_attention_layernorm.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.o_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.0.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.input_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.down_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.mlp.up_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.post_attention_layernorm.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.k_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.1.self_attn.q_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.1.self_attn.v_proj.weight": "model-00001-of-00106.safetensors", + "model.layers.10.input_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.down_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.10.post_attention_layernorm.weight": "model-00009-of-00106.safetensors", + "model.layers.10.self_attn.k_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.o_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.q_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.10.self_attn.v_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.100.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.gate_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.mlp.up_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.100.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.input_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.down_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.mlp.up_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.101.post_attention_layernorm.weight": "model-00077-of-00106.safetensors", + "model.layers.101.self_attn.k_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.o_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.q_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.101.self_attn.v_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.102.input_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.down_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.mlp.gate_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.102.post_attention_layernorm.weight": "model-00078-of-00106.safetensors", + "model.layers.102.self_attn.k_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.o_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.q_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.102.self_attn.v_proj.weight": "model-00077-of-00106.safetensors", + "model.layers.103.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.103.mlp.gate_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.mlp.up_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.103.self_attn.k_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.o_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.q_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.103.self_attn.v_proj.weight": "model-00078-of-00106.safetensors", + "model.layers.104.input_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.down_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.gate_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.mlp.up_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.post_attention_layernorm.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.104.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.input_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.down_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.mlp.up_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.105.post_attention_layernorm.weight": "model-00080-of-00106.safetensors", + "model.layers.105.self_attn.k_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.o_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.q_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.105.self_attn.v_proj.weight": "model-00079-of-00106.safetensors", + "model.layers.106.input_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.down_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.mlp.gate_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.106.post_attention_layernorm.weight": "model-00081-of-00106.safetensors", + "model.layers.106.self_attn.k_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.o_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.q_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.106.self_attn.v_proj.weight": "model-00080-of-00106.safetensors", + "model.layers.107.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.107.mlp.gate_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.mlp.up_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.107.self_attn.k_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.o_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.q_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.107.self_attn.v_proj.weight": "model-00081-of-00106.safetensors", + "model.layers.108.input_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.down_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.gate_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.mlp.up_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.post_attention_layernorm.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.108.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.input_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.down_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.mlp.up_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.109.post_attention_layernorm.weight": "model-00083-of-00106.safetensors", + "model.layers.109.self_attn.k_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.o_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.q_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.109.self_attn.v_proj.weight": "model-00082-of-00106.safetensors", + "model.layers.11.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.11.mlp.gate_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.mlp.up_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.11.self_attn.k_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.o_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.q_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.11.self_attn.v_proj.weight": "model-00009-of-00106.safetensors", + "model.layers.110.input_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.down_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.mlp.gate_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.110.post_attention_layernorm.weight": "model-00084-of-00106.safetensors", + "model.layers.110.self_attn.k_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.o_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.q_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.110.self_attn.v_proj.weight": "model-00083-of-00106.safetensors", + "model.layers.111.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.111.mlp.gate_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.mlp.up_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.111.self_attn.k_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.o_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.q_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.111.self_attn.v_proj.weight": "model-00084-of-00106.safetensors", + "model.layers.112.input_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.down_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.gate_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.mlp.up_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.post_attention_layernorm.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.112.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.input_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.down_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.mlp.up_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.113.post_attention_layernorm.weight": "model-00086-of-00106.safetensors", + "model.layers.113.self_attn.k_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.o_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.q_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.113.self_attn.v_proj.weight": "model-00085-of-00106.safetensors", + "model.layers.114.input_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.down_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.mlp.gate_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.114.post_attention_layernorm.weight": "model-00087-of-00106.safetensors", + "model.layers.114.self_attn.k_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.o_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.q_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.114.self_attn.v_proj.weight": "model-00086-of-00106.safetensors", + "model.layers.115.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.115.mlp.gate_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.mlp.up_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.115.self_attn.k_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.o_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.q_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.115.self_attn.v_proj.weight": "model-00087-of-00106.safetensors", + "model.layers.116.input_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.down_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.gate_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.mlp.up_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.post_attention_layernorm.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.116.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.input_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.down_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.mlp.up_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.117.post_attention_layernorm.weight": "model-00089-of-00106.safetensors", + "model.layers.117.self_attn.k_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.o_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.q_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.117.self_attn.v_proj.weight": "model-00088-of-00106.safetensors", + "model.layers.118.input_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.down_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.mlp.gate_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.118.post_attention_layernorm.weight": "model-00090-of-00106.safetensors", + "model.layers.118.self_attn.k_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.o_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.q_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.118.self_attn.v_proj.weight": "model-00089-of-00106.safetensors", + "model.layers.119.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.119.mlp.gate_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.mlp.up_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.119.self_attn.k_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.o_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.q_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.119.self_attn.v_proj.weight": "model-00090-of-00106.safetensors", + "model.layers.12.input_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.down_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.gate_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.mlp.up_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.post_attention_layernorm.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.12.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.120.input_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.down_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.gate_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.mlp.up_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.post_attention_layernorm.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.120.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.input_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.down_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.mlp.up_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.121.post_attention_layernorm.weight": "model-00092-of-00106.safetensors", + "model.layers.121.self_attn.k_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.o_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.q_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.121.self_attn.v_proj.weight": "model-00091-of-00106.safetensors", + "model.layers.122.input_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.down_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.mlp.gate_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.122.post_attention_layernorm.weight": "model-00093-of-00106.safetensors", + "model.layers.122.self_attn.k_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.o_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.q_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.122.self_attn.v_proj.weight": "model-00092-of-00106.safetensors", + "model.layers.123.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.123.mlp.gate_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.mlp.up_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.123.self_attn.k_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.o_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.q_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.123.self_attn.v_proj.weight": "model-00093-of-00106.safetensors", + "model.layers.124.input_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.down_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.gate_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.mlp.up_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.post_attention_layernorm.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.124.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.input_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.down_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.mlp.up_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.125.post_attention_layernorm.weight": "model-00095-of-00106.safetensors", + "model.layers.125.self_attn.k_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.o_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.q_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.125.self_attn.v_proj.weight": "model-00094-of-00106.safetensors", + "model.layers.126.input_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.down_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.mlp.gate_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.126.post_attention_layernorm.weight": "model-00096-of-00106.safetensors", + "model.layers.126.self_attn.k_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.o_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.q_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.126.self_attn.v_proj.weight": "model-00095-of-00106.safetensors", + "model.layers.127.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.127.mlp.gate_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.mlp.up_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.127.self_attn.k_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.o_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.q_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.127.self_attn.v_proj.weight": "model-00096-of-00106.safetensors", + "model.layers.128.input_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.down_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.gate_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.mlp.up_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.post_attention_layernorm.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.128.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.input_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.down_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.mlp.up_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.129.post_attention_layernorm.weight": "model-00098-of-00106.safetensors", + "model.layers.129.self_attn.k_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.o_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.q_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.129.self_attn.v_proj.weight": "model-00097-of-00106.safetensors", + "model.layers.13.input_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.down_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.mlp.up_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.13.post_attention_layernorm.weight": "model-00011-of-00106.safetensors", + "model.layers.13.self_attn.k_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.o_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.q_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.13.self_attn.v_proj.weight": "model-00010-of-00106.safetensors", + "model.layers.130.input_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.down_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.mlp.gate_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.130.post_attention_layernorm.weight": "model-00099-of-00106.safetensors", + "model.layers.130.self_attn.k_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.o_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.q_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.130.self_attn.v_proj.weight": "model-00098-of-00106.safetensors", + "model.layers.131.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.131.mlp.gate_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.mlp.up_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.131.self_attn.k_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.o_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.q_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.131.self_attn.v_proj.weight": "model-00099-of-00106.safetensors", + "model.layers.132.input_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.down_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.gate_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.mlp.up_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.post_attention_layernorm.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.132.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.input_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.down_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.mlp.up_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.133.post_attention_layernorm.weight": "model-00101-of-00106.safetensors", + "model.layers.133.self_attn.k_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.o_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.q_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.133.self_attn.v_proj.weight": "model-00100-of-00106.safetensors", + "model.layers.134.input_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.down_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.mlp.gate_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.134.post_attention_layernorm.weight": "model-00102-of-00106.safetensors", + "model.layers.134.self_attn.k_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.o_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.q_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.134.self_attn.v_proj.weight": "model-00101-of-00106.safetensors", + "model.layers.135.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.135.mlp.gate_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.mlp.up_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.135.self_attn.k_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.o_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.q_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.135.self_attn.v_proj.weight": "model-00102-of-00106.safetensors", + "model.layers.136.input_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.down_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.gate_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.mlp.up_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.post_attention_layernorm.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.136.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.input_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.down_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.mlp.up_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.137.post_attention_layernorm.weight": "model-00104-of-00106.safetensors", + "model.layers.137.self_attn.k_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.o_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.q_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.137.self_attn.v_proj.weight": "model-00103-of-00106.safetensors", + "model.layers.138.input_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.down_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.mlp.gate_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.138.post_attention_layernorm.weight": "model-00105-of-00106.safetensors", + "model.layers.138.self_attn.k_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.o_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.q_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.138.self_attn.v_proj.weight": "model-00104-of-00106.safetensors", + "model.layers.139.input_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.down_proj.weight": "model-00106-of-00106.safetensors", + "model.layers.139.mlp.gate_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.mlp.up_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.post_attention_layernorm.weight": "model-00106-of-00106.safetensors", + "model.layers.139.self_attn.k_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.o_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.q_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.139.self_attn.v_proj.weight": "model-00105-of-00106.safetensors", + "model.layers.14.input_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.down_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.mlp.gate_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.14.post_attention_layernorm.weight": "model-00012-of-00106.safetensors", + "model.layers.14.self_attn.k_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.o_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.q_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.14.self_attn.v_proj.weight": "model-00011-of-00106.safetensors", + "model.layers.15.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.15.mlp.gate_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.mlp.up_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.15.self_attn.k_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.o_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.q_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.15.self_attn.v_proj.weight": "model-00012-of-00106.safetensors", + "model.layers.16.input_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.down_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.gate_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.mlp.up_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.post_attention_layernorm.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.16.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.input_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.down_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.mlp.up_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.17.post_attention_layernorm.weight": "model-00014-of-00106.safetensors", + "model.layers.17.self_attn.k_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.o_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.q_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.17.self_attn.v_proj.weight": "model-00013-of-00106.safetensors", + "model.layers.18.input_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.down_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.mlp.gate_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.18.post_attention_layernorm.weight": "model-00015-of-00106.safetensors", + "model.layers.18.self_attn.k_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.o_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.q_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.18.self_attn.v_proj.weight": "model-00014-of-00106.safetensors", + "model.layers.19.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.19.mlp.gate_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.mlp.up_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.19.self_attn.k_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.o_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.q_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.19.self_attn.v_proj.weight": "model-00015-of-00106.safetensors", + "model.layers.2.input_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.down_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.mlp.gate_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.2.post_attention_layernorm.weight": "model-00003-of-00106.safetensors", + "model.layers.2.self_attn.k_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.o_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.q_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.2.self_attn.v_proj.weight": "model-00002-of-00106.safetensors", + "model.layers.20.input_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.down_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.gate_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.mlp.up_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.post_attention_layernorm.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.20.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.input_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.down_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.mlp.up_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.21.post_attention_layernorm.weight": "model-00017-of-00106.safetensors", + "model.layers.21.self_attn.k_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.o_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.q_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.21.self_attn.v_proj.weight": "model-00016-of-00106.safetensors", + "model.layers.22.input_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.down_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.mlp.gate_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.22.post_attention_layernorm.weight": "model-00018-of-00106.safetensors", + "model.layers.22.self_attn.k_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.o_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.q_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.22.self_attn.v_proj.weight": "model-00017-of-00106.safetensors", + "model.layers.23.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.23.mlp.gate_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.mlp.up_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.23.self_attn.k_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.o_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.q_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.23.self_attn.v_proj.weight": "model-00018-of-00106.safetensors", + "model.layers.24.input_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.down_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.gate_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.mlp.up_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.post_attention_layernorm.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.24.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.input_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.down_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.mlp.up_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.25.post_attention_layernorm.weight": "model-00020-of-00106.safetensors", + "model.layers.25.self_attn.k_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.o_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.q_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.25.self_attn.v_proj.weight": "model-00019-of-00106.safetensors", + "model.layers.26.input_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.down_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.mlp.gate_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.26.post_attention_layernorm.weight": "model-00021-of-00106.safetensors", + "model.layers.26.self_attn.k_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.o_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.q_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.26.self_attn.v_proj.weight": "model-00020-of-00106.safetensors", + "model.layers.27.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.27.mlp.gate_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.mlp.up_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.27.self_attn.k_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.o_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.q_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.27.self_attn.v_proj.weight": "model-00021-of-00106.safetensors", + "model.layers.28.input_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.down_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.gate_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.mlp.up_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.post_attention_layernorm.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.28.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.input_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.down_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.mlp.up_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.29.post_attention_layernorm.weight": "model-00023-of-00106.safetensors", + "model.layers.29.self_attn.k_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.o_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.q_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.29.self_attn.v_proj.weight": "model-00022-of-00106.safetensors", + "model.layers.3.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.3.mlp.gate_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.mlp.up_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.3.self_attn.k_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.o_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.q_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.3.self_attn.v_proj.weight": "model-00003-of-00106.safetensors", + "model.layers.30.input_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.down_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.mlp.gate_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.30.post_attention_layernorm.weight": "model-00024-of-00106.safetensors", + "model.layers.30.self_attn.k_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.o_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.q_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.30.self_attn.v_proj.weight": "model-00023-of-00106.safetensors", + "model.layers.31.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.31.mlp.gate_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.mlp.up_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.31.self_attn.k_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.o_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.q_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.31.self_attn.v_proj.weight": "model-00024-of-00106.safetensors", + "model.layers.32.input_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.down_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.gate_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.mlp.up_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.post_attention_layernorm.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.32.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.input_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.down_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.mlp.up_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.33.post_attention_layernorm.weight": "model-00026-of-00106.safetensors", + "model.layers.33.self_attn.k_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.o_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.q_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.33.self_attn.v_proj.weight": "model-00025-of-00106.safetensors", + "model.layers.34.input_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.down_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.mlp.gate_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.34.post_attention_layernorm.weight": "model-00027-of-00106.safetensors", + "model.layers.34.self_attn.k_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.o_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.q_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.34.self_attn.v_proj.weight": "model-00026-of-00106.safetensors", + "model.layers.35.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.35.mlp.gate_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.mlp.up_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.35.self_attn.k_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.o_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.q_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.35.self_attn.v_proj.weight": "model-00027-of-00106.safetensors", + "model.layers.36.input_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.down_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.gate_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.mlp.up_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.post_attention_layernorm.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.36.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.input_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.down_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.mlp.up_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.37.post_attention_layernorm.weight": "model-00029-of-00106.safetensors", + "model.layers.37.self_attn.k_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.o_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.q_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.37.self_attn.v_proj.weight": "model-00028-of-00106.safetensors", + "model.layers.38.input_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.down_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.mlp.gate_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.38.post_attention_layernorm.weight": "model-00030-of-00106.safetensors", + "model.layers.38.self_attn.k_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.o_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.q_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.38.self_attn.v_proj.weight": "model-00029-of-00106.safetensors", + "model.layers.39.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.39.mlp.gate_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.mlp.up_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.39.self_attn.k_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.o_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.q_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.39.self_attn.v_proj.weight": "model-00030-of-00106.safetensors", + "model.layers.4.input_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.down_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.gate_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.mlp.up_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.post_attention_layernorm.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.4.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.40.input_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.down_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.gate_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.mlp.up_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.post_attention_layernorm.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.40.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.input_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.down_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.mlp.up_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.41.post_attention_layernorm.weight": "model-00032-of-00106.safetensors", + "model.layers.41.self_attn.k_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.o_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.q_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.41.self_attn.v_proj.weight": "model-00031-of-00106.safetensors", + "model.layers.42.input_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.down_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.mlp.gate_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.42.post_attention_layernorm.weight": "model-00033-of-00106.safetensors", + "model.layers.42.self_attn.k_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.o_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.q_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.42.self_attn.v_proj.weight": "model-00032-of-00106.safetensors", + "model.layers.43.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.43.mlp.gate_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.mlp.up_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.43.self_attn.k_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.o_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.q_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.43.self_attn.v_proj.weight": "model-00033-of-00106.safetensors", + "model.layers.44.input_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.down_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.gate_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.mlp.up_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.post_attention_layernorm.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.44.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.input_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.down_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.mlp.up_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.45.post_attention_layernorm.weight": "model-00035-of-00106.safetensors", + "model.layers.45.self_attn.k_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.o_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.q_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.45.self_attn.v_proj.weight": "model-00034-of-00106.safetensors", + "model.layers.46.input_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.down_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.mlp.gate_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.46.post_attention_layernorm.weight": "model-00036-of-00106.safetensors", + "model.layers.46.self_attn.k_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.o_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.q_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.46.self_attn.v_proj.weight": "model-00035-of-00106.safetensors", + "model.layers.47.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.47.mlp.gate_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.mlp.up_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.47.self_attn.k_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.o_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.q_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.47.self_attn.v_proj.weight": "model-00036-of-00106.safetensors", + "model.layers.48.input_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.down_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.gate_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.mlp.up_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.post_attention_layernorm.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.48.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.input_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.down_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.mlp.up_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.49.post_attention_layernorm.weight": "model-00038-of-00106.safetensors", + "model.layers.49.self_attn.k_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.o_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.q_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.49.self_attn.v_proj.weight": "model-00037-of-00106.safetensors", + "model.layers.5.input_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.down_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.mlp.up_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.5.post_attention_layernorm.weight": "model-00005-of-00106.safetensors", + "model.layers.5.self_attn.k_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.o_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.q_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.5.self_attn.v_proj.weight": "model-00004-of-00106.safetensors", + "model.layers.50.input_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.down_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.mlp.gate_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.50.post_attention_layernorm.weight": "model-00039-of-00106.safetensors", + "model.layers.50.self_attn.k_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.o_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.q_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.50.self_attn.v_proj.weight": "model-00038-of-00106.safetensors", + "model.layers.51.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.51.mlp.gate_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.mlp.up_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.51.self_attn.k_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.o_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.q_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.51.self_attn.v_proj.weight": "model-00039-of-00106.safetensors", + "model.layers.52.input_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.down_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.gate_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.mlp.up_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.post_attention_layernorm.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.52.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.input_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.down_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.mlp.up_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.53.post_attention_layernorm.weight": "model-00041-of-00106.safetensors", + "model.layers.53.self_attn.k_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.o_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.q_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.53.self_attn.v_proj.weight": "model-00040-of-00106.safetensors", + "model.layers.54.input_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.down_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.mlp.gate_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.54.post_attention_layernorm.weight": "model-00042-of-00106.safetensors", + "model.layers.54.self_attn.k_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.o_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.q_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.54.self_attn.v_proj.weight": "model-00041-of-00106.safetensors", + "model.layers.55.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.55.mlp.gate_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.mlp.up_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.55.self_attn.k_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.o_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.q_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.55.self_attn.v_proj.weight": "model-00042-of-00106.safetensors", + "model.layers.56.input_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.down_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.gate_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.mlp.up_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.post_attention_layernorm.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.56.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.input_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.down_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.mlp.up_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.57.post_attention_layernorm.weight": "model-00044-of-00106.safetensors", + "model.layers.57.self_attn.k_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.o_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.q_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.57.self_attn.v_proj.weight": "model-00043-of-00106.safetensors", + "model.layers.58.input_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.down_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.mlp.gate_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.58.post_attention_layernorm.weight": "model-00045-of-00106.safetensors", + "model.layers.58.self_attn.k_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.o_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.q_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.58.self_attn.v_proj.weight": "model-00044-of-00106.safetensors", + "model.layers.59.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.59.mlp.gate_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.mlp.up_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.59.self_attn.k_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.o_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.q_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.59.self_attn.v_proj.weight": "model-00045-of-00106.safetensors", + "model.layers.6.input_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.down_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.mlp.gate_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.6.post_attention_layernorm.weight": "model-00006-of-00106.safetensors", + "model.layers.6.self_attn.k_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.o_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.q_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.6.self_attn.v_proj.weight": "model-00005-of-00106.safetensors", + "model.layers.60.input_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.down_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.gate_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.mlp.up_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.post_attention_layernorm.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.60.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.input_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.down_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.mlp.up_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.61.post_attention_layernorm.weight": "model-00047-of-00106.safetensors", + "model.layers.61.self_attn.k_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.o_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.q_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.61.self_attn.v_proj.weight": "model-00046-of-00106.safetensors", + "model.layers.62.input_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.down_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.mlp.gate_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.62.post_attention_layernorm.weight": "model-00048-of-00106.safetensors", + "model.layers.62.self_attn.k_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.o_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.q_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.62.self_attn.v_proj.weight": "model-00047-of-00106.safetensors", + "model.layers.63.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.63.mlp.gate_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.mlp.up_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.63.self_attn.k_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.o_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.q_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.63.self_attn.v_proj.weight": "model-00048-of-00106.safetensors", + "model.layers.64.input_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.down_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.gate_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.mlp.up_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.post_attention_layernorm.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.64.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.input_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.down_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.mlp.up_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.65.post_attention_layernorm.weight": "model-00050-of-00106.safetensors", + "model.layers.65.self_attn.k_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.o_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.q_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.65.self_attn.v_proj.weight": "model-00049-of-00106.safetensors", + "model.layers.66.input_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.down_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.mlp.gate_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.66.post_attention_layernorm.weight": "model-00051-of-00106.safetensors", + "model.layers.66.self_attn.k_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.o_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.q_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.66.self_attn.v_proj.weight": "model-00050-of-00106.safetensors", + "model.layers.67.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.67.mlp.gate_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.mlp.up_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.67.self_attn.k_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.o_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.q_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.67.self_attn.v_proj.weight": "model-00051-of-00106.safetensors", + "model.layers.68.input_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.down_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.gate_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.mlp.up_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.post_attention_layernorm.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.68.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.input_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.down_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.mlp.up_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.69.post_attention_layernorm.weight": "model-00053-of-00106.safetensors", + "model.layers.69.self_attn.k_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.o_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.q_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.69.self_attn.v_proj.weight": "model-00052-of-00106.safetensors", + "model.layers.7.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.7.mlp.gate_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.mlp.up_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.7.self_attn.k_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.o_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.q_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.7.self_attn.v_proj.weight": "model-00006-of-00106.safetensors", + "model.layers.70.input_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.down_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.mlp.gate_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.70.post_attention_layernorm.weight": "model-00054-of-00106.safetensors", + "model.layers.70.self_attn.k_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.o_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.q_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.70.self_attn.v_proj.weight": "model-00053-of-00106.safetensors", + "model.layers.71.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.71.mlp.gate_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.mlp.up_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.71.self_attn.k_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.o_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.q_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.71.self_attn.v_proj.weight": "model-00054-of-00106.safetensors", + "model.layers.72.input_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.down_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.gate_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.mlp.up_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.post_attention_layernorm.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.72.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.input_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.down_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.mlp.up_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.73.post_attention_layernorm.weight": "model-00056-of-00106.safetensors", + "model.layers.73.self_attn.k_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.o_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.q_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.73.self_attn.v_proj.weight": "model-00055-of-00106.safetensors", + "model.layers.74.input_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.down_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.mlp.gate_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.74.post_attention_layernorm.weight": "model-00057-of-00106.safetensors", + "model.layers.74.self_attn.k_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.o_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.q_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.74.self_attn.v_proj.weight": "model-00056-of-00106.safetensors", + "model.layers.75.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.75.mlp.gate_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.mlp.up_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.75.self_attn.k_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.o_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.q_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.75.self_attn.v_proj.weight": "model-00057-of-00106.safetensors", + "model.layers.76.input_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.down_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.gate_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.mlp.up_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.post_attention_layernorm.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.76.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.input_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.down_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.mlp.up_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.77.post_attention_layernorm.weight": "model-00059-of-00106.safetensors", + "model.layers.77.self_attn.k_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.o_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.q_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.77.self_attn.v_proj.weight": "model-00058-of-00106.safetensors", + "model.layers.78.input_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.down_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.mlp.gate_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.78.post_attention_layernorm.weight": "model-00060-of-00106.safetensors", + "model.layers.78.self_attn.k_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.o_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.q_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.78.self_attn.v_proj.weight": "model-00059-of-00106.safetensors", + "model.layers.79.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.79.mlp.gate_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.mlp.up_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.79.self_attn.k_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.o_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.q_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.79.self_attn.v_proj.weight": "model-00060-of-00106.safetensors", + "model.layers.8.input_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.down_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.gate_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.mlp.up_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.post_attention_layernorm.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.8.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.80.input_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.down_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.gate_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.mlp.up_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.post_attention_layernorm.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.80.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.input_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.down_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.mlp.up_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.81.post_attention_layernorm.weight": "model-00062-of-00106.safetensors", + "model.layers.81.self_attn.k_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.o_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.q_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.81.self_attn.v_proj.weight": "model-00061-of-00106.safetensors", + "model.layers.82.input_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.down_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.mlp.gate_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.82.post_attention_layernorm.weight": "model-00063-of-00106.safetensors", + "model.layers.82.self_attn.k_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.o_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.q_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.82.self_attn.v_proj.weight": "model-00062-of-00106.safetensors", + "model.layers.83.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.83.mlp.gate_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.mlp.up_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.83.self_attn.k_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.o_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.q_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.83.self_attn.v_proj.weight": "model-00063-of-00106.safetensors", + "model.layers.84.input_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.down_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.gate_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.mlp.up_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.post_attention_layernorm.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.84.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.input_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.down_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.mlp.up_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.85.post_attention_layernorm.weight": "model-00065-of-00106.safetensors", + "model.layers.85.self_attn.k_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.o_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.q_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.85.self_attn.v_proj.weight": "model-00064-of-00106.safetensors", + "model.layers.86.input_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.down_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.mlp.gate_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.86.post_attention_layernorm.weight": "model-00066-of-00106.safetensors", + "model.layers.86.self_attn.k_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.o_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.q_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.86.self_attn.v_proj.weight": "model-00065-of-00106.safetensors", + "model.layers.87.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.87.mlp.gate_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.mlp.up_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.87.self_attn.k_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.o_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.q_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.87.self_attn.v_proj.weight": "model-00066-of-00106.safetensors", + "model.layers.88.input_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.down_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.gate_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.mlp.up_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.post_attention_layernorm.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.88.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.input_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.down_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.mlp.up_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.89.post_attention_layernorm.weight": "model-00068-of-00106.safetensors", + "model.layers.89.self_attn.k_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.o_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.q_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.89.self_attn.v_proj.weight": "model-00067-of-00106.safetensors", + "model.layers.9.input_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.down_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.gate_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.mlp.up_proj.weight": "model-00008-of-00106.safetensors", + "model.layers.9.post_attention_layernorm.weight": "model-00008-of-00106.safetensors", + "model.layers.9.self_attn.k_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.o_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.q_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.9.self_attn.v_proj.weight": "model-00007-of-00106.safetensors", + "model.layers.90.input_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.down_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.mlp.gate_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.90.post_attention_layernorm.weight": "model-00069-of-00106.safetensors", + "model.layers.90.self_attn.k_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.o_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.q_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.90.self_attn.v_proj.weight": "model-00068-of-00106.safetensors", + "model.layers.91.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.91.mlp.gate_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.mlp.up_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.91.self_attn.k_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.o_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.q_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.91.self_attn.v_proj.weight": "model-00069-of-00106.safetensors", + "model.layers.92.input_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.down_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.gate_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.mlp.up_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.post_attention_layernorm.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.92.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.input_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.down_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.mlp.up_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.93.post_attention_layernorm.weight": "model-00071-of-00106.safetensors", + "model.layers.93.self_attn.k_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.o_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.q_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.93.self_attn.v_proj.weight": "model-00070-of-00106.safetensors", + "model.layers.94.input_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.down_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.mlp.gate_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.94.post_attention_layernorm.weight": "model-00072-of-00106.safetensors", + "model.layers.94.self_attn.k_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.o_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.q_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.94.self_attn.v_proj.weight": "model-00071-of-00106.safetensors", + "model.layers.95.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.95.mlp.gate_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.mlp.up_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.95.self_attn.k_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.o_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.q_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.95.self_attn.v_proj.weight": "model-00072-of-00106.safetensors", + "model.layers.96.input_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.down_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.gate_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.mlp.up_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.post_attention_layernorm.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.96.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.input_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.down_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.mlp.up_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.97.post_attention_layernorm.weight": "model-00074-of-00106.safetensors", + "model.layers.97.self_attn.k_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.o_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.q_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.97.self_attn.v_proj.weight": "model-00073-of-00106.safetensors", + "model.layers.98.input_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.down_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.mlp.gate_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.98.post_attention_layernorm.weight": "model-00075-of-00106.safetensors", + "model.layers.98.self_attn.k_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.o_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.q_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.98.self_attn.v_proj.weight": "model-00074-of-00106.safetensors", + "model.layers.99.input_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.down_proj.weight": "model-00076-of-00106.safetensors", + "model.layers.99.mlp.gate_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.mlp.up_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.post_attention_layernorm.weight": "model-00076-of-00106.safetensors", + "model.layers.99.self_attn.k_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.o_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.q_proj.weight": "model-00075-of-00106.safetensors", + "model.layers.99.self_attn.v_proj.weight": "model-00075-of-00106.safetensors", + "model.norm.weight": "model-00106-of-00106.safetensors" + } +}