diff --git a/README.md b/README.md new file mode 100644 index 0000000000000000000000000000000000000000..bc5f30d6632ac0efdc7be2e9095e9e9579af2e33 --- /dev/null +++ b/README.md @@ -0,0 +1,199 @@ +--- +library_name: transformers +tags: [] +--- + +# Model Card for Model ID + + + + + +## Model Details + +### Model Description + + + +This is the model card of a 🤗 transformers model that has been pushed on the Hub. This model card has been automatically generated. + +- **Developed by:** [More Information Needed] +- **Funded by [optional]:** [More Information Needed] +- **Shared by [optional]:** [More Information Needed] +- **Model type:** [More Information Needed] +- **Language(s) (NLP):** [More Information Needed] +- **License:** [More Information Needed] +- **Finetuned from model [optional]:** [More Information Needed] + +### Model Sources [optional] + + + +- **Repository:** [More Information Needed] +- **Paper [optional]:** [More Information Needed] +- **Demo [optional]:** [More Information Needed] + +## Uses + + + +### Direct Use + + + +[More Information Needed] + +### Downstream Use [optional] + + + +[More Information Needed] + +### Out-of-Scope Use + + + +[More Information Needed] + +## Bias, Risks, and Limitations + + + +[More Information Needed] + +### Recommendations + + + +Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations. + +## How to Get Started with the Model + +Use the code below to get started with the model. + +[More Information Needed] + +## Training Details + +### Training Data + + + +[More Information Needed] + +### Training Procedure + + + +#### Preprocessing [optional] + +[More Information Needed] + + +#### Training Hyperparameters + +- **Training regime:** [More Information Needed] + +#### Speeds, Sizes, Times [optional] + + + +[More Information Needed] + +## Evaluation + + + +### Testing Data, Factors & Metrics + +#### Testing Data + + + +[More Information Needed] + +#### Factors + + + +[More Information Needed] + +#### Metrics + + + +[More Information Needed] + +### Results + +[More Information Needed] + +#### Summary + + + +## Model Examination [optional] + + + +[More Information Needed] + +## Environmental Impact + + + +Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700). + +- **Hardware Type:** [More Information Needed] +- **Hours used:** [More Information Needed] +- **Cloud Provider:** [More Information Needed] +- **Compute Region:** [More Information Needed] +- **Carbon Emitted:** [More Information Needed] + +## Technical Specifications [optional] + +### Model Architecture and Objective + +[More Information Needed] + +### Compute Infrastructure + +[More Information Needed] + +#### Hardware + +[More Information Needed] + +#### Software + +[More Information Needed] + +## Citation [optional] + + + +**BibTeX:** + +[More Information Needed] + +**APA:** + +[More Information Needed] + +## Glossary [optional] + + + +[More Information Needed] + +## More Information [optional] + +[More Information Needed] + +## Model Card Authors [optional] + +[More Information Needed] + +## Model Card Contact + +[More Information Needed] \ No newline at end of file diff --git a/config.json b/config.json new file mode 100644 index 0000000000000000000000000000000000000000..67a3f453721a3bc349d92a94ae8c85acd89a007b --- /dev/null +++ b/config.json @@ -0,0 +1,35 @@ +{ + "_name_or_path": "/oe-adapt-default/costah/models/final/70B_ppo_1116_mix_best_nojacob_val_0.07_wr_0.1_lr_1e-7_checkpoints/step_40", + "architectures": [ + "LlamaModel" + ], + "attention_bias": false, + "attention_dropout": 0.0, + "bos_token_id": 128000, + "eos_token_id": 128001, + "hidden_act": "silu", + "hidden_size": 8192, + "initializer_range": 0.02, + "intermediate_size": 28672, + "max_position_embeddings": 131072, + "mlp_bias": false, + "model_type": "llama", + "num_attention_heads": 64, + "num_hidden_layers": 80, + "num_key_value_heads": 8, + "pretraining_tp": 1, + "rms_norm_eps": 1e-05, + "rope_scaling": { + "factor": 8.0, + "high_freq_factor": 4.0, + "low_freq_factor": 1.0, + "original_max_position_embeddings": 8192, + "rope_type": "llama3" + }, + "rope_theta": 500000.0, + "tie_word_embeddings": false, + "torch_dtype": "float32", + "transformers_version": "4.43.4", + "use_cache": false, + "vocab_size": 128264 +} diff --git a/model-00001-of-00061.safetensors b/model-00001-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..9c02acb9320f756af78d451697523bafabffd5a8 --- /dev/null +++ b/model-00001-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2dd10b4116a6316eed88a6d6e0daa00ed0ece9575dcd1a4641d337d39b7accd3 +size 4806935104 diff --git a/model-00002-of-00061.safetensors b/model-00002-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..557c2e4e2a666a12fe51bb2f2387cb89f62ab3d5 --- /dev/null +++ b/model-00002-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1428dd289164a5c79f635d943088b1bc96f9970b1f8af8bc7f0310b9c065efb2 +size 4362142808 diff --git a/model-00003-of-00061.safetensors b/model-00003-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..054595674e7af25e5391b8a8f1660a8b25bc910a --- /dev/null +++ b/model-00003-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38292c60262db09177e2342078b690d5f57fb7284a1a3394689a737e1c7463fa +size 4362142800 diff --git a/model-00004-of-00061.safetensors b/model-00004-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..fec9141b815a6b037c14f6b89bf5caced673bc86 --- /dev/null +++ b/model-00004-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:750db17e328908e67e03694c7b75685655354a13cd0f602c1ccbd6dc3f84e57b +size 4966188768 diff --git a/model-00005-of-00061.safetensors b/model-00005-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..711c79b49057eb8480c4fdf5a558ef9f45f25a69 --- /dev/null +++ b/model-00005-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1e9fd6e7f503bc3a19063f90974b3cc7def1a3a152802090bf2dde7ebe07feb7 +size 4362142808 diff --git a/model-00006-of-00061.safetensors b/model-00006-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..ba50189e00bdba03a2920139d4346562f03aed0c --- /dev/null +++ b/model-00006-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ab0e9a5d42b0c0b8372ac447aca36b5a8ffcd346b6021ddea62205f3a303990a +size 4362142800 diff --git a/model-00007-of-00061.safetensors b/model-00007-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..00e30a25e20d4c8c8a70db4e78dd338afa814a5c --- /dev/null +++ b/model-00007-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:9b9337e51f6b882aca0ef6290fc22794b35a4140a71fd5c1f4512d44f48d1c1a +size 4966188768 diff --git a/model-00008-of-00061.safetensors b/model-00008-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8a51b26f1dd0276edd936267967df3d15f7857b0 --- /dev/null +++ b/model-00008-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:52a9761c15fd257579ed37423f17df1e52c935c287333b4ad88f8b8e8cb0fff8 +size 4362142808 diff --git a/model-00009-of-00061.safetensors b/model-00009-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..06f1ca53a16586002bfa5a3adfc89a93e41bae90 --- /dev/null +++ b/model-00009-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:2073ea7d8a49a93c7e44173c4adbab7e39fc2ec6d5f39eb3723b00c69f46088d +size 4362142816 diff --git a/model-00010-of-00061.safetensors b/model-00010-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..573bd864f2aadea1cda4795c4c843c386fba48f7 --- /dev/null +++ b/model-00010-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8832350ab7ab461489731ec0afc2e835744e767dbbefb7567a8d2864a2b5b30 +size 4966188784 diff --git a/model-00011-of-00061.safetensors b/model-00011-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c5143c8130cce51bcadb566c0f8b372bc6a70dd0 --- /dev/null +++ b/model-00011-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a2e2a60faa574e29514263d19e07c29a321cef79f11019f7ee8cc78296731cf2 +size 4362142816 diff --git a/model-00012-of-00061.safetensors b/model-00012-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e0351ba2e70fe5450932f016cd4e8aa85575c8e6 --- /dev/null +++ b/model-00012-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cc79c9e1472d7d7df6f7c46dc52ce41a89c48e7a13e5fe8e1b2ca462b8808d07 +size 4362142816 diff --git a/model-00013-of-00061.safetensors b/model-00013-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d4b3bd7bf105de7f09d9337df5019d8d1ebb3f5f --- /dev/null +++ b/model-00013-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:e5b3a92cbd429525a2069b75e43ff2c6eabb097638b1a86fa433eeb7f0933493 +size 4966188784 diff --git a/model-00014-of-00061.safetensors b/model-00014-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..95c8d08557f9ee1c87f1ac11504a3ffe8044592d --- /dev/null +++ b/model-00014-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:c3368712115f97cfa793d217691f83d2fd24ba6ac2aec7371085a6d70b68ad42 +size 4362142816 diff --git a/model-00015-of-00061.safetensors b/model-00015-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..64e60f593cbdb740ef4a583549d6103fa40acbdf --- /dev/null +++ b/model-00015-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7dc9ca9eecf97af94e6740fd791d90be987dc8936f9f72ca6ace27b166d63b9b +size 4362142816 diff --git a/model-00016-of-00061.safetensors b/model-00016-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8e62ead3de7bd19d1a42ab77b1155f2f148de77c --- /dev/null +++ b/model-00016-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:33b9c5e26e5e718add0f8550bab58e7add6a42731512a95ec7e1e681b5ee7430 +size 4966188784 diff --git a/model-00017-of-00061.safetensors b/model-00017-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5980d423873b78fc62382732a98c7ed9824a5a65 --- /dev/null +++ b/model-00017-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:473abe364619341b76833953722396a2d340b2b5af2eda0be0fc99a85fd1224e +size 4362142816 diff --git a/model-00018-of-00061.safetensors b/model-00018-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d9b6156ae827f5ebb0140d90bcfd377722802d49 --- /dev/null +++ b/model-00018-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1ec8a5047968f50baccf3141855e35e694a705a935fec7f70167371bba201c30 +size 4362142816 diff --git a/model-00019-of-00061.safetensors b/model-00019-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..6ec662719097bf4ae6f4ddf1a8f2ad5f379fce04 --- /dev/null +++ b/model-00019-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:73b8df5d602071fc41230573d4f2c4006a76575f51c493e36eacd23586669ac4 +size 4966188784 diff --git a/model-00020-of-00061.safetensors b/model-00020-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..d50e08cebdcda1f27aabfcacb2bc3ba511cea598 --- /dev/null +++ b/model-00020-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:04242be147991c7324d6db39eaeb8591550057687c25ce80aa6188cd9558eb14 +size 4362142816 diff --git a/model-00021-of-00061.safetensors b/model-00021-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..c4f4c4ee86bd118925c1fe1906d5baec0241bfa9 --- /dev/null +++ b/model-00021-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:69dca4a3936f0a2c1a740b1d074ff087019c3966ab61836f7dba83feb52ffee7 +size 4362142816 diff --git a/model-00022-of-00061.safetensors b/model-00022-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5dbfaf8b4a942a981016c664b22d984185b203dd --- /dev/null +++ b/model-00022-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:00dcf854e61a3fc31b9693a5a71b818fb63f76886969a6688688a37f74541f1c +size 4966188784 diff --git a/model-00023-of-00061.safetensors b/model-00023-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3089e7a4e00256ec9fbea1c4a9e961806d2d8551 --- /dev/null +++ b/model-00023-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0dbb524337a8dd87401390dbd646c781eadfd93c21ddec51f4eed926344e665c +size 4362142816 diff --git a/model-00024-of-00061.safetensors b/model-00024-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7e8f3c74a768c66e73bd9c3a30d6ad7d11372481 --- /dev/null +++ b/model-00024-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:341b702f3fffc42aca8621699634b5422e1c72cb7da2695622b1080e04a5da50 +size 4362142816 diff --git a/model-00025-of-00061.safetensors b/model-00025-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7bce124f11bf31ed8592a9fa0cbc2afc4b065feb --- /dev/null +++ b/model-00025-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:909b6159e3a722522f28bc964652a752cb0c58f32deb94399f2a038a9debdde4 +size 4966188784 diff --git a/model-00026-of-00061.safetensors b/model-00026-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..5bd22b0847837b0ee8a36a0aab83bafbcd602278 --- /dev/null +++ b/model-00026-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fa95c85e5b0d07fd69479e868f996bbab613e3a817b6e23e15cda893b730b55e +size 4362142816 diff --git a/model-00027-of-00061.safetensors b/model-00027-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..875e3c908ba023e8f3155c3eeb972cc26a9d5189 --- /dev/null +++ b/model-00027-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b0858c6e3c21cf95761b649e394fd2a8de42ab48d84d287544e0e09415b8bbc5 +size 4362142816 diff --git a/model-00028-of-00061.safetensors b/model-00028-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1dbd43d30146ddbedea28125e93a99bdd50e583c --- /dev/null +++ b/model-00028-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:923eea523cb22165fe72eddbbdc7f245d3b3c05b8f933f64a80bed219bb7d34e +size 4966188784 diff --git a/model-00029-of-00061.safetensors b/model-00029-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..4738bbcd1f4831602fa12679590a1ec38203bdda --- /dev/null +++ b/model-00029-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:d78db54b48f71ffcf35371e6d9a425a595d39f111b51fe2d17a8f1355da1e7d4 +size 4362142816 diff --git a/model-00030-of-00061.safetensors b/model-00030-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..45c19605650efc6ba50c60bc9673624b2720092b --- /dev/null +++ b/model-00030-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b77c453c7521a9ed741f27d8f949c6fe1b294c9e60b0fcb48b74902de89d7730 +size 4362142816 diff --git a/model-00031-of-00061.safetensors b/model-00031-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..34a6e65e9795f40ea0bcc7b044a7bc667f6d1d4b --- /dev/null +++ b/model-00031-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb98337f22f0cc676dcfbb5b2e024775252895242bb00349fb916c8efdac0006 +size 4966188784 diff --git a/model-00032-of-00061.safetensors b/model-00032-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..77293f2bd0de8fec0c08c6b4604726a757ad8601 --- /dev/null +++ b/model-00032-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:5795f8e6db642c91139a1f6a2fa5e9e1a36032381a833cb4578ec9653c92cb02 +size 4362142816 diff --git a/model-00033-of-00061.safetensors b/model-00033-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c7846e884bc82ecda71d74f5a3b9f351be0e837 --- /dev/null +++ b/model-00033-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:f7be33f8e3bee4866da7379c6ecde310857faefa1daaf4c93d8b85b9e10094f0 +size 4362142816 diff --git a/model-00034-of-00061.safetensors b/model-00034-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..bac4a993a45c46b07e0749137dfe201e5c7b8bda --- /dev/null +++ b/model-00034-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:59100f3ea9b75f6a916c17f0103a9c0058c82212fa71053ab88b152582335458 +size 4966188784 diff --git a/model-00035-of-00061.safetensors b/model-00035-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..35a39ff5d74ab1f72e6d10d0d19c27178ceedb7d --- /dev/null +++ b/model-00035-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:df78e23bccca255f9244dd30e43c2514e93bbe178e32cce90d3a8ff9a20c3ba9 +size 4362142816 diff --git a/model-00036-of-00061.safetensors b/model-00036-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..39e9a7101886098a2b268da7a83563c11b43af22 --- /dev/null +++ b/model-00036-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1b6c0b019b0786b18e519578a980c0ca83ce053717015500e60c99ea0e048a8f +size 4362142816 diff --git a/model-00037-of-00061.safetensors b/model-00037-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..cbff3afff26b4bc070790d2e673a967b6465b095 --- /dev/null +++ b/model-00037-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:62174964f587c986ea33b06e3e00e1162dfbebb81d780e74508af925ffc94ed8 +size 4966188784 diff --git a/model-00038-of-00061.safetensors b/model-00038-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..1a54f5f9214b3850fb08ac239af0531acc74c3ce --- /dev/null +++ b/model-00038-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:38d0269276af4604d0b13c3b2dfa28f00b85502b61d3d0acf70cdfd400bf65cb +size 4362142816 diff --git a/model-00039-of-00061.safetensors b/model-00039-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e585cec79141cacb9382e39d552d99871831d155 --- /dev/null +++ b/model-00039-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:ec4e440354b207e10744ece13874a0f8aaeba787e1c59d0fffcea4470719aa68 +size 4362142816 diff --git a/model-00040-of-00061.safetensors b/model-00040-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..60976f27637627dc319d93d9c81a30e737240124 --- /dev/null +++ b/model-00040-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4b468cfb229f44ffc05ba3655f5f4cc735022a494313185543a25e9b2f1ce472 +size 4966188784 diff --git a/model-00041-of-00061.safetensors b/model-00041-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2a5646b697dcfd73ca576a63d827ccf2b5690ea6 --- /dev/null +++ b/model-00041-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cdcb4f13a9e05aa8ecf5126a3815309cbb4d46d8ab92a444bcd47ca0c96652af +size 4362142816 diff --git a/model-00042-of-00061.safetensors b/model-00042-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8ee67cc05e0256329b2ff713d6c28414c3173790 --- /dev/null +++ b/model-00042-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:28f614d5e2119e2b8470f090e28de381b181e05032aae426cf51d06ab696436c +size 4362142816 diff --git a/model-00043-of-00061.safetensors b/model-00043-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..711580cd971be8caf5e2cd629ce795ebd949fe38 --- /dev/null +++ b/model-00043-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:a6fcaea5dc0e30425ddbb17247b57b92364cb08432b32bd747ea73ee09a6612f +size 4966188784 diff --git a/model-00044-of-00061.safetensors b/model-00044-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..2b256c1f20317c9ae46ddfe46a70977e33beef04 --- /dev/null +++ b/model-00044-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:1f2f92b856fbd1ed1d07e587953b33d812a542cfcbb6c541313605b223479b4e +size 4362142816 diff --git a/model-00045-of-00061.safetensors b/model-00045-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..8c98ba94c483e83006604b82bc4169deafb23005 --- /dev/null +++ b/model-00045-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:05976f1b7ae6f9fbfe222a88babfc976cea11097b825fe2900efe6583d05eb7e +size 4362142816 diff --git a/model-00046-of-00061.safetensors b/model-00046-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..e3395dbf55680ad82d9426163344aedd9eda1bf8 --- /dev/null +++ b/model-00046-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:8aabbf0f22abb3a7aaad20ffc81e3aaadafb78a05429ef29f0c7389e1b8af973 +size 4966188784 diff --git a/model-00047-of-00061.safetensors b/model-00047-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..325b59c79d9d7c573fb44bc4420bdaefe6183a9d --- /dev/null +++ b/model-00047-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:08505cc1bdbcb2c3c7b3593fcb1f6e22b64e22e7a03984da5abbe879382dcc58 +size 4362142816 diff --git a/model-00048-of-00061.safetensors b/model-00048-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7a2c9422a922c6360f7f177bc1ac0e24b7f5d403 --- /dev/null +++ b/model-00048-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:fec8a1bfdf3f61fb3cfbcc740cdc8bcbef84b6b691b6b6a7a0da765546b75b6c +size 4362142816 diff --git a/model-00049-of-00061.safetensors b/model-00049-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b915f940f4d5a42d612a39f90de57d576d888bb2 --- /dev/null +++ b/model-00049-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dbba322eed5c9b52c0641aad8d5f0d7179274554ef7c813dd082d6e095ee7394 +size 4966188784 diff --git a/model-00050-of-00061.safetensors b/model-00050-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..b7f250aca35460913b843de8ccb2e15f9c5ea0db --- /dev/null +++ b/model-00050-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:0b9759d76765eccec140493d671cc142fd631eb7ce1c9849f126205e5daf29bb +size 4362142816 diff --git a/model-00051-of-00061.safetensors b/model-00051-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3abd62a2859aec8554b310cf311677877314a89b --- /dev/null +++ b/model-00051-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:dac535f0f6682d32490f08b76e0424574d4289ae4d8e3ca3848bdedba15f38aa +size 4362142816 diff --git a/model-00052-of-00061.safetensors b/model-00052-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..3352b5dd5a049dbd7b2e429badb1c7cc2f4e744c --- /dev/null +++ b/model-00052-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b8cd62a3f15a6c6975259bde819fbfd440ad99d2f37f0481be73dfb7c1632703 +size 4966188784 diff --git a/model-00053-of-00061.safetensors b/model-00053-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..28bb4df32d78d91adc35e78e07b491ff91f4ef66 --- /dev/null +++ b/model-00053-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:b2578690e0604df65f5446d966de1273439708d7609be4e5073d0a49dde2b500 +size 4362142816 diff --git a/model-00054-of-00061.safetensors b/model-00054-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f848d30ff57b27ecb39e3675d3fcf38ee9f1120f --- /dev/null +++ b/model-00054-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:bfed84964d91737a2caac66ce8bc2748e90050f4cacf69c71d313476d9d54cd4 +size 4362142816 diff --git a/model-00055-of-00061.safetensors b/model-00055-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..980f7aca3b92ebdd4226a8055f938ba3b7257459 --- /dev/null +++ b/model-00055-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:022f51bd609679ea77dcf9862d0b16d908057354c6fed29e9fcca0a21becf046 +size 4966188784 diff --git a/model-00056-of-00061.safetensors b/model-00056-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..a758448e08de2e286fdf0d70353ce308712a7e80 --- /dev/null +++ b/model-00056-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:eeb45ab4e90b7f2f7bb436c7eca1a6dbaf30cf6d8486e0bb287cd1b2c61d4da4 +size 4362142816 diff --git a/model-00057-of-00061.safetensors b/model-00057-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..f2467af964dedf5f69fea4eb6aa900fb88e15ff2 --- /dev/null +++ b/model-00057-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:7628a401d47fda443c3f4ea9a772df5e90499f8a4c318343ffb979f6c1ec8616 +size 4362142816 diff --git a/model-00058-of-00061.safetensors b/model-00058-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..7777b383f206504edecefaf35467f420852f957b --- /dev/null +++ b/model-00058-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:4c4bbcacacbb9aa120008185f211c085b9d3eaeb29137cff0225e89c74db6e8c +size 4966188784 diff --git a/model-00059-of-00061.safetensors b/model-00059-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..736bdb8c5ad81a7d26ba0490a76fdd0c3d1d7cfb --- /dev/null +++ b/model-00059-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:cb2fdb6fee4f34dff177a92bc30fba656c884bc05d9bc371033e7559b4f04884 +size 4362142816 diff --git a/model-00060-of-00061.safetensors b/model-00060-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..86fea4ab904bf7aee1d056f7a984da4d615b4ed8 --- /dev/null +++ b/model-00060-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:39e15825cc27f6861e7a770604abcdae6ba47d68d14bfb3f5b500c8411a3c152 +size 4362142816 diff --git a/model-00061-of-00061.safetensors b/model-00061-of-00061.safetensors new file mode 100644 index 0000000000000000000000000000000000000000..012409c212f8d7358f2d1af596b1ecc6a0a2eb85 --- /dev/null +++ b/model-00061-of-00061.safetensors @@ -0,0 +1,3 @@ +version https://git-lfs.github.com/spec/v1 +oid sha256:6989aaf2ffcb97a0d9a8a968f5d2c6af83c346e241d1a4b6e58bdc0a59ae6344 +size 4362241416 diff --git a/model.safetensors.index.json b/model.safetensors.index.json new file mode 100644 index 0000000000000000000000000000000000000000..9631684c855400a66171857e28a91f12ffe0af40 --- /dev/null +++ b/model.safetensors.index.json @@ -0,0 +1,729 @@ +{ + "metadata": { + "total_size": 278012395520 + }, + "weight_map": { + "embed_tokens.weight": "model-00001-of-00061.safetensors", + "layers.0.input_layernorm.weight": "model-00002-of-00061.safetensors", + "layers.0.mlp.down_proj.weight": "model-00002-of-00061.safetensors", + "layers.0.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "layers.0.mlp.up_proj.weight": "model-00002-of-00061.safetensors", + "layers.0.post_attention_layernorm.weight": "model-00002-of-00061.safetensors", + "layers.0.self_attn.k_proj.weight": "model-00001-of-00061.safetensors", + "layers.0.self_attn.o_proj.weight": "model-00001-of-00061.safetensors", + "layers.0.self_attn.q_proj.weight": "model-00001-of-00061.safetensors", + "layers.0.self_attn.v_proj.weight": "model-00001-of-00061.safetensors", + "layers.1.input_layernorm.weight": "model-00003-of-00061.safetensors", + "layers.1.mlp.down_proj.weight": "model-00003-of-00061.safetensors", + "layers.1.mlp.gate_proj.weight": "model-00002-of-00061.safetensors", + "layers.1.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "layers.1.post_attention_layernorm.weight": "model-00003-of-00061.safetensors", + "layers.1.self_attn.k_proj.weight": "model-00002-of-00061.safetensors", + "layers.1.self_attn.o_proj.weight": "model-00002-of-00061.safetensors", + "layers.1.self_attn.q_proj.weight": "model-00002-of-00061.safetensors", + "layers.1.self_attn.v_proj.weight": "model-00002-of-00061.safetensors", + "layers.10.input_layernorm.weight": "model-00010-of-00061.safetensors", + "layers.10.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "layers.10.mlp.gate_proj.weight": "model-00009-of-00061.safetensors", + "layers.10.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "layers.10.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "layers.10.self_attn.k_proj.weight": "model-00009-of-00061.safetensors", + "layers.10.self_attn.o_proj.weight": "model-00009-of-00061.safetensors", + "layers.10.self_attn.q_proj.weight": "model-00009-of-00061.safetensors", + "layers.10.self_attn.v_proj.weight": "model-00009-of-00061.safetensors", + "layers.11.input_layernorm.weight": "model-00010-of-00061.safetensors", + "layers.11.mlp.down_proj.weight": "model-00010-of-00061.safetensors", + "layers.11.mlp.gate_proj.weight": "model-00010-of-00061.safetensors", + "layers.11.mlp.up_proj.weight": "model-00010-of-00061.safetensors", + "layers.11.post_attention_layernorm.weight": "model-00010-of-00061.safetensors", + "layers.11.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "layers.11.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "layers.11.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "layers.11.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "layers.12.input_layernorm.weight": "model-00011-of-00061.safetensors", + "layers.12.mlp.down_proj.weight": "model-00011-of-00061.safetensors", + "layers.12.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "layers.12.mlp.up_proj.weight": "model-00011-of-00061.safetensors", + "layers.12.post_attention_layernorm.weight": "model-00011-of-00061.safetensors", + "layers.12.self_attn.k_proj.weight": "model-00010-of-00061.safetensors", + "layers.12.self_attn.o_proj.weight": "model-00010-of-00061.safetensors", + "layers.12.self_attn.q_proj.weight": "model-00010-of-00061.safetensors", + "layers.12.self_attn.v_proj.weight": "model-00010-of-00061.safetensors", + "layers.13.input_layernorm.weight": "model-00012-of-00061.safetensors", + "layers.13.mlp.down_proj.weight": "model-00012-of-00061.safetensors", + "layers.13.mlp.gate_proj.weight": "model-00011-of-00061.safetensors", + "layers.13.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "layers.13.post_attention_layernorm.weight": "model-00012-of-00061.safetensors", + "layers.13.self_attn.k_proj.weight": "model-00011-of-00061.safetensors", + "layers.13.self_attn.o_proj.weight": "model-00011-of-00061.safetensors", + "layers.13.self_attn.q_proj.weight": "model-00011-of-00061.safetensors", + "layers.13.self_attn.v_proj.weight": "model-00011-of-00061.safetensors", + "layers.14.input_layernorm.weight": "model-00013-of-00061.safetensors", + "layers.14.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "layers.14.mlp.gate_proj.weight": "model-00012-of-00061.safetensors", + "layers.14.mlp.up_proj.weight": "model-00012-of-00061.safetensors", + "layers.14.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "layers.14.self_attn.k_proj.weight": "model-00012-of-00061.safetensors", + "layers.14.self_attn.o_proj.weight": "model-00012-of-00061.safetensors", + "layers.14.self_attn.q_proj.weight": "model-00012-of-00061.safetensors", + "layers.14.self_attn.v_proj.weight": "model-00012-of-00061.safetensors", + "layers.15.input_layernorm.weight": "model-00013-of-00061.safetensors", + "layers.15.mlp.down_proj.weight": "model-00013-of-00061.safetensors", + "layers.15.mlp.gate_proj.weight": "model-00013-of-00061.safetensors", + "layers.15.mlp.up_proj.weight": "model-00013-of-00061.safetensors", + "layers.15.post_attention_layernorm.weight": "model-00013-of-00061.safetensors", + "layers.15.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "layers.15.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "layers.15.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "layers.15.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "layers.16.input_layernorm.weight": "model-00014-of-00061.safetensors", + "layers.16.mlp.down_proj.weight": "model-00014-of-00061.safetensors", + "layers.16.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "layers.16.mlp.up_proj.weight": "model-00014-of-00061.safetensors", + "layers.16.post_attention_layernorm.weight": "model-00014-of-00061.safetensors", + "layers.16.self_attn.k_proj.weight": "model-00013-of-00061.safetensors", + "layers.16.self_attn.o_proj.weight": "model-00013-of-00061.safetensors", + "layers.16.self_attn.q_proj.weight": "model-00013-of-00061.safetensors", + "layers.16.self_attn.v_proj.weight": "model-00013-of-00061.safetensors", + "layers.17.input_layernorm.weight": "model-00015-of-00061.safetensors", + "layers.17.mlp.down_proj.weight": "model-00015-of-00061.safetensors", + "layers.17.mlp.gate_proj.weight": "model-00014-of-00061.safetensors", + "layers.17.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "layers.17.post_attention_layernorm.weight": "model-00015-of-00061.safetensors", + "layers.17.self_attn.k_proj.weight": "model-00014-of-00061.safetensors", + "layers.17.self_attn.o_proj.weight": "model-00014-of-00061.safetensors", + "layers.17.self_attn.q_proj.weight": "model-00014-of-00061.safetensors", + "layers.17.self_attn.v_proj.weight": "model-00014-of-00061.safetensors", + "layers.18.input_layernorm.weight": "model-00016-of-00061.safetensors", + "layers.18.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "layers.18.mlp.gate_proj.weight": "model-00015-of-00061.safetensors", + "layers.18.mlp.up_proj.weight": "model-00015-of-00061.safetensors", + "layers.18.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "layers.18.self_attn.k_proj.weight": "model-00015-of-00061.safetensors", + "layers.18.self_attn.o_proj.weight": "model-00015-of-00061.safetensors", + "layers.18.self_attn.q_proj.weight": "model-00015-of-00061.safetensors", + "layers.18.self_attn.v_proj.weight": "model-00015-of-00061.safetensors", + "layers.19.input_layernorm.weight": "model-00016-of-00061.safetensors", + "layers.19.mlp.down_proj.weight": "model-00016-of-00061.safetensors", + "layers.19.mlp.gate_proj.weight": "model-00016-of-00061.safetensors", + "layers.19.mlp.up_proj.weight": "model-00016-of-00061.safetensors", + "layers.19.post_attention_layernorm.weight": "model-00016-of-00061.safetensors", + "layers.19.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "layers.19.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "layers.19.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "layers.19.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "layers.2.input_layernorm.weight": "model-00004-of-00061.safetensors", + "layers.2.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "layers.2.mlp.gate_proj.weight": "model-00003-of-00061.safetensors", + "layers.2.mlp.up_proj.weight": "model-00003-of-00061.safetensors", + "layers.2.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "layers.2.self_attn.k_proj.weight": "model-00003-of-00061.safetensors", + "layers.2.self_attn.o_proj.weight": "model-00003-of-00061.safetensors", + "layers.2.self_attn.q_proj.weight": "model-00003-of-00061.safetensors", + "layers.2.self_attn.v_proj.weight": "model-00003-of-00061.safetensors", + "layers.20.input_layernorm.weight": "model-00017-of-00061.safetensors", + "layers.20.mlp.down_proj.weight": "model-00017-of-00061.safetensors", + "layers.20.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "layers.20.mlp.up_proj.weight": "model-00017-of-00061.safetensors", + "layers.20.post_attention_layernorm.weight": "model-00017-of-00061.safetensors", + "layers.20.self_attn.k_proj.weight": "model-00016-of-00061.safetensors", + "layers.20.self_attn.o_proj.weight": "model-00016-of-00061.safetensors", + "layers.20.self_attn.q_proj.weight": "model-00016-of-00061.safetensors", + "layers.20.self_attn.v_proj.weight": "model-00016-of-00061.safetensors", + "layers.21.input_layernorm.weight": "model-00018-of-00061.safetensors", + "layers.21.mlp.down_proj.weight": "model-00018-of-00061.safetensors", + "layers.21.mlp.gate_proj.weight": "model-00017-of-00061.safetensors", + "layers.21.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "layers.21.post_attention_layernorm.weight": "model-00018-of-00061.safetensors", + "layers.21.self_attn.k_proj.weight": "model-00017-of-00061.safetensors", + "layers.21.self_attn.o_proj.weight": "model-00017-of-00061.safetensors", + "layers.21.self_attn.q_proj.weight": "model-00017-of-00061.safetensors", + "layers.21.self_attn.v_proj.weight": "model-00017-of-00061.safetensors", + "layers.22.input_layernorm.weight": "model-00019-of-00061.safetensors", + "layers.22.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "layers.22.mlp.gate_proj.weight": "model-00018-of-00061.safetensors", + "layers.22.mlp.up_proj.weight": "model-00018-of-00061.safetensors", + "layers.22.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "layers.22.self_attn.k_proj.weight": "model-00018-of-00061.safetensors", + "layers.22.self_attn.o_proj.weight": "model-00018-of-00061.safetensors", + "layers.22.self_attn.q_proj.weight": "model-00018-of-00061.safetensors", + "layers.22.self_attn.v_proj.weight": "model-00018-of-00061.safetensors", + "layers.23.input_layernorm.weight": "model-00019-of-00061.safetensors", + "layers.23.mlp.down_proj.weight": "model-00019-of-00061.safetensors", + "layers.23.mlp.gate_proj.weight": "model-00019-of-00061.safetensors", + "layers.23.mlp.up_proj.weight": "model-00019-of-00061.safetensors", + "layers.23.post_attention_layernorm.weight": "model-00019-of-00061.safetensors", + "layers.23.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "layers.23.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "layers.23.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "layers.23.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "layers.24.input_layernorm.weight": "model-00020-of-00061.safetensors", + "layers.24.mlp.down_proj.weight": "model-00020-of-00061.safetensors", + "layers.24.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "layers.24.mlp.up_proj.weight": "model-00020-of-00061.safetensors", + "layers.24.post_attention_layernorm.weight": "model-00020-of-00061.safetensors", + "layers.24.self_attn.k_proj.weight": "model-00019-of-00061.safetensors", + "layers.24.self_attn.o_proj.weight": "model-00019-of-00061.safetensors", + "layers.24.self_attn.q_proj.weight": "model-00019-of-00061.safetensors", + "layers.24.self_attn.v_proj.weight": "model-00019-of-00061.safetensors", + "layers.25.input_layernorm.weight": "model-00021-of-00061.safetensors", + "layers.25.mlp.down_proj.weight": "model-00021-of-00061.safetensors", + "layers.25.mlp.gate_proj.weight": "model-00020-of-00061.safetensors", + "layers.25.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "layers.25.post_attention_layernorm.weight": "model-00021-of-00061.safetensors", + "layers.25.self_attn.k_proj.weight": "model-00020-of-00061.safetensors", + "layers.25.self_attn.o_proj.weight": "model-00020-of-00061.safetensors", + "layers.25.self_attn.q_proj.weight": "model-00020-of-00061.safetensors", + "layers.25.self_attn.v_proj.weight": "model-00020-of-00061.safetensors", + "layers.26.input_layernorm.weight": "model-00022-of-00061.safetensors", + "layers.26.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "layers.26.mlp.gate_proj.weight": "model-00021-of-00061.safetensors", + "layers.26.mlp.up_proj.weight": "model-00021-of-00061.safetensors", + "layers.26.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "layers.26.self_attn.k_proj.weight": "model-00021-of-00061.safetensors", + "layers.26.self_attn.o_proj.weight": "model-00021-of-00061.safetensors", + "layers.26.self_attn.q_proj.weight": "model-00021-of-00061.safetensors", + "layers.26.self_attn.v_proj.weight": "model-00021-of-00061.safetensors", + "layers.27.input_layernorm.weight": "model-00022-of-00061.safetensors", + "layers.27.mlp.down_proj.weight": "model-00022-of-00061.safetensors", + "layers.27.mlp.gate_proj.weight": "model-00022-of-00061.safetensors", + "layers.27.mlp.up_proj.weight": "model-00022-of-00061.safetensors", + "layers.27.post_attention_layernorm.weight": "model-00022-of-00061.safetensors", + "layers.27.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "layers.27.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "layers.27.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "layers.27.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "layers.28.input_layernorm.weight": "model-00023-of-00061.safetensors", + "layers.28.mlp.down_proj.weight": "model-00023-of-00061.safetensors", + "layers.28.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "layers.28.mlp.up_proj.weight": "model-00023-of-00061.safetensors", + "layers.28.post_attention_layernorm.weight": "model-00023-of-00061.safetensors", + "layers.28.self_attn.k_proj.weight": "model-00022-of-00061.safetensors", + "layers.28.self_attn.o_proj.weight": "model-00022-of-00061.safetensors", + "layers.28.self_attn.q_proj.weight": "model-00022-of-00061.safetensors", + "layers.28.self_attn.v_proj.weight": "model-00022-of-00061.safetensors", + "layers.29.input_layernorm.weight": "model-00024-of-00061.safetensors", + "layers.29.mlp.down_proj.weight": "model-00024-of-00061.safetensors", + "layers.29.mlp.gate_proj.weight": "model-00023-of-00061.safetensors", + "layers.29.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "layers.29.post_attention_layernorm.weight": "model-00024-of-00061.safetensors", + "layers.29.self_attn.k_proj.weight": "model-00023-of-00061.safetensors", + "layers.29.self_attn.o_proj.weight": "model-00023-of-00061.safetensors", + "layers.29.self_attn.q_proj.weight": "model-00023-of-00061.safetensors", + "layers.29.self_attn.v_proj.weight": "model-00023-of-00061.safetensors", + "layers.3.input_layernorm.weight": "model-00004-of-00061.safetensors", + "layers.3.mlp.down_proj.weight": "model-00004-of-00061.safetensors", + "layers.3.mlp.gate_proj.weight": "model-00004-of-00061.safetensors", + "layers.3.mlp.up_proj.weight": "model-00004-of-00061.safetensors", + "layers.3.post_attention_layernorm.weight": "model-00004-of-00061.safetensors", + "layers.3.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "layers.3.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "layers.3.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "layers.3.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "layers.30.input_layernorm.weight": "model-00025-of-00061.safetensors", + "layers.30.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "layers.30.mlp.gate_proj.weight": "model-00024-of-00061.safetensors", + "layers.30.mlp.up_proj.weight": "model-00024-of-00061.safetensors", + "layers.30.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "layers.30.self_attn.k_proj.weight": "model-00024-of-00061.safetensors", + "layers.30.self_attn.o_proj.weight": "model-00024-of-00061.safetensors", + "layers.30.self_attn.q_proj.weight": "model-00024-of-00061.safetensors", + "layers.30.self_attn.v_proj.weight": "model-00024-of-00061.safetensors", + "layers.31.input_layernorm.weight": "model-00025-of-00061.safetensors", + "layers.31.mlp.down_proj.weight": "model-00025-of-00061.safetensors", + "layers.31.mlp.gate_proj.weight": "model-00025-of-00061.safetensors", + "layers.31.mlp.up_proj.weight": "model-00025-of-00061.safetensors", + "layers.31.post_attention_layernorm.weight": "model-00025-of-00061.safetensors", + "layers.31.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "layers.31.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "layers.31.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "layers.31.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "layers.32.input_layernorm.weight": "model-00026-of-00061.safetensors", + "layers.32.mlp.down_proj.weight": "model-00026-of-00061.safetensors", + "layers.32.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "layers.32.mlp.up_proj.weight": "model-00026-of-00061.safetensors", + "layers.32.post_attention_layernorm.weight": "model-00026-of-00061.safetensors", + "layers.32.self_attn.k_proj.weight": "model-00025-of-00061.safetensors", + "layers.32.self_attn.o_proj.weight": "model-00025-of-00061.safetensors", + "layers.32.self_attn.q_proj.weight": "model-00025-of-00061.safetensors", + "layers.32.self_attn.v_proj.weight": "model-00025-of-00061.safetensors", + "layers.33.input_layernorm.weight": "model-00027-of-00061.safetensors", + "layers.33.mlp.down_proj.weight": "model-00027-of-00061.safetensors", + "layers.33.mlp.gate_proj.weight": "model-00026-of-00061.safetensors", + "layers.33.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "layers.33.post_attention_layernorm.weight": "model-00027-of-00061.safetensors", + "layers.33.self_attn.k_proj.weight": "model-00026-of-00061.safetensors", + "layers.33.self_attn.o_proj.weight": "model-00026-of-00061.safetensors", + "layers.33.self_attn.q_proj.weight": "model-00026-of-00061.safetensors", + "layers.33.self_attn.v_proj.weight": "model-00026-of-00061.safetensors", + "layers.34.input_layernorm.weight": "model-00028-of-00061.safetensors", + "layers.34.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "layers.34.mlp.gate_proj.weight": "model-00027-of-00061.safetensors", + "layers.34.mlp.up_proj.weight": "model-00027-of-00061.safetensors", + "layers.34.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "layers.34.self_attn.k_proj.weight": "model-00027-of-00061.safetensors", + "layers.34.self_attn.o_proj.weight": "model-00027-of-00061.safetensors", + "layers.34.self_attn.q_proj.weight": "model-00027-of-00061.safetensors", + "layers.34.self_attn.v_proj.weight": "model-00027-of-00061.safetensors", + "layers.35.input_layernorm.weight": "model-00028-of-00061.safetensors", + "layers.35.mlp.down_proj.weight": "model-00028-of-00061.safetensors", + "layers.35.mlp.gate_proj.weight": "model-00028-of-00061.safetensors", + "layers.35.mlp.up_proj.weight": "model-00028-of-00061.safetensors", + "layers.35.post_attention_layernorm.weight": "model-00028-of-00061.safetensors", + "layers.35.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "layers.35.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "layers.35.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "layers.35.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "layers.36.input_layernorm.weight": "model-00029-of-00061.safetensors", + "layers.36.mlp.down_proj.weight": "model-00029-of-00061.safetensors", + "layers.36.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "layers.36.mlp.up_proj.weight": "model-00029-of-00061.safetensors", + "layers.36.post_attention_layernorm.weight": "model-00029-of-00061.safetensors", + "layers.36.self_attn.k_proj.weight": "model-00028-of-00061.safetensors", + "layers.36.self_attn.o_proj.weight": "model-00028-of-00061.safetensors", + "layers.36.self_attn.q_proj.weight": "model-00028-of-00061.safetensors", + "layers.36.self_attn.v_proj.weight": "model-00028-of-00061.safetensors", + "layers.37.input_layernorm.weight": "model-00030-of-00061.safetensors", + "layers.37.mlp.down_proj.weight": "model-00030-of-00061.safetensors", + "layers.37.mlp.gate_proj.weight": "model-00029-of-00061.safetensors", + "layers.37.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "layers.37.post_attention_layernorm.weight": "model-00030-of-00061.safetensors", + "layers.37.self_attn.k_proj.weight": "model-00029-of-00061.safetensors", + "layers.37.self_attn.o_proj.weight": "model-00029-of-00061.safetensors", + "layers.37.self_attn.q_proj.weight": "model-00029-of-00061.safetensors", + "layers.37.self_attn.v_proj.weight": "model-00029-of-00061.safetensors", + "layers.38.input_layernorm.weight": "model-00031-of-00061.safetensors", + "layers.38.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "layers.38.mlp.gate_proj.weight": "model-00030-of-00061.safetensors", + "layers.38.mlp.up_proj.weight": "model-00030-of-00061.safetensors", + "layers.38.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "layers.38.self_attn.k_proj.weight": "model-00030-of-00061.safetensors", + "layers.38.self_attn.o_proj.weight": "model-00030-of-00061.safetensors", + "layers.38.self_attn.q_proj.weight": "model-00030-of-00061.safetensors", + "layers.38.self_attn.v_proj.weight": "model-00030-of-00061.safetensors", + "layers.39.input_layernorm.weight": "model-00031-of-00061.safetensors", + "layers.39.mlp.down_proj.weight": "model-00031-of-00061.safetensors", + "layers.39.mlp.gate_proj.weight": "model-00031-of-00061.safetensors", + "layers.39.mlp.up_proj.weight": "model-00031-of-00061.safetensors", + "layers.39.post_attention_layernorm.weight": "model-00031-of-00061.safetensors", + "layers.39.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "layers.39.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "layers.39.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "layers.39.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "layers.4.input_layernorm.weight": "model-00005-of-00061.safetensors", + "layers.4.mlp.down_proj.weight": "model-00005-of-00061.safetensors", + "layers.4.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "layers.4.mlp.up_proj.weight": "model-00005-of-00061.safetensors", + "layers.4.post_attention_layernorm.weight": "model-00005-of-00061.safetensors", + "layers.4.self_attn.k_proj.weight": "model-00004-of-00061.safetensors", + "layers.4.self_attn.o_proj.weight": "model-00004-of-00061.safetensors", + "layers.4.self_attn.q_proj.weight": "model-00004-of-00061.safetensors", + "layers.4.self_attn.v_proj.weight": "model-00004-of-00061.safetensors", + "layers.40.input_layernorm.weight": "model-00032-of-00061.safetensors", + "layers.40.mlp.down_proj.weight": "model-00032-of-00061.safetensors", + "layers.40.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "layers.40.mlp.up_proj.weight": "model-00032-of-00061.safetensors", + "layers.40.post_attention_layernorm.weight": "model-00032-of-00061.safetensors", + "layers.40.self_attn.k_proj.weight": "model-00031-of-00061.safetensors", + "layers.40.self_attn.o_proj.weight": "model-00031-of-00061.safetensors", + "layers.40.self_attn.q_proj.weight": "model-00031-of-00061.safetensors", + "layers.40.self_attn.v_proj.weight": "model-00031-of-00061.safetensors", + "layers.41.input_layernorm.weight": "model-00033-of-00061.safetensors", + "layers.41.mlp.down_proj.weight": "model-00033-of-00061.safetensors", + "layers.41.mlp.gate_proj.weight": "model-00032-of-00061.safetensors", + "layers.41.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "layers.41.post_attention_layernorm.weight": "model-00033-of-00061.safetensors", + "layers.41.self_attn.k_proj.weight": "model-00032-of-00061.safetensors", + "layers.41.self_attn.o_proj.weight": "model-00032-of-00061.safetensors", + "layers.41.self_attn.q_proj.weight": "model-00032-of-00061.safetensors", + "layers.41.self_attn.v_proj.weight": "model-00032-of-00061.safetensors", + "layers.42.input_layernorm.weight": "model-00034-of-00061.safetensors", + "layers.42.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "layers.42.mlp.gate_proj.weight": "model-00033-of-00061.safetensors", + "layers.42.mlp.up_proj.weight": "model-00033-of-00061.safetensors", + "layers.42.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "layers.42.self_attn.k_proj.weight": "model-00033-of-00061.safetensors", + "layers.42.self_attn.o_proj.weight": "model-00033-of-00061.safetensors", + "layers.42.self_attn.q_proj.weight": "model-00033-of-00061.safetensors", + "layers.42.self_attn.v_proj.weight": "model-00033-of-00061.safetensors", + "layers.43.input_layernorm.weight": "model-00034-of-00061.safetensors", + "layers.43.mlp.down_proj.weight": "model-00034-of-00061.safetensors", + "layers.43.mlp.gate_proj.weight": "model-00034-of-00061.safetensors", + "layers.43.mlp.up_proj.weight": "model-00034-of-00061.safetensors", + "layers.43.post_attention_layernorm.weight": "model-00034-of-00061.safetensors", + "layers.43.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "layers.43.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "layers.43.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "layers.43.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "layers.44.input_layernorm.weight": "model-00035-of-00061.safetensors", + "layers.44.mlp.down_proj.weight": "model-00035-of-00061.safetensors", + "layers.44.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "layers.44.mlp.up_proj.weight": "model-00035-of-00061.safetensors", + "layers.44.post_attention_layernorm.weight": "model-00035-of-00061.safetensors", + "layers.44.self_attn.k_proj.weight": "model-00034-of-00061.safetensors", + "layers.44.self_attn.o_proj.weight": "model-00034-of-00061.safetensors", + "layers.44.self_attn.q_proj.weight": "model-00034-of-00061.safetensors", + "layers.44.self_attn.v_proj.weight": "model-00034-of-00061.safetensors", + "layers.45.input_layernorm.weight": "model-00036-of-00061.safetensors", + "layers.45.mlp.down_proj.weight": "model-00036-of-00061.safetensors", + "layers.45.mlp.gate_proj.weight": "model-00035-of-00061.safetensors", + "layers.45.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "layers.45.post_attention_layernorm.weight": "model-00036-of-00061.safetensors", + "layers.45.self_attn.k_proj.weight": "model-00035-of-00061.safetensors", + "layers.45.self_attn.o_proj.weight": "model-00035-of-00061.safetensors", + "layers.45.self_attn.q_proj.weight": "model-00035-of-00061.safetensors", + "layers.45.self_attn.v_proj.weight": "model-00035-of-00061.safetensors", + "layers.46.input_layernorm.weight": "model-00037-of-00061.safetensors", + "layers.46.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "layers.46.mlp.gate_proj.weight": "model-00036-of-00061.safetensors", + "layers.46.mlp.up_proj.weight": "model-00036-of-00061.safetensors", + "layers.46.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "layers.46.self_attn.k_proj.weight": "model-00036-of-00061.safetensors", + "layers.46.self_attn.o_proj.weight": "model-00036-of-00061.safetensors", + "layers.46.self_attn.q_proj.weight": "model-00036-of-00061.safetensors", + "layers.46.self_attn.v_proj.weight": "model-00036-of-00061.safetensors", + "layers.47.input_layernorm.weight": "model-00037-of-00061.safetensors", + "layers.47.mlp.down_proj.weight": "model-00037-of-00061.safetensors", + "layers.47.mlp.gate_proj.weight": "model-00037-of-00061.safetensors", + "layers.47.mlp.up_proj.weight": "model-00037-of-00061.safetensors", + "layers.47.post_attention_layernorm.weight": "model-00037-of-00061.safetensors", + "layers.47.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "layers.47.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "layers.47.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "layers.47.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "layers.48.input_layernorm.weight": "model-00038-of-00061.safetensors", + "layers.48.mlp.down_proj.weight": "model-00038-of-00061.safetensors", + "layers.48.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "layers.48.mlp.up_proj.weight": "model-00038-of-00061.safetensors", + "layers.48.post_attention_layernorm.weight": "model-00038-of-00061.safetensors", + "layers.48.self_attn.k_proj.weight": "model-00037-of-00061.safetensors", + "layers.48.self_attn.o_proj.weight": "model-00037-of-00061.safetensors", + "layers.48.self_attn.q_proj.weight": "model-00037-of-00061.safetensors", + "layers.48.self_attn.v_proj.weight": "model-00037-of-00061.safetensors", + "layers.49.input_layernorm.weight": "model-00039-of-00061.safetensors", + "layers.49.mlp.down_proj.weight": "model-00039-of-00061.safetensors", + "layers.49.mlp.gate_proj.weight": "model-00038-of-00061.safetensors", + "layers.49.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "layers.49.post_attention_layernorm.weight": "model-00039-of-00061.safetensors", + "layers.49.self_attn.k_proj.weight": "model-00038-of-00061.safetensors", + "layers.49.self_attn.o_proj.weight": "model-00038-of-00061.safetensors", + "layers.49.self_attn.q_proj.weight": "model-00038-of-00061.safetensors", + "layers.49.self_attn.v_proj.weight": "model-00038-of-00061.safetensors", + "layers.5.input_layernorm.weight": "model-00006-of-00061.safetensors", + "layers.5.mlp.down_proj.weight": "model-00006-of-00061.safetensors", + "layers.5.mlp.gate_proj.weight": "model-00005-of-00061.safetensors", + "layers.5.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "layers.5.post_attention_layernorm.weight": "model-00006-of-00061.safetensors", + "layers.5.self_attn.k_proj.weight": "model-00005-of-00061.safetensors", + "layers.5.self_attn.o_proj.weight": "model-00005-of-00061.safetensors", + "layers.5.self_attn.q_proj.weight": "model-00005-of-00061.safetensors", + "layers.5.self_attn.v_proj.weight": "model-00005-of-00061.safetensors", + "layers.50.input_layernorm.weight": "model-00040-of-00061.safetensors", + "layers.50.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "layers.50.mlp.gate_proj.weight": "model-00039-of-00061.safetensors", + "layers.50.mlp.up_proj.weight": "model-00039-of-00061.safetensors", + "layers.50.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "layers.50.self_attn.k_proj.weight": "model-00039-of-00061.safetensors", + "layers.50.self_attn.o_proj.weight": "model-00039-of-00061.safetensors", + "layers.50.self_attn.q_proj.weight": "model-00039-of-00061.safetensors", + "layers.50.self_attn.v_proj.weight": "model-00039-of-00061.safetensors", + "layers.51.input_layernorm.weight": "model-00040-of-00061.safetensors", + "layers.51.mlp.down_proj.weight": "model-00040-of-00061.safetensors", + "layers.51.mlp.gate_proj.weight": "model-00040-of-00061.safetensors", + "layers.51.mlp.up_proj.weight": "model-00040-of-00061.safetensors", + "layers.51.post_attention_layernorm.weight": "model-00040-of-00061.safetensors", + "layers.51.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "layers.51.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "layers.51.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "layers.51.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "layers.52.input_layernorm.weight": "model-00041-of-00061.safetensors", + "layers.52.mlp.down_proj.weight": "model-00041-of-00061.safetensors", + "layers.52.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "layers.52.mlp.up_proj.weight": "model-00041-of-00061.safetensors", + "layers.52.post_attention_layernorm.weight": "model-00041-of-00061.safetensors", + "layers.52.self_attn.k_proj.weight": "model-00040-of-00061.safetensors", + "layers.52.self_attn.o_proj.weight": "model-00040-of-00061.safetensors", + "layers.52.self_attn.q_proj.weight": "model-00040-of-00061.safetensors", + "layers.52.self_attn.v_proj.weight": "model-00040-of-00061.safetensors", + "layers.53.input_layernorm.weight": "model-00042-of-00061.safetensors", + "layers.53.mlp.down_proj.weight": "model-00042-of-00061.safetensors", + "layers.53.mlp.gate_proj.weight": "model-00041-of-00061.safetensors", + "layers.53.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "layers.53.post_attention_layernorm.weight": "model-00042-of-00061.safetensors", + "layers.53.self_attn.k_proj.weight": "model-00041-of-00061.safetensors", + "layers.53.self_attn.o_proj.weight": "model-00041-of-00061.safetensors", + "layers.53.self_attn.q_proj.weight": "model-00041-of-00061.safetensors", + "layers.53.self_attn.v_proj.weight": "model-00041-of-00061.safetensors", + "layers.54.input_layernorm.weight": "model-00043-of-00061.safetensors", + "layers.54.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "layers.54.mlp.gate_proj.weight": "model-00042-of-00061.safetensors", + "layers.54.mlp.up_proj.weight": "model-00042-of-00061.safetensors", + "layers.54.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "layers.54.self_attn.k_proj.weight": "model-00042-of-00061.safetensors", + "layers.54.self_attn.o_proj.weight": "model-00042-of-00061.safetensors", + "layers.54.self_attn.q_proj.weight": "model-00042-of-00061.safetensors", + "layers.54.self_attn.v_proj.weight": "model-00042-of-00061.safetensors", + "layers.55.input_layernorm.weight": "model-00043-of-00061.safetensors", + "layers.55.mlp.down_proj.weight": "model-00043-of-00061.safetensors", + "layers.55.mlp.gate_proj.weight": "model-00043-of-00061.safetensors", + "layers.55.mlp.up_proj.weight": "model-00043-of-00061.safetensors", + "layers.55.post_attention_layernorm.weight": "model-00043-of-00061.safetensors", + "layers.55.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "layers.55.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "layers.55.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "layers.55.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "layers.56.input_layernorm.weight": "model-00044-of-00061.safetensors", + "layers.56.mlp.down_proj.weight": "model-00044-of-00061.safetensors", + "layers.56.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "layers.56.mlp.up_proj.weight": "model-00044-of-00061.safetensors", + "layers.56.post_attention_layernorm.weight": "model-00044-of-00061.safetensors", + "layers.56.self_attn.k_proj.weight": "model-00043-of-00061.safetensors", + "layers.56.self_attn.o_proj.weight": "model-00043-of-00061.safetensors", + "layers.56.self_attn.q_proj.weight": "model-00043-of-00061.safetensors", + "layers.56.self_attn.v_proj.weight": "model-00043-of-00061.safetensors", + "layers.57.input_layernorm.weight": "model-00045-of-00061.safetensors", + "layers.57.mlp.down_proj.weight": "model-00045-of-00061.safetensors", + "layers.57.mlp.gate_proj.weight": "model-00044-of-00061.safetensors", + "layers.57.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "layers.57.post_attention_layernorm.weight": "model-00045-of-00061.safetensors", + "layers.57.self_attn.k_proj.weight": "model-00044-of-00061.safetensors", + "layers.57.self_attn.o_proj.weight": "model-00044-of-00061.safetensors", + "layers.57.self_attn.q_proj.weight": "model-00044-of-00061.safetensors", + "layers.57.self_attn.v_proj.weight": "model-00044-of-00061.safetensors", + "layers.58.input_layernorm.weight": "model-00046-of-00061.safetensors", + "layers.58.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "layers.58.mlp.gate_proj.weight": "model-00045-of-00061.safetensors", + "layers.58.mlp.up_proj.weight": "model-00045-of-00061.safetensors", + "layers.58.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "layers.58.self_attn.k_proj.weight": "model-00045-of-00061.safetensors", + "layers.58.self_attn.o_proj.weight": "model-00045-of-00061.safetensors", + "layers.58.self_attn.q_proj.weight": "model-00045-of-00061.safetensors", + "layers.58.self_attn.v_proj.weight": "model-00045-of-00061.safetensors", + "layers.59.input_layernorm.weight": "model-00046-of-00061.safetensors", + "layers.59.mlp.down_proj.weight": "model-00046-of-00061.safetensors", + "layers.59.mlp.gate_proj.weight": "model-00046-of-00061.safetensors", + "layers.59.mlp.up_proj.weight": "model-00046-of-00061.safetensors", + "layers.59.post_attention_layernorm.weight": "model-00046-of-00061.safetensors", + "layers.59.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "layers.59.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "layers.59.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "layers.59.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "layers.6.input_layernorm.weight": "model-00007-of-00061.safetensors", + "layers.6.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "layers.6.mlp.gate_proj.weight": "model-00006-of-00061.safetensors", + "layers.6.mlp.up_proj.weight": "model-00006-of-00061.safetensors", + "layers.6.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "layers.6.self_attn.k_proj.weight": "model-00006-of-00061.safetensors", + "layers.6.self_attn.o_proj.weight": "model-00006-of-00061.safetensors", + "layers.6.self_attn.q_proj.weight": "model-00006-of-00061.safetensors", + "layers.6.self_attn.v_proj.weight": "model-00006-of-00061.safetensors", + "layers.60.input_layernorm.weight": "model-00047-of-00061.safetensors", + "layers.60.mlp.down_proj.weight": "model-00047-of-00061.safetensors", + "layers.60.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "layers.60.mlp.up_proj.weight": "model-00047-of-00061.safetensors", + "layers.60.post_attention_layernorm.weight": "model-00047-of-00061.safetensors", + "layers.60.self_attn.k_proj.weight": "model-00046-of-00061.safetensors", + "layers.60.self_attn.o_proj.weight": "model-00046-of-00061.safetensors", + "layers.60.self_attn.q_proj.weight": "model-00046-of-00061.safetensors", + "layers.60.self_attn.v_proj.weight": "model-00046-of-00061.safetensors", + "layers.61.input_layernorm.weight": "model-00048-of-00061.safetensors", + "layers.61.mlp.down_proj.weight": "model-00048-of-00061.safetensors", + "layers.61.mlp.gate_proj.weight": "model-00047-of-00061.safetensors", + "layers.61.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "layers.61.post_attention_layernorm.weight": "model-00048-of-00061.safetensors", + "layers.61.self_attn.k_proj.weight": "model-00047-of-00061.safetensors", + "layers.61.self_attn.o_proj.weight": "model-00047-of-00061.safetensors", + "layers.61.self_attn.q_proj.weight": "model-00047-of-00061.safetensors", + "layers.61.self_attn.v_proj.weight": "model-00047-of-00061.safetensors", + "layers.62.input_layernorm.weight": "model-00049-of-00061.safetensors", + "layers.62.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "layers.62.mlp.gate_proj.weight": "model-00048-of-00061.safetensors", + "layers.62.mlp.up_proj.weight": "model-00048-of-00061.safetensors", + "layers.62.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "layers.62.self_attn.k_proj.weight": "model-00048-of-00061.safetensors", + "layers.62.self_attn.o_proj.weight": "model-00048-of-00061.safetensors", + "layers.62.self_attn.q_proj.weight": "model-00048-of-00061.safetensors", + "layers.62.self_attn.v_proj.weight": "model-00048-of-00061.safetensors", + "layers.63.input_layernorm.weight": "model-00049-of-00061.safetensors", + "layers.63.mlp.down_proj.weight": "model-00049-of-00061.safetensors", + "layers.63.mlp.gate_proj.weight": "model-00049-of-00061.safetensors", + "layers.63.mlp.up_proj.weight": "model-00049-of-00061.safetensors", + "layers.63.post_attention_layernorm.weight": "model-00049-of-00061.safetensors", + "layers.63.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "layers.63.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "layers.63.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "layers.63.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "layers.64.input_layernorm.weight": "model-00050-of-00061.safetensors", + "layers.64.mlp.down_proj.weight": "model-00050-of-00061.safetensors", + "layers.64.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "layers.64.mlp.up_proj.weight": "model-00050-of-00061.safetensors", + "layers.64.post_attention_layernorm.weight": "model-00050-of-00061.safetensors", + "layers.64.self_attn.k_proj.weight": "model-00049-of-00061.safetensors", + "layers.64.self_attn.o_proj.weight": "model-00049-of-00061.safetensors", + "layers.64.self_attn.q_proj.weight": "model-00049-of-00061.safetensors", + "layers.64.self_attn.v_proj.weight": "model-00049-of-00061.safetensors", + "layers.65.input_layernorm.weight": "model-00051-of-00061.safetensors", + "layers.65.mlp.down_proj.weight": "model-00051-of-00061.safetensors", + "layers.65.mlp.gate_proj.weight": "model-00050-of-00061.safetensors", + "layers.65.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "layers.65.post_attention_layernorm.weight": "model-00051-of-00061.safetensors", + "layers.65.self_attn.k_proj.weight": "model-00050-of-00061.safetensors", + "layers.65.self_attn.o_proj.weight": "model-00050-of-00061.safetensors", + "layers.65.self_attn.q_proj.weight": "model-00050-of-00061.safetensors", + "layers.65.self_attn.v_proj.weight": "model-00050-of-00061.safetensors", + "layers.66.input_layernorm.weight": "model-00052-of-00061.safetensors", + "layers.66.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "layers.66.mlp.gate_proj.weight": "model-00051-of-00061.safetensors", + "layers.66.mlp.up_proj.weight": "model-00051-of-00061.safetensors", + "layers.66.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "layers.66.self_attn.k_proj.weight": "model-00051-of-00061.safetensors", + "layers.66.self_attn.o_proj.weight": "model-00051-of-00061.safetensors", + "layers.66.self_attn.q_proj.weight": "model-00051-of-00061.safetensors", + "layers.66.self_attn.v_proj.weight": "model-00051-of-00061.safetensors", + "layers.67.input_layernorm.weight": "model-00052-of-00061.safetensors", + "layers.67.mlp.down_proj.weight": "model-00052-of-00061.safetensors", + "layers.67.mlp.gate_proj.weight": "model-00052-of-00061.safetensors", + "layers.67.mlp.up_proj.weight": "model-00052-of-00061.safetensors", + "layers.67.post_attention_layernorm.weight": "model-00052-of-00061.safetensors", + "layers.67.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "layers.67.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "layers.67.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "layers.67.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "layers.68.input_layernorm.weight": "model-00053-of-00061.safetensors", + "layers.68.mlp.down_proj.weight": "model-00053-of-00061.safetensors", + "layers.68.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "layers.68.mlp.up_proj.weight": "model-00053-of-00061.safetensors", + "layers.68.post_attention_layernorm.weight": "model-00053-of-00061.safetensors", + "layers.68.self_attn.k_proj.weight": "model-00052-of-00061.safetensors", + "layers.68.self_attn.o_proj.weight": "model-00052-of-00061.safetensors", + "layers.68.self_attn.q_proj.weight": "model-00052-of-00061.safetensors", + "layers.68.self_attn.v_proj.weight": "model-00052-of-00061.safetensors", + "layers.69.input_layernorm.weight": "model-00054-of-00061.safetensors", + "layers.69.mlp.down_proj.weight": "model-00054-of-00061.safetensors", + "layers.69.mlp.gate_proj.weight": "model-00053-of-00061.safetensors", + "layers.69.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "layers.69.post_attention_layernorm.weight": "model-00054-of-00061.safetensors", + "layers.69.self_attn.k_proj.weight": "model-00053-of-00061.safetensors", + "layers.69.self_attn.o_proj.weight": "model-00053-of-00061.safetensors", + "layers.69.self_attn.q_proj.weight": "model-00053-of-00061.safetensors", + "layers.69.self_attn.v_proj.weight": "model-00053-of-00061.safetensors", + "layers.7.input_layernorm.weight": "model-00007-of-00061.safetensors", + "layers.7.mlp.down_proj.weight": "model-00007-of-00061.safetensors", + "layers.7.mlp.gate_proj.weight": "model-00007-of-00061.safetensors", + "layers.7.mlp.up_proj.weight": "model-00007-of-00061.safetensors", + "layers.7.post_attention_layernorm.weight": "model-00007-of-00061.safetensors", + "layers.7.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "layers.7.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "layers.7.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "layers.7.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "layers.70.input_layernorm.weight": "model-00055-of-00061.safetensors", + "layers.70.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "layers.70.mlp.gate_proj.weight": "model-00054-of-00061.safetensors", + "layers.70.mlp.up_proj.weight": "model-00054-of-00061.safetensors", + "layers.70.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "layers.70.self_attn.k_proj.weight": "model-00054-of-00061.safetensors", + "layers.70.self_attn.o_proj.weight": "model-00054-of-00061.safetensors", + "layers.70.self_attn.q_proj.weight": "model-00054-of-00061.safetensors", + "layers.70.self_attn.v_proj.weight": "model-00054-of-00061.safetensors", + "layers.71.input_layernorm.weight": "model-00055-of-00061.safetensors", + "layers.71.mlp.down_proj.weight": "model-00055-of-00061.safetensors", + "layers.71.mlp.gate_proj.weight": "model-00055-of-00061.safetensors", + "layers.71.mlp.up_proj.weight": "model-00055-of-00061.safetensors", + "layers.71.post_attention_layernorm.weight": "model-00055-of-00061.safetensors", + "layers.71.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "layers.71.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "layers.71.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "layers.71.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "layers.72.input_layernorm.weight": "model-00056-of-00061.safetensors", + "layers.72.mlp.down_proj.weight": "model-00056-of-00061.safetensors", + "layers.72.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "layers.72.mlp.up_proj.weight": "model-00056-of-00061.safetensors", + "layers.72.post_attention_layernorm.weight": "model-00056-of-00061.safetensors", + "layers.72.self_attn.k_proj.weight": "model-00055-of-00061.safetensors", + "layers.72.self_attn.o_proj.weight": "model-00055-of-00061.safetensors", + "layers.72.self_attn.q_proj.weight": "model-00055-of-00061.safetensors", + "layers.72.self_attn.v_proj.weight": "model-00055-of-00061.safetensors", + "layers.73.input_layernorm.weight": "model-00057-of-00061.safetensors", + "layers.73.mlp.down_proj.weight": "model-00057-of-00061.safetensors", + "layers.73.mlp.gate_proj.weight": "model-00056-of-00061.safetensors", + "layers.73.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "layers.73.post_attention_layernorm.weight": "model-00057-of-00061.safetensors", + "layers.73.self_attn.k_proj.weight": "model-00056-of-00061.safetensors", + "layers.73.self_attn.o_proj.weight": "model-00056-of-00061.safetensors", + "layers.73.self_attn.q_proj.weight": "model-00056-of-00061.safetensors", + "layers.73.self_attn.v_proj.weight": "model-00056-of-00061.safetensors", + "layers.74.input_layernorm.weight": "model-00058-of-00061.safetensors", + "layers.74.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "layers.74.mlp.gate_proj.weight": "model-00057-of-00061.safetensors", + "layers.74.mlp.up_proj.weight": "model-00057-of-00061.safetensors", + "layers.74.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "layers.74.self_attn.k_proj.weight": "model-00057-of-00061.safetensors", + "layers.74.self_attn.o_proj.weight": "model-00057-of-00061.safetensors", + "layers.74.self_attn.q_proj.weight": "model-00057-of-00061.safetensors", + "layers.74.self_attn.v_proj.weight": "model-00057-of-00061.safetensors", + "layers.75.input_layernorm.weight": "model-00058-of-00061.safetensors", + "layers.75.mlp.down_proj.weight": "model-00058-of-00061.safetensors", + "layers.75.mlp.gate_proj.weight": "model-00058-of-00061.safetensors", + "layers.75.mlp.up_proj.weight": "model-00058-of-00061.safetensors", + "layers.75.post_attention_layernorm.weight": "model-00058-of-00061.safetensors", + "layers.75.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "layers.75.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "layers.75.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "layers.75.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "layers.76.input_layernorm.weight": "model-00059-of-00061.safetensors", + "layers.76.mlp.down_proj.weight": "model-00059-of-00061.safetensors", + "layers.76.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "layers.76.mlp.up_proj.weight": "model-00059-of-00061.safetensors", + "layers.76.post_attention_layernorm.weight": "model-00059-of-00061.safetensors", + "layers.76.self_attn.k_proj.weight": "model-00058-of-00061.safetensors", + "layers.76.self_attn.o_proj.weight": "model-00058-of-00061.safetensors", + "layers.76.self_attn.q_proj.weight": "model-00058-of-00061.safetensors", + "layers.76.self_attn.v_proj.weight": "model-00058-of-00061.safetensors", + "layers.77.input_layernorm.weight": "model-00060-of-00061.safetensors", + "layers.77.mlp.down_proj.weight": "model-00060-of-00061.safetensors", + "layers.77.mlp.gate_proj.weight": "model-00059-of-00061.safetensors", + "layers.77.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "layers.77.post_attention_layernorm.weight": "model-00060-of-00061.safetensors", + "layers.77.self_attn.k_proj.weight": "model-00059-of-00061.safetensors", + "layers.77.self_attn.o_proj.weight": "model-00059-of-00061.safetensors", + "layers.77.self_attn.q_proj.weight": "model-00059-of-00061.safetensors", + "layers.77.self_attn.v_proj.weight": "model-00059-of-00061.safetensors", + "layers.78.input_layernorm.weight": "model-00061-of-00061.safetensors", + "layers.78.mlp.down_proj.weight": "model-00061-of-00061.safetensors", + "layers.78.mlp.gate_proj.weight": "model-00060-of-00061.safetensors", + "layers.78.mlp.up_proj.weight": "model-00060-of-00061.safetensors", + "layers.78.post_attention_layernorm.weight": "model-00061-of-00061.safetensors", + "layers.78.self_attn.k_proj.weight": "model-00060-of-00061.safetensors", + "layers.78.self_attn.o_proj.weight": "model-00060-of-00061.safetensors", + "layers.78.self_attn.q_proj.weight": "model-00060-of-00061.safetensors", + "layers.78.self_attn.v_proj.weight": "model-00060-of-00061.safetensors", + "layers.79.input_layernorm.weight": "model-00061-of-00061.safetensors", + "layers.79.mlp.down_proj.weight": "model-00061-of-00061.safetensors", + "layers.79.mlp.gate_proj.weight": "model-00061-of-00061.safetensors", + "layers.79.mlp.up_proj.weight": "model-00061-of-00061.safetensors", + "layers.79.post_attention_layernorm.weight": "model-00061-of-00061.safetensors", + "layers.79.self_attn.k_proj.weight": "model-00061-of-00061.safetensors", + "layers.79.self_attn.o_proj.weight": "model-00061-of-00061.safetensors", + "layers.79.self_attn.q_proj.weight": "model-00061-of-00061.safetensors", + "layers.79.self_attn.v_proj.weight": "model-00061-of-00061.safetensors", + "layers.8.input_layernorm.weight": "model-00008-of-00061.safetensors", + "layers.8.mlp.down_proj.weight": "model-00008-of-00061.safetensors", + "layers.8.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "layers.8.mlp.up_proj.weight": "model-00008-of-00061.safetensors", + "layers.8.post_attention_layernorm.weight": "model-00008-of-00061.safetensors", + "layers.8.self_attn.k_proj.weight": "model-00007-of-00061.safetensors", + "layers.8.self_attn.o_proj.weight": "model-00007-of-00061.safetensors", + "layers.8.self_attn.q_proj.weight": "model-00007-of-00061.safetensors", + "layers.8.self_attn.v_proj.weight": "model-00007-of-00061.safetensors", + "layers.9.input_layernorm.weight": "model-00009-of-00061.safetensors", + "layers.9.mlp.down_proj.weight": "model-00009-of-00061.safetensors", + "layers.9.mlp.gate_proj.weight": "model-00008-of-00061.safetensors", + "layers.9.mlp.up_proj.weight": "model-00009-of-00061.safetensors", + "layers.9.post_attention_layernorm.weight": "model-00009-of-00061.safetensors", + "layers.9.self_attn.k_proj.weight": "model-00008-of-00061.safetensors", + "layers.9.self_attn.o_proj.weight": "model-00008-of-00061.safetensors", + "layers.9.self_attn.q_proj.weight": "model-00008-of-00061.safetensors", + "layers.9.self_attn.v_proj.weight": "model-00008-of-00061.safetensors", + "norm.weight": "model-00061-of-00061.safetensors" + } +}