diff --git a/README.md b/README.md
index 154df8298fab5ecf322016157858e08cd1bccbe1..1fc0eba30f5d1b3cc856457fb7fddc5e6fdd6ac2 100644
--- a/README.md
+++ b/README.md
@@ -1,3 +1,153 @@
---
license: apache-2.0
+library_name: peft
+tags:
+- generated_from_trainer
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+model-index:
+- name: qlora-out
+ results: []
---
+
+
+
+[](https://github.com/OpenAccess-AI-Collective/axolotl)
+See axolotl config
+
+axolotl version: `0.4.0`
+```yaml
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+model_type: LlamaForCausalLM
+tokenizer_type: LlamaTokenizer
+
+load_in_8bit: false
+load_in_4bit: true
+strict: false
+
+datasets:
+ - path: totally-not-an-llm/ZorgonChat
+ type: alpaca
+dataset_prepared_path:
+val_set_size: 0.05
+output_dir: ./qlora-out
+
+adapter: qlora
+lora_model_dir:
+
+sequence_len: 4096
+sample_packing: false
+pad_to_sequence_len: true
+
+lora_r: 32
+lora_alpha: 16
+lora_dropout: 0.05
+lora_target_modules:
+lora_target_linear: true
+lora_fan_in_fan_out:
+
+wandb_project:
+wandb_entity:
+wandb_watch:
+wandb_name:
+wandb_log_model:
+
+gradient_accumulation_steps: 4
+micro_batch_size: 2
+num_epochs: 4
+optimizer: paged_adamw_32bit
+lr_scheduler: cosine
+learning_rate: 0.0002
+
+train_on_inputs: false
+group_by_length: false
+bf16: auto
+fp16:
+tf32: false
+
+gradient_checkpointing: true
+early_stopping_patience:
+resume_from_checkpoint:
+local_rank:
+logging_steps: 1
+xformers_attention:
+flash_attention: true
+
+warmup_steps: 10
+evals_per_epoch: 4
+saves_per_epoch: 1
+debug:
+deepspeed:
+weight_decay: 0.0
+fsdp:
+fsdp_config:
+special_tokens:
+
+```
+
+
+
+# qlora-out
+
+This model is a fine-tuned version of [TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T](https://huggingface.co/TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T) on the None dataset.
+It achieves the following results on the evaluation set:
+- Loss: 2.3466
+
+## Model description
+
+More information needed
+
+## Intended uses & limitations
+
+More information needed
+
+## Training and evaluation data
+
+More information needed
+
+## Training procedure
+
+### Training hyperparameters
+
+The following hyperparameters were used during training:
+- learning_rate: 0.0002
+- train_batch_size: 2
+- eval_batch_size: 2
+- seed: 42
+- gradient_accumulation_steps: 4
+- total_train_batch_size: 8
+- optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
+- lr_scheduler_type: cosine
+- lr_scheduler_warmup_steps: 10
+- num_epochs: 4
+
+### Training results
+
+| Training Loss | Epoch | Step | Validation Loss |
+|:-------------:|:-----:|:----:|:---------------:|
+| 3.9295 | 0.03 | 1 | 3.9073 |
+| 3.5364 | 0.25 | 8 | 3.6199 |
+| 3.263 | 0.5 | 16 | 3.1821 |
+| 2.798 | 0.75 | 24 | 2.8962 |
+| 2.7787 | 1.0 | 32 | 2.6773 |
+| 2.5959 | 1.25 | 40 | 2.5506 |
+| 2.4793 | 1.5 | 48 | 2.4955 |
+| 2.5221 | 1.75 | 56 | 2.4613 |
+| 2.4384 | 2.0 | 64 | 2.4055 |
+| 2.295 | 2.25 | 72 | 2.3923 |
+| 2.3943 | 2.5 | 80 | 2.3862 |
+| 2.2398 | 2.75 | 88 | 2.3605 |
+| 2.2693 | 3.0 | 96 | 2.3526 |
+| 2.425 | 3.25 | 104 | 2.3471 |
+| 2.2857 | 3.5 | 112 | 2.3468 |
+| 2.2448 | 3.75 | 120 | 2.3451 |
+| 2.1836 | 4.0 | 128 | 2.3466 |
+
+
+### Framework versions
+
+- PEFT 0.10.0
+- Transformers 4.40.0.dev0
+- Pytorch 2.1.2+cu118
+- Datasets 2.15.0
+- Tokenizers 0.15.0
\ No newline at end of file
diff --git a/adapter_config.json b/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..28a5d05472d8cd3bfe780706c0eded9ae55cc02a
--- /dev/null
+++ b/adapter_config.json
@@ -0,0 +1,34 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
+ "bias": "none",
+ "fan_in_fan_out": null,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 32,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "up_proj",
+ "v_proj",
+ "o_proj",
+ "down_proj",
+ "q_proj",
+ "k_proj",
+ "gate_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/adapter_model.bin b/adapter_model.bin
new file mode 100644
index 0000000000000000000000000000000000000000..eb3ca9d56364a8e85cbccf2959527429e61776b7
--- /dev/null
+++ b/adapter_model.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:3863c09db8a8ee37f6c92e820fd3918dba26368631f34bba226156413c6f68bd
+size 50573978
diff --git a/checkpoint-128/README.md b/checkpoint-128/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..78aadd3258d05c05cabc678473b6f0942cd441c7
--- /dev/null
+++ b/checkpoint-128/README.md
@@ -0,0 +1,202 @@
+---
+library_name: peft
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.10.0
\ No newline at end of file
diff --git a/checkpoint-128/adapter_config.json b/checkpoint-128/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..28a5d05472d8cd3bfe780706c0eded9ae55cc02a
--- /dev/null
+++ b/checkpoint-128/adapter_config.json
@@ -0,0 +1,34 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
+ "bias": "none",
+ "fan_in_fan_out": null,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 32,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "up_proj",
+ "v_proj",
+ "o_proj",
+ "down_proj",
+ "q_proj",
+ "k_proj",
+ "gate_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-128/adapter_model.safetensors b/checkpoint-128/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..196e6d0818b109c6c700d7652c88c5280c1afb60
--- /dev/null
+++ b/checkpoint-128/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:528e075df49b7bbd123228a3b95349c09b3e4aff529a6761814d25530514b3ec
+size 50503848
diff --git a/checkpoint-128/optimizer.pt b/checkpoint-128/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..3eee76973544f0a79ba522d8f6aaa319b2075d4d
--- /dev/null
+++ b/checkpoint-128/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:f2662bde842785223ccfc20324eb816ad2974916f34f883ea7c17e0bfd3e93bd
+size 202035450
diff --git a/checkpoint-128/rng_state.pth b/checkpoint-128/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..adab1e7675bf55fc0dff7d84393e57e138d23f40
--- /dev/null
+++ b/checkpoint-128/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:e46048a3692b100b154cf7c38df5e2ce5c28c9c56bbbf5e17fe2196b8374656f
+size 14244
diff --git a/checkpoint-128/scheduler.pt b/checkpoint-128/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..ee5b484f425ff0c2699435e9c95a9e3b42b52b87
--- /dev/null
+++ b/checkpoint-128/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:968a6dcbda34982ad43f2af2e04e5edf94e043c521201e71b1583695497d18e0
+size 1064
diff --git a/checkpoint-128/special_tokens_map.json b/checkpoint-128/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..72ecfeeb7e14d244c936169d2ed139eeae235ef1
--- /dev/null
+++ b/checkpoint-128/special_tokens_map.json
@@ -0,0 +1,24 @@
+{
+ "bos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "unk_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/checkpoint-128/tokenizer.model b/checkpoint-128/tokenizer.model
new file mode 100644
index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899
--- /dev/null
+++ b/checkpoint-128/tokenizer.model
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
+size 499723
diff --git a/checkpoint-128/tokenizer_config.json b/checkpoint-128/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..0773857a13ba5a27453a0b462624fe76e8e82a86
--- /dev/null
+++ b/checkpoint-128/tokenizer_config.json
@@ -0,0 +1,44 @@
+{
+ "add_bos_token": true,
+ "add_eos_token": false,
+ "add_prefix_space": true,
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "legacy": false,
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "padding_side": "right",
+ "sp_model_kwargs": {},
+ "spaces_between_special_tokens": false,
+ "tokenizer_class": "LlamaTokenizer",
+ "unk_token": "",
+ "use_default_system_prompt": false,
+ "use_fast": true
+}
diff --git a/checkpoint-128/trainer_state.json b/checkpoint-128/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..4beddca44366228dc8e5a3adac5d1b24ffa119b7
--- /dev/null
+++ b/checkpoint-128/trainer_state.json
@@ -0,0 +1,1053 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 4.0,
+ "eval_steps": 8,
+ "global_step": 128,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "grad_norm": 0.5859375,
+ "learning_rate": 2e-05,
+ "loss": 3.9295,
+ "step": 1
+ },
+ {
+ "epoch": 0.03,
+ "eval_loss": 3.907318115234375,
+ "eval_runtime": 1.2713,
+ "eval_samples_per_second": 11.012,
+ "eval_steps_per_second": 5.506,
+ "step": 1
+ },
+ {
+ "epoch": 0.06,
+ "grad_norm": 0.5078125,
+ "learning_rate": 4e-05,
+ "loss": 3.805,
+ "step": 2
+ },
+ {
+ "epoch": 0.09,
+ "grad_norm": 0.63671875,
+ "learning_rate": 6e-05,
+ "loss": 3.8521,
+ "step": 3
+ },
+ {
+ "epoch": 0.12,
+ "grad_norm": 0.609375,
+ "learning_rate": 8e-05,
+ "loss": 3.8947,
+ "step": 4
+ },
+ {
+ "epoch": 0.16,
+ "grad_norm": 0.546875,
+ "learning_rate": 0.0001,
+ "loss": 3.6494,
+ "step": 5
+ },
+ {
+ "epoch": 0.19,
+ "grad_norm": 0.50390625,
+ "learning_rate": 0.00012,
+ "loss": 3.6457,
+ "step": 6
+ },
+ {
+ "epoch": 0.22,
+ "grad_norm": 0.640625,
+ "learning_rate": 0.00014,
+ "loss": 3.967,
+ "step": 7
+ },
+ {
+ "epoch": 0.25,
+ "grad_norm": 0.494140625,
+ "learning_rate": 0.00016,
+ "loss": 3.5364,
+ "step": 8
+ },
+ {
+ "epoch": 0.25,
+ "eval_loss": 3.6198840141296387,
+ "eval_runtime": 1.2681,
+ "eval_samples_per_second": 11.04,
+ "eval_steps_per_second": 5.52,
+ "step": 8
+ },
+ {
+ "epoch": 0.28,
+ "grad_norm": 0.51171875,
+ "learning_rate": 0.00018,
+ "loss": 3.5216,
+ "step": 9
+ },
+ {
+ "epoch": 0.31,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.0002,
+ "loss": 3.677,
+ "step": 10
+ },
+ {
+ "epoch": 0.34,
+ "grad_norm": 0.734375,
+ "learning_rate": 0.00019996456111234527,
+ "loss": 3.058,
+ "step": 11
+ },
+ {
+ "epoch": 0.38,
+ "grad_norm": 0.66796875,
+ "learning_rate": 0.0001998582695676762,
+ "loss": 3.1333,
+ "step": 12
+ },
+ {
+ "epoch": 0.41,
+ "grad_norm": 0.8359375,
+ "learning_rate": 0.000199681200703075,
+ "loss": 3.38,
+ "step": 13
+ },
+ {
+ "epoch": 0.44,
+ "grad_norm": 0.62109375,
+ "learning_rate": 0.00019943348002101371,
+ "loss": 3.1371,
+ "step": 14
+ },
+ {
+ "epoch": 0.47,
+ "grad_norm": 0.65625,
+ "learning_rate": 0.00019911528310040074,
+ "loss": 3.1479,
+ "step": 15
+ },
+ {
+ "epoch": 0.5,
+ "grad_norm": 0.6484375,
+ "learning_rate": 0.00019872683547213446,
+ "loss": 3.263,
+ "step": 16
+ },
+ {
+ "epoch": 0.5,
+ "eval_loss": 3.1820719242095947,
+ "eval_runtime": 1.2692,
+ "eval_samples_per_second": 11.031,
+ "eval_steps_per_second": 5.515,
+ "step": 16
+ },
+ {
+ "epoch": 0.53,
+ "grad_norm": 0.53125,
+ "learning_rate": 0.00019826841245925212,
+ "loss": 2.9833,
+ "step": 17
+ },
+ {
+ "epoch": 0.56,
+ "grad_norm": 0.5703125,
+ "learning_rate": 0.00019774033898178667,
+ "loss": 3.0787,
+ "step": 18
+ },
+ {
+ "epoch": 0.59,
+ "grad_norm": 0.71484375,
+ "learning_rate": 0.00019714298932647098,
+ "loss": 3.4132,
+ "step": 19
+ },
+ {
+ "epoch": 0.62,
+ "grad_norm": 0.73046875,
+ "learning_rate": 0.0001964767868814516,
+ "loss": 2.7304,
+ "step": 20
+ },
+ {
+ "epoch": 0.66,
+ "grad_norm": 0.55078125,
+ "learning_rate": 0.00019574220383620055,
+ "loss": 3.0116,
+ "step": 21
+ },
+ {
+ "epoch": 0.69,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.00019493976084683813,
+ "loss": 2.9474,
+ "step": 22
+ },
+ {
+ "epoch": 0.72,
+ "grad_norm": 0.61328125,
+ "learning_rate": 0.00019407002666710336,
+ "loss": 2.9415,
+ "step": 23
+ },
+ {
+ "epoch": 0.75,
+ "grad_norm": 0.6328125,
+ "learning_rate": 0.00019313361774523385,
+ "loss": 2.798,
+ "step": 24
+ },
+ {
+ "epoch": 0.75,
+ "eval_loss": 2.896176815032959,
+ "eval_runtime": 1.2765,
+ "eval_samples_per_second": 10.967,
+ "eval_steps_per_second": 5.484,
+ "step": 24
+ },
+ {
+ "epoch": 0.78,
+ "grad_norm": 0.8046875,
+ "learning_rate": 0.00019213119778704128,
+ "loss": 3.2157,
+ "step": 25
+ },
+ {
+ "epoch": 0.81,
+ "grad_norm": 0.7109375,
+ "learning_rate": 0.00019106347728549135,
+ "loss": 3.0666,
+ "step": 26
+ },
+ {
+ "epoch": 0.84,
+ "grad_norm": 0.59765625,
+ "learning_rate": 0.00018993121301712193,
+ "loss": 2.8219,
+ "step": 27
+ },
+ {
+ "epoch": 0.88,
+ "grad_norm": 0.75390625,
+ "learning_rate": 0.00018873520750565718,
+ "loss": 3.1164,
+ "step": 28
+ },
+ {
+ "epoch": 0.91,
+ "grad_norm": 0.67578125,
+ "learning_rate": 0.00018747630845319612,
+ "loss": 2.7154,
+ "step": 29
+ },
+ {
+ "epoch": 0.94,
+ "grad_norm": 0.625,
+ "learning_rate": 0.0001861554081393806,
+ "loss": 2.7395,
+ "step": 30
+ },
+ {
+ "epoch": 0.97,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.0001847734427889671,
+ "loss": 2.8282,
+ "step": 31
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.86328125,
+ "learning_rate": 0.0001833313919082515,
+ "loss": 2.7787,
+ "step": 32
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 2.67726731300354,
+ "eval_runtime": 1.2769,
+ "eval_samples_per_second": 10.964,
+ "eval_steps_per_second": 5.482,
+ "step": 32
+ },
+ {
+ "epoch": 1.03,
+ "grad_norm": 0.703125,
+ "learning_rate": 0.0001818302775908169,
+ "loss": 2.5957,
+ "step": 33
+ },
+ {
+ "epoch": 1.06,
+ "grad_norm": 0.76953125,
+ "learning_rate": 0.00018027116379309638,
+ "loss": 2.7011,
+ "step": 34
+ },
+ {
+ "epoch": 1.09,
+ "grad_norm": 0.84765625,
+ "learning_rate": 0.00017865515558026428,
+ "loss": 2.6043,
+ "step": 35
+ },
+ {
+ "epoch": 1.12,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00017698339834299061,
+ "loss": 2.8607,
+ "step": 36
+ },
+ {
+ "epoch": 1.16,
+ "grad_norm": 0.7578125,
+ "learning_rate": 0.00017525707698561385,
+ "loss": 2.5949,
+ "step": 37
+ },
+ {
+ "epoch": 1.19,
+ "grad_norm": 0.94921875,
+ "learning_rate": 0.00017347741508630672,
+ "loss": 2.7476,
+ "step": 38
+ },
+ {
+ "epoch": 1.22,
+ "grad_norm": 0.88671875,
+ "learning_rate": 0.00017164567402983152,
+ "loss": 2.7991,
+ "step": 39
+ },
+ {
+ "epoch": 1.25,
+ "grad_norm": 0.8671875,
+ "learning_rate": 0.0001697631521134985,
+ "loss": 2.5959,
+ "step": 40
+ },
+ {
+ "epoch": 1.25,
+ "eval_loss": 2.5505764484405518,
+ "eval_runtime": 1.2761,
+ "eval_samples_per_second": 10.971,
+ "eval_steps_per_second": 5.486,
+ "step": 40
+ },
+ {
+ "epoch": 1.28,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00016783118362696163,
+ "loss": 2.5342,
+ "step": 41
+ },
+ {
+ "epoch": 1.31,
+ "grad_norm": 0.921875,
+ "learning_rate": 0.00016585113790650388,
+ "loss": 2.4969,
+ "step": 42
+ },
+ {
+ "epoch": 1.34,
+ "grad_norm": 1.125,
+ "learning_rate": 0.00016382441836448202,
+ "loss": 2.5723,
+ "step": 43
+ },
+ {
+ "epoch": 1.38,
+ "grad_norm": 1.015625,
+ "learning_rate": 0.0001617524614946192,
+ "loss": 2.6166,
+ "step": 44
+ },
+ {
+ "epoch": 1.41,
+ "grad_norm": 0.98828125,
+ "learning_rate": 0.00015963673585385016,
+ "loss": 2.4574,
+ "step": 45
+ },
+ {
+ "epoch": 1.44,
+ "grad_norm": 0.94921875,
+ "learning_rate": 0.0001574787410214407,
+ "loss": 2.5074,
+ "step": 46
+ },
+ {
+ "epoch": 1.47,
+ "grad_norm": 0.953125,
+ "learning_rate": 0.00015528000653611935,
+ "loss": 2.5394,
+ "step": 47
+ },
+ {
+ "epoch": 1.5,
+ "grad_norm": 0.828125,
+ "learning_rate": 0.00015304209081197425,
+ "loss": 2.4793,
+ "step": 48
+ },
+ {
+ "epoch": 1.5,
+ "eval_loss": 2.495466470718384,
+ "eval_runtime": 1.2814,
+ "eval_samples_per_second": 10.926,
+ "eval_steps_per_second": 5.463,
+ "step": 48
+ },
+ {
+ "epoch": 1.53,
+ "grad_norm": 0.921875,
+ "learning_rate": 0.000150766580033884,
+ "loss": 2.2909,
+ "step": 49
+ },
+ {
+ "epoch": 1.56,
+ "grad_norm": 0.74609375,
+ "learning_rate": 0.00014845508703326504,
+ "loss": 2.3424,
+ "step": 50
+ },
+ {
+ "epoch": 1.59,
+ "grad_norm": 0.87890625,
+ "learning_rate": 0.0001461092501449326,
+ "loss": 2.506,
+ "step": 51
+ },
+ {
+ "epoch": 1.62,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00014373073204588556,
+ "loss": 2.4829,
+ "step": 52
+ },
+ {
+ "epoch": 1.66,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00014132121857683783,
+ "loss": 2.481,
+ "step": 53
+ },
+ {
+ "epoch": 1.69,
+ "grad_norm": 0.9296875,
+ "learning_rate": 0.00013888241754733208,
+ "loss": 2.5512,
+ "step": 54
+ },
+ {
+ "epoch": 1.72,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00013641605752528224,
+ "loss": 2.5405,
+ "step": 55
+ },
+ {
+ "epoch": 1.75,
+ "grad_norm": 0.95703125,
+ "learning_rate": 0.00013392388661180303,
+ "loss": 2.5221,
+ "step": 56
+ },
+ {
+ "epoch": 1.75,
+ "eval_loss": 2.461298704147339,
+ "eval_runtime": 1.28,
+ "eval_samples_per_second": 10.937,
+ "eval_steps_per_second": 5.469,
+ "step": 56
+ },
+ {
+ "epoch": 1.78,
+ "grad_norm": 0.98046875,
+ "learning_rate": 0.0001314076712021949,
+ "loss": 2.5646,
+ "step": 57
+ },
+ {
+ "epoch": 1.81,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.0001288691947339621,
+ "loss": 2.5079,
+ "step": 58
+ },
+ {
+ "epoch": 1.84,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.00012631025642275212,
+ "loss": 2.4743,
+ "step": 59
+ },
+ {
+ "epoch": 1.88,
+ "grad_norm": 0.60546875,
+ "learning_rate": 0.0001237326699871115,
+ "loss": 2.3103,
+ "step": 60
+ },
+ {
+ "epoch": 1.91,
+ "grad_norm": 0.80859375,
+ "learning_rate": 0.00012113826236296244,
+ "loss": 2.4229,
+ "step": 61
+ },
+ {
+ "epoch": 1.94,
+ "grad_norm": 0.671875,
+ "learning_rate": 0.00011852887240871145,
+ "loss": 2.2709,
+ "step": 62
+ },
+ {
+ "epoch": 1.97,
+ "grad_norm": 0.8203125,
+ "learning_rate": 0.00011590634960190721,
+ "loss": 2.4868,
+ "step": 63
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.62109375,
+ "learning_rate": 0.00011327255272837221,
+ "loss": 2.4384,
+ "step": 64
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 2.4055111408233643,
+ "eval_runtime": 1.2798,
+ "eval_samples_per_second": 10.94,
+ "eval_steps_per_second": 5.47,
+ "step": 64
+ },
+ {
+ "epoch": 2.03,
+ "grad_norm": 0.6328125,
+ "learning_rate": 0.00011062934856473655,
+ "loss": 2.3671,
+ "step": 65
+ },
+ {
+ "epoch": 2.06,
+ "grad_norm": 0.69140625,
+ "learning_rate": 0.00010797861055530831,
+ "loss": 2.3381,
+ "step": 66
+ },
+ {
+ "epoch": 2.09,
+ "grad_norm": 0.671875,
+ "learning_rate": 0.00010532221748421787,
+ "loss": 2.2134,
+ "step": 67
+ },
+ {
+ "epoch": 2.12,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00010266205214377748,
+ "loss": 2.2687,
+ "step": 68
+ },
+ {
+ "epoch": 2.16,
+ "grad_norm": 0.75,
+ "learning_rate": 0.0001,
+ "loss": 2.4273,
+ "step": 69
+ },
+ {
+ "epoch": 2.19,
+ "grad_norm": 0.8671875,
+ "learning_rate": 9.733794785622253e-05,
+ "loss": 2.4439,
+ "step": 70
+ },
+ {
+ "epoch": 2.22,
+ "grad_norm": 0.953125,
+ "learning_rate": 9.467778251578217e-05,
+ "loss": 2.6631,
+ "step": 71
+ },
+ {
+ "epoch": 2.25,
+ "grad_norm": 0.78125,
+ "learning_rate": 9.202138944469168e-05,
+ "loss": 2.295,
+ "step": 72
+ },
+ {
+ "epoch": 2.25,
+ "eval_loss": 2.3922784328460693,
+ "eval_runtime": 1.282,
+ "eval_samples_per_second": 10.921,
+ "eval_steps_per_second": 5.46,
+ "step": 72
+ },
+ {
+ "epoch": 2.28,
+ "grad_norm": 0.734375,
+ "learning_rate": 8.937065143526347e-05,
+ "loss": 2.4963,
+ "step": 73
+ },
+ {
+ "epoch": 2.31,
+ "grad_norm": 0.68359375,
+ "learning_rate": 8.672744727162781e-05,
+ "loss": 2.4274,
+ "step": 74
+ },
+ {
+ "epoch": 2.34,
+ "grad_norm": 0.9765625,
+ "learning_rate": 8.409365039809281e-05,
+ "loss": 2.4988,
+ "step": 75
+ },
+ {
+ "epoch": 2.38,
+ "grad_norm": 0.75390625,
+ "learning_rate": 8.147112759128859e-05,
+ "loss": 2.2886,
+ "step": 76
+ },
+ {
+ "epoch": 2.41,
+ "grad_norm": 0.80078125,
+ "learning_rate": 7.886173763703757e-05,
+ "loss": 2.1944,
+ "step": 77
+ },
+ {
+ "epoch": 2.44,
+ "grad_norm": 0.8828125,
+ "learning_rate": 7.626733001288851e-05,
+ "loss": 2.3283,
+ "step": 78
+ },
+ {
+ "epoch": 2.47,
+ "grad_norm": 0.82421875,
+ "learning_rate": 7.368974357724789e-05,
+ "loss": 2.3855,
+ "step": 79
+ },
+ {
+ "epoch": 2.5,
+ "grad_norm": 0.7890625,
+ "learning_rate": 7.113080526603792e-05,
+ "loss": 2.3943,
+ "step": 80
+ },
+ {
+ "epoch": 2.5,
+ "eval_loss": 2.386228084564209,
+ "eval_runtime": 1.2835,
+ "eval_samples_per_second": 10.908,
+ "eval_steps_per_second": 5.454,
+ "step": 80
+ },
+ {
+ "epoch": 2.53,
+ "grad_norm": 0.94921875,
+ "learning_rate": 6.859232879780515e-05,
+ "loss": 2.2725,
+ "step": 81
+ },
+ {
+ "epoch": 2.56,
+ "grad_norm": 0.75,
+ "learning_rate": 6.607611338819697e-05,
+ "loss": 2.1989,
+ "step": 82
+ },
+ {
+ "epoch": 2.59,
+ "grad_norm": 0.703125,
+ "learning_rate": 6.358394247471778e-05,
+ "loss": 2.1219,
+ "step": 83
+ },
+ {
+ "epoch": 2.62,
+ "grad_norm": 1.0390625,
+ "learning_rate": 6.111758245266794e-05,
+ "loss": 2.3478,
+ "step": 84
+ },
+ {
+ "epoch": 2.66,
+ "grad_norm": 0.78125,
+ "learning_rate": 5.867878142316221e-05,
+ "loss": 2.4196,
+ "step": 85
+ },
+ {
+ "epoch": 2.69,
+ "grad_norm": 0.6796875,
+ "learning_rate": 5.626926795411447e-05,
+ "loss": 2.1882,
+ "step": 86
+ },
+ {
+ "epoch": 2.72,
+ "grad_norm": 0.8203125,
+ "learning_rate": 5.38907498550674e-05,
+ "loss": 2.4703,
+ "step": 87
+ },
+ {
+ "epoch": 2.75,
+ "grad_norm": 0.8671875,
+ "learning_rate": 5.1544912966734994e-05,
+ "loss": 2.2398,
+ "step": 88
+ },
+ {
+ "epoch": 2.75,
+ "eval_loss": 2.3605105876922607,
+ "eval_runtime": 1.2887,
+ "eval_samples_per_second": 10.864,
+ "eval_steps_per_second": 5.432,
+ "step": 88
+ },
+ {
+ "epoch": 2.78,
+ "grad_norm": 0.81640625,
+ "learning_rate": 4.9233419966116036e-05,
+ "loss": 2.2939,
+ "step": 89
+ },
+ {
+ "epoch": 2.81,
+ "grad_norm": 1.0234375,
+ "learning_rate": 4.695790918802576e-05,
+ "loss": 2.2835,
+ "step": 90
+ },
+ {
+ "epoch": 2.84,
+ "grad_norm": 0.85546875,
+ "learning_rate": 4.47199934638807e-05,
+ "loss": 2.3145,
+ "step": 91
+ },
+ {
+ "epoch": 2.88,
+ "grad_norm": 0.5546875,
+ "learning_rate": 4.252125897855932e-05,
+ "loss": 2.2521,
+ "step": 92
+ },
+ {
+ "epoch": 2.91,
+ "grad_norm": 0.76953125,
+ "learning_rate": 4.036326414614985e-05,
+ "loss": 2.377,
+ "step": 93
+ },
+ {
+ "epoch": 2.94,
+ "grad_norm": 0.78125,
+ "learning_rate": 3.824753850538082e-05,
+ "loss": 2.4343,
+ "step": 94
+ },
+ {
+ "epoch": 2.97,
+ "grad_norm": 0.8046875,
+ "learning_rate": 3.617558163551802e-05,
+ "loss": 2.2051,
+ "step": 95
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.88671875,
+ "learning_rate": 3.414886209349615e-05,
+ "loss": 2.2693,
+ "step": 96
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 2.3525888919830322,
+ "eval_runtime": 1.2821,
+ "eval_samples_per_second": 10.919,
+ "eval_steps_per_second": 5.46,
+ "step": 96
+ },
+ {
+ "epoch": 3.03,
+ "grad_norm": 0.74609375,
+ "learning_rate": 3.216881637303839e-05,
+ "loss": 2.5048,
+ "step": 97
+ },
+ {
+ "epoch": 3.06,
+ "grad_norm": 0.95703125,
+ "learning_rate": 3.0236847886501542e-05,
+ "loss": 2.3404,
+ "step": 98
+ },
+ {
+ "epoch": 3.09,
+ "grad_norm": 0.69140625,
+ "learning_rate": 2.8354325970168484e-05,
+ "loss": 2.3147,
+ "step": 99
+ },
+ {
+ "epoch": 3.12,
+ "grad_norm": 0.765625,
+ "learning_rate": 2.6522584913693294e-05,
+ "loss": 2.448,
+ "step": 100
+ },
+ {
+ "epoch": 3.16,
+ "grad_norm": 0.671875,
+ "learning_rate": 2.4742923014386156e-05,
+ "loss": 2.2212,
+ "step": 101
+ },
+ {
+ "epoch": 3.19,
+ "grad_norm": 0.625,
+ "learning_rate": 2.301660165700936e-05,
+ "loss": 2.3427,
+ "step": 102
+ },
+ {
+ "epoch": 3.22,
+ "grad_norm": 0.6484375,
+ "learning_rate": 2.1344844419735755e-05,
+ "loss": 2.1,
+ "step": 103
+ },
+ {
+ "epoch": 3.25,
+ "grad_norm": 0.73046875,
+ "learning_rate": 1.9728836206903656e-05,
+ "loss": 2.425,
+ "step": 104
+ },
+ {
+ "epoch": 3.25,
+ "eval_loss": 2.3470816612243652,
+ "eval_runtime": 1.2874,
+ "eval_samples_per_second": 10.875,
+ "eval_steps_per_second": 5.437,
+ "step": 104
+ },
+ {
+ "epoch": 3.28,
+ "grad_norm": 0.734375,
+ "learning_rate": 1.8169722409183097e-05,
+ "loss": 2.2268,
+ "step": 105
+ },
+ {
+ "epoch": 3.31,
+ "grad_norm": 0.8359375,
+ "learning_rate": 1.6668608091748495e-05,
+ "loss": 2.3695,
+ "step": 106
+ },
+ {
+ "epoch": 3.34,
+ "grad_norm": 0.58203125,
+ "learning_rate": 1.522655721103291e-05,
+ "loss": 2.1864,
+ "step": 107
+ },
+ {
+ "epoch": 3.38,
+ "grad_norm": 0.5703125,
+ "learning_rate": 1.3844591860619383e-05,
+ "loss": 2.2613,
+ "step": 108
+ },
+ {
+ "epoch": 3.41,
+ "grad_norm": 0.7265625,
+ "learning_rate": 1.2523691546803873e-05,
+ "loss": 2.2771,
+ "step": 109
+ },
+ {
+ "epoch": 3.44,
+ "grad_norm": 0.59765625,
+ "learning_rate": 1.1264792494342857e-05,
+ "loss": 2.2056,
+ "step": 110
+ },
+ {
+ "epoch": 3.47,
+ "grad_norm": 0.6875,
+ "learning_rate": 1.0068786982878087e-05,
+ "loss": 2.4039,
+ "step": 111
+ },
+ {
+ "epoch": 3.5,
+ "grad_norm": 0.77734375,
+ "learning_rate": 8.936522714508678e-06,
+ "loss": 2.2857,
+ "step": 112
+ },
+ {
+ "epoch": 3.5,
+ "eval_loss": 2.3467748165130615,
+ "eval_runtime": 1.2808,
+ "eval_samples_per_second": 10.931,
+ "eval_steps_per_second": 5.466,
+ "step": 112
+ },
+ {
+ "epoch": 3.53,
+ "grad_norm": 0.671875,
+ "learning_rate": 7.868802212958703e-06,
+ "loss": 2.257,
+ "step": 113
+ },
+ {
+ "epoch": 3.56,
+ "grad_norm": 1.21875,
+ "learning_rate": 6.866382254766157e-06,
+ "loss": 2.4278,
+ "step": 114
+ },
+ {
+ "epoch": 3.59,
+ "grad_norm": 0.859375,
+ "learning_rate": 5.929973332896677e-06,
+ "loss": 2.2232,
+ "step": 115
+ },
+ {
+ "epoch": 3.62,
+ "grad_norm": 0.66015625,
+ "learning_rate": 5.060239153161872e-06,
+ "loss": 2.225,
+ "step": 116
+ },
+ {
+ "epoch": 3.66,
+ "grad_norm": 0.6875,
+ "learning_rate": 4.257796163799455e-06,
+ "loss": 2.2133,
+ "step": 117
+ },
+ {
+ "epoch": 3.69,
+ "grad_norm": 0.8671875,
+ "learning_rate": 3.5232131185484076e-06,
+ "loss": 2.2191,
+ "step": 118
+ },
+ {
+ "epoch": 3.72,
+ "grad_norm": 0.77734375,
+ "learning_rate": 2.857010673529015e-06,
+ "loss": 2.0715,
+ "step": 119
+ },
+ {
+ "epoch": 3.75,
+ "grad_norm": 0.72265625,
+ "learning_rate": 2.259661018213333e-06,
+ "loss": 2.2448,
+ "step": 120
+ },
+ {
+ "epoch": 3.75,
+ "eval_loss": 2.3450927734375,
+ "eval_runtime": 1.2817,
+ "eval_samples_per_second": 10.923,
+ "eval_steps_per_second": 5.461,
+ "step": 120
+ },
+ {
+ "epoch": 3.78,
+ "grad_norm": 0.62109375,
+ "learning_rate": 1.7315875407479032e-06,
+ "loss": 2.1337,
+ "step": 121
+ },
+ {
+ "epoch": 3.81,
+ "grad_norm": 0.765625,
+ "learning_rate": 1.2731645278655445e-06,
+ "loss": 2.4334,
+ "step": 122
+ },
+ {
+ "epoch": 3.84,
+ "grad_norm": 0.73828125,
+ "learning_rate": 8.847168995992916e-07,
+ "loss": 2.3989,
+ "step": 123
+ },
+ {
+ "epoch": 3.88,
+ "grad_norm": 0.71484375,
+ "learning_rate": 5.665199789862907e-07,
+ "loss": 2.3509,
+ "step": 124
+ },
+ {
+ "epoch": 3.91,
+ "grad_norm": 0.78125,
+ "learning_rate": 3.1879929692498757e-07,
+ "loss": 2.3366,
+ "step": 125
+ },
+ {
+ "epoch": 3.94,
+ "grad_norm": 0.625,
+ "learning_rate": 1.4173043232380557e-07,
+ "loss": 2.071,
+ "step": 126
+ },
+ {
+ "epoch": 3.97,
+ "grad_norm": 0.6015625,
+ "learning_rate": 3.5438887654737355e-08,
+ "loss": 2.2108,
+ "step": 127
+ },
+ {
+ "epoch": 4.0,
+ "grad_norm": 0.7578125,
+ "learning_rate": 0.0,
+ "loss": 2.1836,
+ "step": 128
+ },
+ {
+ "epoch": 4.0,
+ "eval_loss": 2.3465805053710938,
+ "eval_runtime": 1.2864,
+ "eval_samples_per_second": 10.883,
+ "eval_steps_per_second": 5.441,
+ "step": 128
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 128,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 4,
+ "save_steps": 32,
+ "total_flos": 2.6669324546605056e+16,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-128/training_args.bin b/checkpoint-128/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..18814c2c9b675fc3be2d153f474e77fb88d5038e
--- /dev/null
+++ b/checkpoint-128/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5fca5508434f64a69e54fc8c46f68907814d91f48e751b7a0eeb4050e5ae3225
+size 5816
diff --git a/checkpoint-32/README.md b/checkpoint-32/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..78aadd3258d05c05cabc678473b6f0942cd441c7
--- /dev/null
+++ b/checkpoint-32/README.md
@@ -0,0 +1,202 @@
+---
+library_name: peft
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.10.0
\ No newline at end of file
diff --git a/checkpoint-32/adapter_config.json b/checkpoint-32/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..28a5d05472d8cd3bfe780706c0eded9ae55cc02a
--- /dev/null
+++ b/checkpoint-32/adapter_config.json
@@ -0,0 +1,34 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
+ "bias": "none",
+ "fan_in_fan_out": null,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 32,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "up_proj",
+ "v_proj",
+ "o_proj",
+ "down_proj",
+ "q_proj",
+ "k_proj",
+ "gate_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-32/adapter_model.safetensors b/checkpoint-32/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..5d77977993e6a3ec786583aa43499c136292bc1b
--- /dev/null
+++ b/checkpoint-32/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:ec430d847c06078c87671f113c67363bf58b51f662fe0c2c9aac1aea9197c05e
+size 50503848
diff --git a/checkpoint-32/optimizer.pt b/checkpoint-32/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c34b5013e20e2b0762dfdde502bfa17350bb704f
--- /dev/null
+++ b/checkpoint-32/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:b2dcfc1311f5586ce9bc6ab8405e3706e4dcb6ef0050771e2c9bbe97b1fc7874
+size 202035450
diff --git a/checkpoint-32/rng_state.pth b/checkpoint-32/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..58a6ba31ec07fa4f9a8b982f55b673ab48aafb20
--- /dev/null
+++ b/checkpoint-32/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fba88830bbf3f3a231c5f7c643ca357bcc703c4a14675504974ab2e003369a61
+size 14244
diff --git a/checkpoint-32/scheduler.pt b/checkpoint-32/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..e17718c2286bf3678a89033727e8dc0d7f6b14ca
--- /dev/null
+++ b/checkpoint-32/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:73fa7fed0c0930926a3524b49ae450b78da12282a81cf57d30f2d1a4044247d8
+size 1064
diff --git a/checkpoint-32/special_tokens_map.json b/checkpoint-32/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..72ecfeeb7e14d244c936169d2ed139eeae235ef1
--- /dev/null
+++ b/checkpoint-32/special_tokens_map.json
@@ -0,0 +1,24 @@
+{
+ "bos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "unk_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/checkpoint-32/tokenizer.model b/checkpoint-32/tokenizer.model
new file mode 100644
index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899
--- /dev/null
+++ b/checkpoint-32/tokenizer.model
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
+size 499723
diff --git a/checkpoint-32/tokenizer_config.json b/checkpoint-32/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..0773857a13ba5a27453a0b462624fe76e8e82a86
--- /dev/null
+++ b/checkpoint-32/tokenizer_config.json
@@ -0,0 +1,44 @@
+{
+ "add_bos_token": true,
+ "add_eos_token": false,
+ "add_prefix_space": true,
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "legacy": false,
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "padding_side": "right",
+ "sp_model_kwargs": {},
+ "spaces_between_special_tokens": false,
+ "tokenizer_class": "LlamaTokenizer",
+ "unk_token": "",
+ "use_default_system_prompt": false,
+ "use_fast": true
+}
diff --git a/checkpoint-32/trainer_state.json b/checkpoint-32/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..6d4c385a701d5d69e028762c30a114f372d0f102
--- /dev/null
+++ b/checkpoint-32/trainer_state.json
@@ -0,0 +1,285 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 1.0,
+ "eval_steps": 8,
+ "global_step": 32,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "grad_norm": 0.5859375,
+ "learning_rate": 2e-05,
+ "loss": 3.9295,
+ "step": 1
+ },
+ {
+ "epoch": 0.03,
+ "eval_loss": 3.907318115234375,
+ "eval_runtime": 1.2713,
+ "eval_samples_per_second": 11.012,
+ "eval_steps_per_second": 5.506,
+ "step": 1
+ },
+ {
+ "epoch": 0.06,
+ "grad_norm": 0.5078125,
+ "learning_rate": 4e-05,
+ "loss": 3.805,
+ "step": 2
+ },
+ {
+ "epoch": 0.09,
+ "grad_norm": 0.63671875,
+ "learning_rate": 6e-05,
+ "loss": 3.8521,
+ "step": 3
+ },
+ {
+ "epoch": 0.12,
+ "grad_norm": 0.609375,
+ "learning_rate": 8e-05,
+ "loss": 3.8947,
+ "step": 4
+ },
+ {
+ "epoch": 0.16,
+ "grad_norm": 0.546875,
+ "learning_rate": 0.0001,
+ "loss": 3.6494,
+ "step": 5
+ },
+ {
+ "epoch": 0.19,
+ "grad_norm": 0.50390625,
+ "learning_rate": 0.00012,
+ "loss": 3.6457,
+ "step": 6
+ },
+ {
+ "epoch": 0.22,
+ "grad_norm": 0.640625,
+ "learning_rate": 0.00014,
+ "loss": 3.967,
+ "step": 7
+ },
+ {
+ "epoch": 0.25,
+ "grad_norm": 0.494140625,
+ "learning_rate": 0.00016,
+ "loss": 3.5364,
+ "step": 8
+ },
+ {
+ "epoch": 0.25,
+ "eval_loss": 3.6198840141296387,
+ "eval_runtime": 1.2681,
+ "eval_samples_per_second": 11.04,
+ "eval_steps_per_second": 5.52,
+ "step": 8
+ },
+ {
+ "epoch": 0.28,
+ "grad_norm": 0.51171875,
+ "learning_rate": 0.00018,
+ "loss": 3.5216,
+ "step": 9
+ },
+ {
+ "epoch": 0.31,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.0002,
+ "loss": 3.677,
+ "step": 10
+ },
+ {
+ "epoch": 0.34,
+ "grad_norm": 0.734375,
+ "learning_rate": 0.00019996456111234527,
+ "loss": 3.058,
+ "step": 11
+ },
+ {
+ "epoch": 0.38,
+ "grad_norm": 0.66796875,
+ "learning_rate": 0.0001998582695676762,
+ "loss": 3.1333,
+ "step": 12
+ },
+ {
+ "epoch": 0.41,
+ "grad_norm": 0.8359375,
+ "learning_rate": 0.000199681200703075,
+ "loss": 3.38,
+ "step": 13
+ },
+ {
+ "epoch": 0.44,
+ "grad_norm": 0.62109375,
+ "learning_rate": 0.00019943348002101371,
+ "loss": 3.1371,
+ "step": 14
+ },
+ {
+ "epoch": 0.47,
+ "grad_norm": 0.65625,
+ "learning_rate": 0.00019911528310040074,
+ "loss": 3.1479,
+ "step": 15
+ },
+ {
+ "epoch": 0.5,
+ "grad_norm": 0.6484375,
+ "learning_rate": 0.00019872683547213446,
+ "loss": 3.263,
+ "step": 16
+ },
+ {
+ "epoch": 0.5,
+ "eval_loss": 3.1820719242095947,
+ "eval_runtime": 1.2692,
+ "eval_samples_per_second": 11.031,
+ "eval_steps_per_second": 5.515,
+ "step": 16
+ },
+ {
+ "epoch": 0.53,
+ "grad_norm": 0.53125,
+ "learning_rate": 0.00019826841245925212,
+ "loss": 2.9833,
+ "step": 17
+ },
+ {
+ "epoch": 0.56,
+ "grad_norm": 0.5703125,
+ "learning_rate": 0.00019774033898178667,
+ "loss": 3.0787,
+ "step": 18
+ },
+ {
+ "epoch": 0.59,
+ "grad_norm": 0.71484375,
+ "learning_rate": 0.00019714298932647098,
+ "loss": 3.4132,
+ "step": 19
+ },
+ {
+ "epoch": 0.62,
+ "grad_norm": 0.73046875,
+ "learning_rate": 0.0001964767868814516,
+ "loss": 2.7304,
+ "step": 20
+ },
+ {
+ "epoch": 0.66,
+ "grad_norm": 0.55078125,
+ "learning_rate": 0.00019574220383620055,
+ "loss": 3.0116,
+ "step": 21
+ },
+ {
+ "epoch": 0.69,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.00019493976084683813,
+ "loss": 2.9474,
+ "step": 22
+ },
+ {
+ "epoch": 0.72,
+ "grad_norm": 0.61328125,
+ "learning_rate": 0.00019407002666710336,
+ "loss": 2.9415,
+ "step": 23
+ },
+ {
+ "epoch": 0.75,
+ "grad_norm": 0.6328125,
+ "learning_rate": 0.00019313361774523385,
+ "loss": 2.798,
+ "step": 24
+ },
+ {
+ "epoch": 0.75,
+ "eval_loss": 2.896176815032959,
+ "eval_runtime": 1.2765,
+ "eval_samples_per_second": 10.967,
+ "eval_steps_per_second": 5.484,
+ "step": 24
+ },
+ {
+ "epoch": 0.78,
+ "grad_norm": 0.8046875,
+ "learning_rate": 0.00019213119778704128,
+ "loss": 3.2157,
+ "step": 25
+ },
+ {
+ "epoch": 0.81,
+ "grad_norm": 0.7109375,
+ "learning_rate": 0.00019106347728549135,
+ "loss": 3.0666,
+ "step": 26
+ },
+ {
+ "epoch": 0.84,
+ "grad_norm": 0.59765625,
+ "learning_rate": 0.00018993121301712193,
+ "loss": 2.8219,
+ "step": 27
+ },
+ {
+ "epoch": 0.88,
+ "grad_norm": 0.75390625,
+ "learning_rate": 0.00018873520750565718,
+ "loss": 3.1164,
+ "step": 28
+ },
+ {
+ "epoch": 0.91,
+ "grad_norm": 0.67578125,
+ "learning_rate": 0.00018747630845319612,
+ "loss": 2.7154,
+ "step": 29
+ },
+ {
+ "epoch": 0.94,
+ "grad_norm": 0.625,
+ "learning_rate": 0.0001861554081393806,
+ "loss": 2.7395,
+ "step": 30
+ },
+ {
+ "epoch": 0.97,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.0001847734427889671,
+ "loss": 2.8282,
+ "step": 31
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.86328125,
+ "learning_rate": 0.0001833313919082515,
+ "loss": 2.7787,
+ "step": 32
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 2.67726731300354,
+ "eval_runtime": 1.2769,
+ "eval_samples_per_second": 10.964,
+ "eval_steps_per_second": 5.482,
+ "step": 32
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 128,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 4,
+ "save_steps": 32,
+ "total_flos": 6667331136651264.0,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-32/training_args.bin b/checkpoint-32/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..18814c2c9b675fc3be2d153f474e77fb88d5038e
--- /dev/null
+++ b/checkpoint-32/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5fca5508434f64a69e54fc8c46f68907814d91f48e751b7a0eeb4050e5ae3225
+size 5816
diff --git a/checkpoint-64/README.md b/checkpoint-64/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..78aadd3258d05c05cabc678473b6f0942cd441c7
--- /dev/null
+++ b/checkpoint-64/README.md
@@ -0,0 +1,202 @@
+---
+library_name: peft
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.10.0
\ No newline at end of file
diff --git a/checkpoint-64/adapter_config.json b/checkpoint-64/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..28a5d05472d8cd3bfe780706c0eded9ae55cc02a
--- /dev/null
+++ b/checkpoint-64/adapter_config.json
@@ -0,0 +1,34 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
+ "bias": "none",
+ "fan_in_fan_out": null,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 32,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "up_proj",
+ "v_proj",
+ "o_proj",
+ "down_proj",
+ "q_proj",
+ "k_proj",
+ "gate_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-64/adapter_model.safetensors b/checkpoint-64/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..f64158055261b7dbcd940a09727dfa4714abe709
--- /dev/null
+++ b/checkpoint-64/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:478e962adcf5cf2208e94455da60f82ca1058e1b9684714f9c16065ef2896b84
+size 50503848
diff --git a/checkpoint-64/optimizer.pt b/checkpoint-64/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..c2be7aab5a469d5d346dac676d36939320e43484
--- /dev/null
+++ b/checkpoint-64/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:1f7fae12b68510716cf6c13d76435aadee62c6caf362f7880a7604a4378d4478
+size 202035450
diff --git a/checkpoint-64/rng_state.pth b/checkpoint-64/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..ca068a57650e3ee3e1329872d2058d2114c12a06
--- /dev/null
+++ b/checkpoint-64/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:926ec53b6c5e5969c7a9dedbcd5e4c4ff81ecd6859afd62f93fd8684f6ac758d
+size 14244
diff --git a/checkpoint-64/scheduler.pt b/checkpoint-64/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..383157bccc0c6456709a28e9b1a3b19603f9651e
--- /dev/null
+++ b/checkpoint-64/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:fc2f121208237af7a1af0cb966f4edc3bb705d4fe4813c302cf1425b7f9228d4
+size 1064
diff --git a/checkpoint-64/special_tokens_map.json b/checkpoint-64/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..72ecfeeb7e14d244c936169d2ed139eeae235ef1
--- /dev/null
+++ b/checkpoint-64/special_tokens_map.json
@@ -0,0 +1,24 @@
+{
+ "bos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "unk_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/checkpoint-64/tokenizer.model b/checkpoint-64/tokenizer.model
new file mode 100644
index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899
--- /dev/null
+++ b/checkpoint-64/tokenizer.model
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
+size 499723
diff --git a/checkpoint-64/tokenizer_config.json b/checkpoint-64/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..0773857a13ba5a27453a0b462624fe76e8e82a86
--- /dev/null
+++ b/checkpoint-64/tokenizer_config.json
@@ -0,0 +1,44 @@
+{
+ "add_bos_token": true,
+ "add_eos_token": false,
+ "add_prefix_space": true,
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "legacy": false,
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "padding_side": "right",
+ "sp_model_kwargs": {},
+ "spaces_between_special_tokens": false,
+ "tokenizer_class": "LlamaTokenizer",
+ "unk_token": "",
+ "use_default_system_prompt": false,
+ "use_fast": true
+}
diff --git a/checkpoint-64/trainer_state.json b/checkpoint-64/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..dc8634c0b9dcd481438ba361b8bbb770ad44d783
--- /dev/null
+++ b/checkpoint-64/trainer_state.json
@@ -0,0 +1,541 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 2.0,
+ "eval_steps": 8,
+ "global_step": 64,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "grad_norm": 0.5859375,
+ "learning_rate": 2e-05,
+ "loss": 3.9295,
+ "step": 1
+ },
+ {
+ "epoch": 0.03,
+ "eval_loss": 3.907318115234375,
+ "eval_runtime": 1.2713,
+ "eval_samples_per_second": 11.012,
+ "eval_steps_per_second": 5.506,
+ "step": 1
+ },
+ {
+ "epoch": 0.06,
+ "grad_norm": 0.5078125,
+ "learning_rate": 4e-05,
+ "loss": 3.805,
+ "step": 2
+ },
+ {
+ "epoch": 0.09,
+ "grad_norm": 0.63671875,
+ "learning_rate": 6e-05,
+ "loss": 3.8521,
+ "step": 3
+ },
+ {
+ "epoch": 0.12,
+ "grad_norm": 0.609375,
+ "learning_rate": 8e-05,
+ "loss": 3.8947,
+ "step": 4
+ },
+ {
+ "epoch": 0.16,
+ "grad_norm": 0.546875,
+ "learning_rate": 0.0001,
+ "loss": 3.6494,
+ "step": 5
+ },
+ {
+ "epoch": 0.19,
+ "grad_norm": 0.50390625,
+ "learning_rate": 0.00012,
+ "loss": 3.6457,
+ "step": 6
+ },
+ {
+ "epoch": 0.22,
+ "grad_norm": 0.640625,
+ "learning_rate": 0.00014,
+ "loss": 3.967,
+ "step": 7
+ },
+ {
+ "epoch": 0.25,
+ "grad_norm": 0.494140625,
+ "learning_rate": 0.00016,
+ "loss": 3.5364,
+ "step": 8
+ },
+ {
+ "epoch": 0.25,
+ "eval_loss": 3.6198840141296387,
+ "eval_runtime": 1.2681,
+ "eval_samples_per_second": 11.04,
+ "eval_steps_per_second": 5.52,
+ "step": 8
+ },
+ {
+ "epoch": 0.28,
+ "grad_norm": 0.51171875,
+ "learning_rate": 0.00018,
+ "loss": 3.5216,
+ "step": 9
+ },
+ {
+ "epoch": 0.31,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.0002,
+ "loss": 3.677,
+ "step": 10
+ },
+ {
+ "epoch": 0.34,
+ "grad_norm": 0.734375,
+ "learning_rate": 0.00019996456111234527,
+ "loss": 3.058,
+ "step": 11
+ },
+ {
+ "epoch": 0.38,
+ "grad_norm": 0.66796875,
+ "learning_rate": 0.0001998582695676762,
+ "loss": 3.1333,
+ "step": 12
+ },
+ {
+ "epoch": 0.41,
+ "grad_norm": 0.8359375,
+ "learning_rate": 0.000199681200703075,
+ "loss": 3.38,
+ "step": 13
+ },
+ {
+ "epoch": 0.44,
+ "grad_norm": 0.62109375,
+ "learning_rate": 0.00019943348002101371,
+ "loss": 3.1371,
+ "step": 14
+ },
+ {
+ "epoch": 0.47,
+ "grad_norm": 0.65625,
+ "learning_rate": 0.00019911528310040074,
+ "loss": 3.1479,
+ "step": 15
+ },
+ {
+ "epoch": 0.5,
+ "grad_norm": 0.6484375,
+ "learning_rate": 0.00019872683547213446,
+ "loss": 3.263,
+ "step": 16
+ },
+ {
+ "epoch": 0.5,
+ "eval_loss": 3.1820719242095947,
+ "eval_runtime": 1.2692,
+ "eval_samples_per_second": 11.031,
+ "eval_steps_per_second": 5.515,
+ "step": 16
+ },
+ {
+ "epoch": 0.53,
+ "grad_norm": 0.53125,
+ "learning_rate": 0.00019826841245925212,
+ "loss": 2.9833,
+ "step": 17
+ },
+ {
+ "epoch": 0.56,
+ "grad_norm": 0.5703125,
+ "learning_rate": 0.00019774033898178667,
+ "loss": 3.0787,
+ "step": 18
+ },
+ {
+ "epoch": 0.59,
+ "grad_norm": 0.71484375,
+ "learning_rate": 0.00019714298932647098,
+ "loss": 3.4132,
+ "step": 19
+ },
+ {
+ "epoch": 0.62,
+ "grad_norm": 0.73046875,
+ "learning_rate": 0.0001964767868814516,
+ "loss": 2.7304,
+ "step": 20
+ },
+ {
+ "epoch": 0.66,
+ "grad_norm": 0.55078125,
+ "learning_rate": 0.00019574220383620055,
+ "loss": 3.0116,
+ "step": 21
+ },
+ {
+ "epoch": 0.69,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.00019493976084683813,
+ "loss": 2.9474,
+ "step": 22
+ },
+ {
+ "epoch": 0.72,
+ "grad_norm": 0.61328125,
+ "learning_rate": 0.00019407002666710336,
+ "loss": 2.9415,
+ "step": 23
+ },
+ {
+ "epoch": 0.75,
+ "grad_norm": 0.6328125,
+ "learning_rate": 0.00019313361774523385,
+ "loss": 2.798,
+ "step": 24
+ },
+ {
+ "epoch": 0.75,
+ "eval_loss": 2.896176815032959,
+ "eval_runtime": 1.2765,
+ "eval_samples_per_second": 10.967,
+ "eval_steps_per_second": 5.484,
+ "step": 24
+ },
+ {
+ "epoch": 0.78,
+ "grad_norm": 0.8046875,
+ "learning_rate": 0.00019213119778704128,
+ "loss": 3.2157,
+ "step": 25
+ },
+ {
+ "epoch": 0.81,
+ "grad_norm": 0.7109375,
+ "learning_rate": 0.00019106347728549135,
+ "loss": 3.0666,
+ "step": 26
+ },
+ {
+ "epoch": 0.84,
+ "grad_norm": 0.59765625,
+ "learning_rate": 0.00018993121301712193,
+ "loss": 2.8219,
+ "step": 27
+ },
+ {
+ "epoch": 0.88,
+ "grad_norm": 0.75390625,
+ "learning_rate": 0.00018873520750565718,
+ "loss": 3.1164,
+ "step": 28
+ },
+ {
+ "epoch": 0.91,
+ "grad_norm": 0.67578125,
+ "learning_rate": 0.00018747630845319612,
+ "loss": 2.7154,
+ "step": 29
+ },
+ {
+ "epoch": 0.94,
+ "grad_norm": 0.625,
+ "learning_rate": 0.0001861554081393806,
+ "loss": 2.7395,
+ "step": 30
+ },
+ {
+ "epoch": 0.97,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.0001847734427889671,
+ "loss": 2.8282,
+ "step": 31
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.86328125,
+ "learning_rate": 0.0001833313919082515,
+ "loss": 2.7787,
+ "step": 32
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 2.67726731300354,
+ "eval_runtime": 1.2769,
+ "eval_samples_per_second": 10.964,
+ "eval_steps_per_second": 5.482,
+ "step": 32
+ },
+ {
+ "epoch": 1.03,
+ "grad_norm": 0.703125,
+ "learning_rate": 0.0001818302775908169,
+ "loss": 2.5957,
+ "step": 33
+ },
+ {
+ "epoch": 1.06,
+ "grad_norm": 0.76953125,
+ "learning_rate": 0.00018027116379309638,
+ "loss": 2.7011,
+ "step": 34
+ },
+ {
+ "epoch": 1.09,
+ "grad_norm": 0.84765625,
+ "learning_rate": 0.00017865515558026428,
+ "loss": 2.6043,
+ "step": 35
+ },
+ {
+ "epoch": 1.12,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00017698339834299061,
+ "loss": 2.8607,
+ "step": 36
+ },
+ {
+ "epoch": 1.16,
+ "grad_norm": 0.7578125,
+ "learning_rate": 0.00017525707698561385,
+ "loss": 2.5949,
+ "step": 37
+ },
+ {
+ "epoch": 1.19,
+ "grad_norm": 0.94921875,
+ "learning_rate": 0.00017347741508630672,
+ "loss": 2.7476,
+ "step": 38
+ },
+ {
+ "epoch": 1.22,
+ "grad_norm": 0.88671875,
+ "learning_rate": 0.00017164567402983152,
+ "loss": 2.7991,
+ "step": 39
+ },
+ {
+ "epoch": 1.25,
+ "grad_norm": 0.8671875,
+ "learning_rate": 0.0001697631521134985,
+ "loss": 2.5959,
+ "step": 40
+ },
+ {
+ "epoch": 1.25,
+ "eval_loss": 2.5505764484405518,
+ "eval_runtime": 1.2761,
+ "eval_samples_per_second": 10.971,
+ "eval_steps_per_second": 5.486,
+ "step": 40
+ },
+ {
+ "epoch": 1.28,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00016783118362696163,
+ "loss": 2.5342,
+ "step": 41
+ },
+ {
+ "epoch": 1.31,
+ "grad_norm": 0.921875,
+ "learning_rate": 0.00016585113790650388,
+ "loss": 2.4969,
+ "step": 42
+ },
+ {
+ "epoch": 1.34,
+ "grad_norm": 1.125,
+ "learning_rate": 0.00016382441836448202,
+ "loss": 2.5723,
+ "step": 43
+ },
+ {
+ "epoch": 1.38,
+ "grad_norm": 1.015625,
+ "learning_rate": 0.0001617524614946192,
+ "loss": 2.6166,
+ "step": 44
+ },
+ {
+ "epoch": 1.41,
+ "grad_norm": 0.98828125,
+ "learning_rate": 0.00015963673585385016,
+ "loss": 2.4574,
+ "step": 45
+ },
+ {
+ "epoch": 1.44,
+ "grad_norm": 0.94921875,
+ "learning_rate": 0.0001574787410214407,
+ "loss": 2.5074,
+ "step": 46
+ },
+ {
+ "epoch": 1.47,
+ "grad_norm": 0.953125,
+ "learning_rate": 0.00015528000653611935,
+ "loss": 2.5394,
+ "step": 47
+ },
+ {
+ "epoch": 1.5,
+ "grad_norm": 0.828125,
+ "learning_rate": 0.00015304209081197425,
+ "loss": 2.4793,
+ "step": 48
+ },
+ {
+ "epoch": 1.5,
+ "eval_loss": 2.495466470718384,
+ "eval_runtime": 1.2814,
+ "eval_samples_per_second": 10.926,
+ "eval_steps_per_second": 5.463,
+ "step": 48
+ },
+ {
+ "epoch": 1.53,
+ "grad_norm": 0.921875,
+ "learning_rate": 0.000150766580033884,
+ "loss": 2.2909,
+ "step": 49
+ },
+ {
+ "epoch": 1.56,
+ "grad_norm": 0.74609375,
+ "learning_rate": 0.00014845508703326504,
+ "loss": 2.3424,
+ "step": 50
+ },
+ {
+ "epoch": 1.59,
+ "grad_norm": 0.87890625,
+ "learning_rate": 0.0001461092501449326,
+ "loss": 2.506,
+ "step": 51
+ },
+ {
+ "epoch": 1.62,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00014373073204588556,
+ "loss": 2.4829,
+ "step": 52
+ },
+ {
+ "epoch": 1.66,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00014132121857683783,
+ "loss": 2.481,
+ "step": 53
+ },
+ {
+ "epoch": 1.69,
+ "grad_norm": 0.9296875,
+ "learning_rate": 0.00013888241754733208,
+ "loss": 2.5512,
+ "step": 54
+ },
+ {
+ "epoch": 1.72,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00013641605752528224,
+ "loss": 2.5405,
+ "step": 55
+ },
+ {
+ "epoch": 1.75,
+ "grad_norm": 0.95703125,
+ "learning_rate": 0.00013392388661180303,
+ "loss": 2.5221,
+ "step": 56
+ },
+ {
+ "epoch": 1.75,
+ "eval_loss": 2.461298704147339,
+ "eval_runtime": 1.28,
+ "eval_samples_per_second": 10.937,
+ "eval_steps_per_second": 5.469,
+ "step": 56
+ },
+ {
+ "epoch": 1.78,
+ "grad_norm": 0.98046875,
+ "learning_rate": 0.0001314076712021949,
+ "loss": 2.5646,
+ "step": 57
+ },
+ {
+ "epoch": 1.81,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.0001288691947339621,
+ "loss": 2.5079,
+ "step": 58
+ },
+ {
+ "epoch": 1.84,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.00012631025642275212,
+ "loss": 2.4743,
+ "step": 59
+ },
+ {
+ "epoch": 1.88,
+ "grad_norm": 0.60546875,
+ "learning_rate": 0.0001237326699871115,
+ "loss": 2.3103,
+ "step": 60
+ },
+ {
+ "epoch": 1.91,
+ "grad_norm": 0.80859375,
+ "learning_rate": 0.00012113826236296244,
+ "loss": 2.4229,
+ "step": 61
+ },
+ {
+ "epoch": 1.94,
+ "grad_norm": 0.671875,
+ "learning_rate": 0.00011852887240871145,
+ "loss": 2.2709,
+ "step": 62
+ },
+ {
+ "epoch": 1.97,
+ "grad_norm": 0.8203125,
+ "learning_rate": 0.00011590634960190721,
+ "loss": 2.4868,
+ "step": 63
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.62109375,
+ "learning_rate": 0.00011327255272837221,
+ "loss": 2.4384,
+ "step": 64
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 2.4055111408233643,
+ "eval_runtime": 1.2798,
+ "eval_samples_per_second": 10.94,
+ "eval_steps_per_second": 5.47,
+ "step": 64
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 128,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 4,
+ "save_steps": 32,
+ "total_flos": 1.3334662273302528e+16,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-64/training_args.bin b/checkpoint-64/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..18814c2c9b675fc3be2d153f474e77fb88d5038e
--- /dev/null
+++ b/checkpoint-64/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5fca5508434f64a69e54fc8c46f68907814d91f48e751b7a0eeb4050e5ae3225
+size 5816
diff --git a/checkpoint-96/README.md b/checkpoint-96/README.md
new file mode 100644
index 0000000000000000000000000000000000000000..78aadd3258d05c05cabc678473b6f0942cd441c7
--- /dev/null
+++ b/checkpoint-96/README.md
@@ -0,0 +1,202 @@
+---
+library_name: peft
+base_model: TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T
+---
+
+# Model Card for Model ID
+
+
+
+
+
+## Model Details
+
+### Model Description
+
+
+
+
+
+- **Developed by:** [More Information Needed]
+- **Funded by [optional]:** [More Information Needed]
+- **Shared by [optional]:** [More Information Needed]
+- **Model type:** [More Information Needed]
+- **Language(s) (NLP):** [More Information Needed]
+- **License:** [More Information Needed]
+- **Finetuned from model [optional]:** [More Information Needed]
+
+### Model Sources [optional]
+
+
+
+- **Repository:** [More Information Needed]
+- **Paper [optional]:** [More Information Needed]
+- **Demo [optional]:** [More Information Needed]
+
+## Uses
+
+
+
+### Direct Use
+
+
+
+[More Information Needed]
+
+### Downstream Use [optional]
+
+
+
+[More Information Needed]
+
+### Out-of-Scope Use
+
+
+
+[More Information Needed]
+
+## Bias, Risks, and Limitations
+
+
+
+[More Information Needed]
+
+### Recommendations
+
+
+
+Users (both direct and downstream) should be made aware of the risks, biases and limitations of the model. More information needed for further recommendations.
+
+## How to Get Started with the Model
+
+Use the code below to get started with the model.
+
+[More Information Needed]
+
+## Training Details
+
+### Training Data
+
+
+
+[More Information Needed]
+
+### Training Procedure
+
+
+
+#### Preprocessing [optional]
+
+[More Information Needed]
+
+
+#### Training Hyperparameters
+
+- **Training regime:** [More Information Needed]
+
+#### Speeds, Sizes, Times [optional]
+
+
+
+[More Information Needed]
+
+## Evaluation
+
+
+
+### Testing Data, Factors & Metrics
+
+#### Testing Data
+
+
+
+[More Information Needed]
+
+#### Factors
+
+
+
+[More Information Needed]
+
+#### Metrics
+
+
+
+[More Information Needed]
+
+### Results
+
+[More Information Needed]
+
+#### Summary
+
+
+
+## Model Examination [optional]
+
+
+
+[More Information Needed]
+
+## Environmental Impact
+
+
+
+Carbon emissions can be estimated using the [Machine Learning Impact calculator](https://mlco2.github.io/impact#compute) presented in [Lacoste et al. (2019)](https://arxiv.org/abs/1910.09700).
+
+- **Hardware Type:** [More Information Needed]
+- **Hours used:** [More Information Needed]
+- **Cloud Provider:** [More Information Needed]
+- **Compute Region:** [More Information Needed]
+- **Carbon Emitted:** [More Information Needed]
+
+## Technical Specifications [optional]
+
+### Model Architecture and Objective
+
+[More Information Needed]
+
+### Compute Infrastructure
+
+[More Information Needed]
+
+#### Hardware
+
+[More Information Needed]
+
+#### Software
+
+[More Information Needed]
+
+## Citation [optional]
+
+
+
+**BibTeX:**
+
+[More Information Needed]
+
+**APA:**
+
+[More Information Needed]
+
+## Glossary [optional]
+
+
+
+[More Information Needed]
+
+## More Information [optional]
+
+[More Information Needed]
+
+## Model Card Authors [optional]
+
+[More Information Needed]
+
+## Model Card Contact
+
+[More Information Needed]
+### Framework versions
+
+- PEFT 0.10.0
\ No newline at end of file
diff --git a/checkpoint-96/adapter_config.json b/checkpoint-96/adapter_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..28a5d05472d8cd3bfe780706c0eded9ae55cc02a
--- /dev/null
+++ b/checkpoint-96/adapter_config.json
@@ -0,0 +1,34 @@
+{
+ "alpha_pattern": {},
+ "auto_mapping": null,
+ "base_model_name_or_path": "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
+ "bias": "none",
+ "fan_in_fan_out": null,
+ "inference_mode": true,
+ "init_lora_weights": true,
+ "layer_replication": null,
+ "layers_pattern": null,
+ "layers_to_transform": null,
+ "loftq_config": {},
+ "lora_alpha": 16,
+ "lora_dropout": 0.05,
+ "megatron_config": null,
+ "megatron_core": "megatron.core",
+ "modules_to_save": null,
+ "peft_type": "LORA",
+ "r": 32,
+ "rank_pattern": {},
+ "revision": null,
+ "target_modules": [
+ "up_proj",
+ "v_proj",
+ "o_proj",
+ "down_proj",
+ "q_proj",
+ "k_proj",
+ "gate_proj"
+ ],
+ "task_type": "CAUSAL_LM",
+ "use_dora": false,
+ "use_rslora": false
+}
\ No newline at end of file
diff --git a/checkpoint-96/adapter_model.safetensors b/checkpoint-96/adapter_model.safetensors
new file mode 100644
index 0000000000000000000000000000000000000000..b5d711bc4baca835941019d8c3024888b49ac2e4
--- /dev/null
+++ b/checkpoint-96/adapter_model.safetensors
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:c7f91dafd56822f172630cecc2b2e1daa597fdbbad4c82963c27fc7d9324b236
+size 50503848
diff --git a/checkpoint-96/optimizer.pt b/checkpoint-96/optimizer.pt
new file mode 100644
index 0000000000000000000000000000000000000000..35ba7846dec077fdeed05ee25f6ab01ee121ad94
--- /dev/null
+++ b/checkpoint-96/optimizer.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:201f8255e7b843a6dce3b0a0f20c92b3051e4c17064d38df3bcda3aea4710751
+size 202035450
diff --git a/checkpoint-96/rng_state.pth b/checkpoint-96/rng_state.pth
new file mode 100644
index 0000000000000000000000000000000000000000..2cfa30af03d9380f0d09d29e8bedfe3893e13861
--- /dev/null
+++ b/checkpoint-96/rng_state.pth
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:0cdd17d0d9f68ae884a6c41c42875839a612f5c867257621410c8670377c4664
+size 14244
diff --git a/checkpoint-96/scheduler.pt b/checkpoint-96/scheduler.pt
new file mode 100644
index 0000000000000000000000000000000000000000..86646154be5a560bcc549fed7f3d82dd6fac0066
--- /dev/null
+++ b/checkpoint-96/scheduler.pt
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:6b76c8662d73a5a13047466dda569fd16697d4a2a715b2ad8f3567f25cfa42c3
+size 1064
diff --git a/checkpoint-96/special_tokens_map.json b/checkpoint-96/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..72ecfeeb7e14d244c936169d2ed139eeae235ef1
--- /dev/null
+++ b/checkpoint-96/special_tokens_map.json
@@ -0,0 +1,24 @@
+{
+ "bos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "unk_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/checkpoint-96/tokenizer.model b/checkpoint-96/tokenizer.model
new file mode 100644
index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899
--- /dev/null
+++ b/checkpoint-96/tokenizer.model
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
+size 499723
diff --git a/checkpoint-96/tokenizer_config.json b/checkpoint-96/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..0773857a13ba5a27453a0b462624fe76e8e82a86
--- /dev/null
+++ b/checkpoint-96/tokenizer_config.json
@@ -0,0 +1,44 @@
+{
+ "add_bos_token": true,
+ "add_eos_token": false,
+ "add_prefix_space": true,
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "legacy": false,
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "padding_side": "right",
+ "sp_model_kwargs": {},
+ "spaces_between_special_tokens": false,
+ "tokenizer_class": "LlamaTokenizer",
+ "unk_token": "",
+ "use_default_system_prompt": false,
+ "use_fast": true
+}
diff --git a/checkpoint-96/trainer_state.json b/checkpoint-96/trainer_state.json
new file mode 100644
index 0000000000000000000000000000000000000000..9dd58faee91a92c197bf32775a805312c059cdac
--- /dev/null
+++ b/checkpoint-96/trainer_state.json
@@ -0,0 +1,797 @@
+{
+ "best_metric": null,
+ "best_model_checkpoint": null,
+ "epoch": 3.0,
+ "eval_steps": 8,
+ "global_step": 96,
+ "is_hyper_param_search": false,
+ "is_local_process_zero": true,
+ "is_world_process_zero": true,
+ "log_history": [
+ {
+ "epoch": 0.03,
+ "grad_norm": 0.5859375,
+ "learning_rate": 2e-05,
+ "loss": 3.9295,
+ "step": 1
+ },
+ {
+ "epoch": 0.03,
+ "eval_loss": 3.907318115234375,
+ "eval_runtime": 1.2713,
+ "eval_samples_per_second": 11.012,
+ "eval_steps_per_second": 5.506,
+ "step": 1
+ },
+ {
+ "epoch": 0.06,
+ "grad_norm": 0.5078125,
+ "learning_rate": 4e-05,
+ "loss": 3.805,
+ "step": 2
+ },
+ {
+ "epoch": 0.09,
+ "grad_norm": 0.63671875,
+ "learning_rate": 6e-05,
+ "loss": 3.8521,
+ "step": 3
+ },
+ {
+ "epoch": 0.12,
+ "grad_norm": 0.609375,
+ "learning_rate": 8e-05,
+ "loss": 3.8947,
+ "step": 4
+ },
+ {
+ "epoch": 0.16,
+ "grad_norm": 0.546875,
+ "learning_rate": 0.0001,
+ "loss": 3.6494,
+ "step": 5
+ },
+ {
+ "epoch": 0.19,
+ "grad_norm": 0.50390625,
+ "learning_rate": 0.00012,
+ "loss": 3.6457,
+ "step": 6
+ },
+ {
+ "epoch": 0.22,
+ "grad_norm": 0.640625,
+ "learning_rate": 0.00014,
+ "loss": 3.967,
+ "step": 7
+ },
+ {
+ "epoch": 0.25,
+ "grad_norm": 0.494140625,
+ "learning_rate": 0.00016,
+ "loss": 3.5364,
+ "step": 8
+ },
+ {
+ "epoch": 0.25,
+ "eval_loss": 3.6198840141296387,
+ "eval_runtime": 1.2681,
+ "eval_samples_per_second": 11.04,
+ "eval_steps_per_second": 5.52,
+ "step": 8
+ },
+ {
+ "epoch": 0.28,
+ "grad_norm": 0.51171875,
+ "learning_rate": 0.00018,
+ "loss": 3.5216,
+ "step": 9
+ },
+ {
+ "epoch": 0.31,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.0002,
+ "loss": 3.677,
+ "step": 10
+ },
+ {
+ "epoch": 0.34,
+ "grad_norm": 0.734375,
+ "learning_rate": 0.00019996456111234527,
+ "loss": 3.058,
+ "step": 11
+ },
+ {
+ "epoch": 0.38,
+ "grad_norm": 0.66796875,
+ "learning_rate": 0.0001998582695676762,
+ "loss": 3.1333,
+ "step": 12
+ },
+ {
+ "epoch": 0.41,
+ "grad_norm": 0.8359375,
+ "learning_rate": 0.000199681200703075,
+ "loss": 3.38,
+ "step": 13
+ },
+ {
+ "epoch": 0.44,
+ "grad_norm": 0.62109375,
+ "learning_rate": 0.00019943348002101371,
+ "loss": 3.1371,
+ "step": 14
+ },
+ {
+ "epoch": 0.47,
+ "grad_norm": 0.65625,
+ "learning_rate": 0.00019911528310040074,
+ "loss": 3.1479,
+ "step": 15
+ },
+ {
+ "epoch": 0.5,
+ "grad_norm": 0.6484375,
+ "learning_rate": 0.00019872683547213446,
+ "loss": 3.263,
+ "step": 16
+ },
+ {
+ "epoch": 0.5,
+ "eval_loss": 3.1820719242095947,
+ "eval_runtime": 1.2692,
+ "eval_samples_per_second": 11.031,
+ "eval_steps_per_second": 5.515,
+ "step": 16
+ },
+ {
+ "epoch": 0.53,
+ "grad_norm": 0.53125,
+ "learning_rate": 0.00019826841245925212,
+ "loss": 2.9833,
+ "step": 17
+ },
+ {
+ "epoch": 0.56,
+ "grad_norm": 0.5703125,
+ "learning_rate": 0.00019774033898178667,
+ "loss": 3.0787,
+ "step": 18
+ },
+ {
+ "epoch": 0.59,
+ "grad_norm": 0.71484375,
+ "learning_rate": 0.00019714298932647098,
+ "loss": 3.4132,
+ "step": 19
+ },
+ {
+ "epoch": 0.62,
+ "grad_norm": 0.73046875,
+ "learning_rate": 0.0001964767868814516,
+ "loss": 2.7304,
+ "step": 20
+ },
+ {
+ "epoch": 0.66,
+ "grad_norm": 0.55078125,
+ "learning_rate": 0.00019574220383620055,
+ "loss": 3.0116,
+ "step": 21
+ },
+ {
+ "epoch": 0.69,
+ "grad_norm": 0.6171875,
+ "learning_rate": 0.00019493976084683813,
+ "loss": 2.9474,
+ "step": 22
+ },
+ {
+ "epoch": 0.72,
+ "grad_norm": 0.61328125,
+ "learning_rate": 0.00019407002666710336,
+ "loss": 2.9415,
+ "step": 23
+ },
+ {
+ "epoch": 0.75,
+ "grad_norm": 0.6328125,
+ "learning_rate": 0.00019313361774523385,
+ "loss": 2.798,
+ "step": 24
+ },
+ {
+ "epoch": 0.75,
+ "eval_loss": 2.896176815032959,
+ "eval_runtime": 1.2765,
+ "eval_samples_per_second": 10.967,
+ "eval_steps_per_second": 5.484,
+ "step": 24
+ },
+ {
+ "epoch": 0.78,
+ "grad_norm": 0.8046875,
+ "learning_rate": 0.00019213119778704128,
+ "loss": 3.2157,
+ "step": 25
+ },
+ {
+ "epoch": 0.81,
+ "grad_norm": 0.7109375,
+ "learning_rate": 0.00019106347728549135,
+ "loss": 3.0666,
+ "step": 26
+ },
+ {
+ "epoch": 0.84,
+ "grad_norm": 0.59765625,
+ "learning_rate": 0.00018993121301712193,
+ "loss": 2.8219,
+ "step": 27
+ },
+ {
+ "epoch": 0.88,
+ "grad_norm": 0.75390625,
+ "learning_rate": 0.00018873520750565718,
+ "loss": 3.1164,
+ "step": 28
+ },
+ {
+ "epoch": 0.91,
+ "grad_norm": 0.67578125,
+ "learning_rate": 0.00018747630845319612,
+ "loss": 2.7154,
+ "step": 29
+ },
+ {
+ "epoch": 0.94,
+ "grad_norm": 0.625,
+ "learning_rate": 0.0001861554081393806,
+ "loss": 2.7395,
+ "step": 30
+ },
+ {
+ "epoch": 0.97,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.0001847734427889671,
+ "loss": 2.8282,
+ "step": 31
+ },
+ {
+ "epoch": 1.0,
+ "grad_norm": 0.86328125,
+ "learning_rate": 0.0001833313919082515,
+ "loss": 2.7787,
+ "step": 32
+ },
+ {
+ "epoch": 1.0,
+ "eval_loss": 2.67726731300354,
+ "eval_runtime": 1.2769,
+ "eval_samples_per_second": 10.964,
+ "eval_steps_per_second": 5.482,
+ "step": 32
+ },
+ {
+ "epoch": 1.03,
+ "grad_norm": 0.703125,
+ "learning_rate": 0.0001818302775908169,
+ "loss": 2.5957,
+ "step": 33
+ },
+ {
+ "epoch": 1.06,
+ "grad_norm": 0.76953125,
+ "learning_rate": 0.00018027116379309638,
+ "loss": 2.7011,
+ "step": 34
+ },
+ {
+ "epoch": 1.09,
+ "grad_norm": 0.84765625,
+ "learning_rate": 0.00017865515558026428,
+ "loss": 2.6043,
+ "step": 35
+ },
+ {
+ "epoch": 1.12,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00017698339834299061,
+ "loss": 2.8607,
+ "step": 36
+ },
+ {
+ "epoch": 1.16,
+ "grad_norm": 0.7578125,
+ "learning_rate": 0.00017525707698561385,
+ "loss": 2.5949,
+ "step": 37
+ },
+ {
+ "epoch": 1.19,
+ "grad_norm": 0.94921875,
+ "learning_rate": 0.00017347741508630672,
+ "loss": 2.7476,
+ "step": 38
+ },
+ {
+ "epoch": 1.22,
+ "grad_norm": 0.88671875,
+ "learning_rate": 0.00017164567402983152,
+ "loss": 2.7991,
+ "step": 39
+ },
+ {
+ "epoch": 1.25,
+ "grad_norm": 0.8671875,
+ "learning_rate": 0.0001697631521134985,
+ "loss": 2.5959,
+ "step": 40
+ },
+ {
+ "epoch": 1.25,
+ "eval_loss": 2.5505764484405518,
+ "eval_runtime": 1.2761,
+ "eval_samples_per_second": 10.971,
+ "eval_steps_per_second": 5.486,
+ "step": 40
+ },
+ {
+ "epoch": 1.28,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00016783118362696163,
+ "loss": 2.5342,
+ "step": 41
+ },
+ {
+ "epoch": 1.31,
+ "grad_norm": 0.921875,
+ "learning_rate": 0.00016585113790650388,
+ "loss": 2.4969,
+ "step": 42
+ },
+ {
+ "epoch": 1.34,
+ "grad_norm": 1.125,
+ "learning_rate": 0.00016382441836448202,
+ "loss": 2.5723,
+ "step": 43
+ },
+ {
+ "epoch": 1.38,
+ "grad_norm": 1.015625,
+ "learning_rate": 0.0001617524614946192,
+ "loss": 2.6166,
+ "step": 44
+ },
+ {
+ "epoch": 1.41,
+ "grad_norm": 0.98828125,
+ "learning_rate": 0.00015963673585385016,
+ "loss": 2.4574,
+ "step": 45
+ },
+ {
+ "epoch": 1.44,
+ "grad_norm": 0.94921875,
+ "learning_rate": 0.0001574787410214407,
+ "loss": 2.5074,
+ "step": 46
+ },
+ {
+ "epoch": 1.47,
+ "grad_norm": 0.953125,
+ "learning_rate": 0.00015528000653611935,
+ "loss": 2.5394,
+ "step": 47
+ },
+ {
+ "epoch": 1.5,
+ "grad_norm": 0.828125,
+ "learning_rate": 0.00015304209081197425,
+ "loss": 2.4793,
+ "step": 48
+ },
+ {
+ "epoch": 1.5,
+ "eval_loss": 2.495466470718384,
+ "eval_runtime": 1.2814,
+ "eval_samples_per_second": 10.926,
+ "eval_steps_per_second": 5.463,
+ "step": 48
+ },
+ {
+ "epoch": 1.53,
+ "grad_norm": 0.921875,
+ "learning_rate": 0.000150766580033884,
+ "loss": 2.2909,
+ "step": 49
+ },
+ {
+ "epoch": 1.56,
+ "grad_norm": 0.74609375,
+ "learning_rate": 0.00014845508703326504,
+ "loss": 2.3424,
+ "step": 50
+ },
+ {
+ "epoch": 1.59,
+ "grad_norm": 0.87890625,
+ "learning_rate": 0.0001461092501449326,
+ "loss": 2.506,
+ "step": 51
+ },
+ {
+ "epoch": 1.62,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00014373073204588556,
+ "loss": 2.4829,
+ "step": 52
+ },
+ {
+ "epoch": 1.66,
+ "grad_norm": 0.984375,
+ "learning_rate": 0.00014132121857683783,
+ "loss": 2.481,
+ "step": 53
+ },
+ {
+ "epoch": 1.69,
+ "grad_norm": 0.9296875,
+ "learning_rate": 0.00013888241754733208,
+ "loss": 2.5512,
+ "step": 54
+ },
+ {
+ "epoch": 1.72,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00013641605752528224,
+ "loss": 2.5405,
+ "step": 55
+ },
+ {
+ "epoch": 1.75,
+ "grad_norm": 0.95703125,
+ "learning_rate": 0.00013392388661180303,
+ "loss": 2.5221,
+ "step": 56
+ },
+ {
+ "epoch": 1.75,
+ "eval_loss": 2.461298704147339,
+ "eval_runtime": 1.28,
+ "eval_samples_per_second": 10.937,
+ "eval_steps_per_second": 5.469,
+ "step": 56
+ },
+ {
+ "epoch": 1.78,
+ "grad_norm": 0.98046875,
+ "learning_rate": 0.0001314076712021949,
+ "loss": 2.5646,
+ "step": 57
+ },
+ {
+ "epoch": 1.81,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.0001288691947339621,
+ "loss": 2.5079,
+ "step": 58
+ },
+ {
+ "epoch": 1.84,
+ "grad_norm": 0.8125,
+ "learning_rate": 0.00012631025642275212,
+ "loss": 2.4743,
+ "step": 59
+ },
+ {
+ "epoch": 1.88,
+ "grad_norm": 0.60546875,
+ "learning_rate": 0.0001237326699871115,
+ "loss": 2.3103,
+ "step": 60
+ },
+ {
+ "epoch": 1.91,
+ "grad_norm": 0.80859375,
+ "learning_rate": 0.00012113826236296244,
+ "loss": 2.4229,
+ "step": 61
+ },
+ {
+ "epoch": 1.94,
+ "grad_norm": 0.671875,
+ "learning_rate": 0.00011852887240871145,
+ "loss": 2.2709,
+ "step": 62
+ },
+ {
+ "epoch": 1.97,
+ "grad_norm": 0.8203125,
+ "learning_rate": 0.00011590634960190721,
+ "loss": 2.4868,
+ "step": 63
+ },
+ {
+ "epoch": 2.0,
+ "grad_norm": 0.62109375,
+ "learning_rate": 0.00011327255272837221,
+ "loss": 2.4384,
+ "step": 64
+ },
+ {
+ "epoch": 2.0,
+ "eval_loss": 2.4055111408233643,
+ "eval_runtime": 1.2798,
+ "eval_samples_per_second": 10.94,
+ "eval_steps_per_second": 5.47,
+ "step": 64
+ },
+ {
+ "epoch": 2.03,
+ "grad_norm": 0.6328125,
+ "learning_rate": 0.00011062934856473655,
+ "loss": 2.3671,
+ "step": 65
+ },
+ {
+ "epoch": 2.06,
+ "grad_norm": 0.69140625,
+ "learning_rate": 0.00010797861055530831,
+ "loss": 2.3381,
+ "step": 66
+ },
+ {
+ "epoch": 2.09,
+ "grad_norm": 0.671875,
+ "learning_rate": 0.00010532221748421787,
+ "loss": 2.2134,
+ "step": 67
+ },
+ {
+ "epoch": 2.12,
+ "grad_norm": 0.7890625,
+ "learning_rate": 0.00010266205214377748,
+ "loss": 2.2687,
+ "step": 68
+ },
+ {
+ "epoch": 2.16,
+ "grad_norm": 0.75,
+ "learning_rate": 0.0001,
+ "loss": 2.4273,
+ "step": 69
+ },
+ {
+ "epoch": 2.19,
+ "grad_norm": 0.8671875,
+ "learning_rate": 9.733794785622253e-05,
+ "loss": 2.4439,
+ "step": 70
+ },
+ {
+ "epoch": 2.22,
+ "grad_norm": 0.953125,
+ "learning_rate": 9.467778251578217e-05,
+ "loss": 2.6631,
+ "step": 71
+ },
+ {
+ "epoch": 2.25,
+ "grad_norm": 0.78125,
+ "learning_rate": 9.202138944469168e-05,
+ "loss": 2.295,
+ "step": 72
+ },
+ {
+ "epoch": 2.25,
+ "eval_loss": 2.3922784328460693,
+ "eval_runtime": 1.282,
+ "eval_samples_per_second": 10.921,
+ "eval_steps_per_second": 5.46,
+ "step": 72
+ },
+ {
+ "epoch": 2.28,
+ "grad_norm": 0.734375,
+ "learning_rate": 8.937065143526347e-05,
+ "loss": 2.4963,
+ "step": 73
+ },
+ {
+ "epoch": 2.31,
+ "grad_norm": 0.68359375,
+ "learning_rate": 8.672744727162781e-05,
+ "loss": 2.4274,
+ "step": 74
+ },
+ {
+ "epoch": 2.34,
+ "grad_norm": 0.9765625,
+ "learning_rate": 8.409365039809281e-05,
+ "loss": 2.4988,
+ "step": 75
+ },
+ {
+ "epoch": 2.38,
+ "grad_norm": 0.75390625,
+ "learning_rate": 8.147112759128859e-05,
+ "loss": 2.2886,
+ "step": 76
+ },
+ {
+ "epoch": 2.41,
+ "grad_norm": 0.80078125,
+ "learning_rate": 7.886173763703757e-05,
+ "loss": 2.1944,
+ "step": 77
+ },
+ {
+ "epoch": 2.44,
+ "grad_norm": 0.8828125,
+ "learning_rate": 7.626733001288851e-05,
+ "loss": 2.3283,
+ "step": 78
+ },
+ {
+ "epoch": 2.47,
+ "grad_norm": 0.82421875,
+ "learning_rate": 7.368974357724789e-05,
+ "loss": 2.3855,
+ "step": 79
+ },
+ {
+ "epoch": 2.5,
+ "grad_norm": 0.7890625,
+ "learning_rate": 7.113080526603792e-05,
+ "loss": 2.3943,
+ "step": 80
+ },
+ {
+ "epoch": 2.5,
+ "eval_loss": 2.386228084564209,
+ "eval_runtime": 1.2835,
+ "eval_samples_per_second": 10.908,
+ "eval_steps_per_second": 5.454,
+ "step": 80
+ },
+ {
+ "epoch": 2.53,
+ "grad_norm": 0.94921875,
+ "learning_rate": 6.859232879780515e-05,
+ "loss": 2.2725,
+ "step": 81
+ },
+ {
+ "epoch": 2.56,
+ "grad_norm": 0.75,
+ "learning_rate": 6.607611338819697e-05,
+ "loss": 2.1989,
+ "step": 82
+ },
+ {
+ "epoch": 2.59,
+ "grad_norm": 0.703125,
+ "learning_rate": 6.358394247471778e-05,
+ "loss": 2.1219,
+ "step": 83
+ },
+ {
+ "epoch": 2.62,
+ "grad_norm": 1.0390625,
+ "learning_rate": 6.111758245266794e-05,
+ "loss": 2.3478,
+ "step": 84
+ },
+ {
+ "epoch": 2.66,
+ "grad_norm": 0.78125,
+ "learning_rate": 5.867878142316221e-05,
+ "loss": 2.4196,
+ "step": 85
+ },
+ {
+ "epoch": 2.69,
+ "grad_norm": 0.6796875,
+ "learning_rate": 5.626926795411447e-05,
+ "loss": 2.1882,
+ "step": 86
+ },
+ {
+ "epoch": 2.72,
+ "grad_norm": 0.8203125,
+ "learning_rate": 5.38907498550674e-05,
+ "loss": 2.4703,
+ "step": 87
+ },
+ {
+ "epoch": 2.75,
+ "grad_norm": 0.8671875,
+ "learning_rate": 5.1544912966734994e-05,
+ "loss": 2.2398,
+ "step": 88
+ },
+ {
+ "epoch": 2.75,
+ "eval_loss": 2.3605105876922607,
+ "eval_runtime": 1.2887,
+ "eval_samples_per_second": 10.864,
+ "eval_steps_per_second": 5.432,
+ "step": 88
+ },
+ {
+ "epoch": 2.78,
+ "grad_norm": 0.81640625,
+ "learning_rate": 4.9233419966116036e-05,
+ "loss": 2.2939,
+ "step": 89
+ },
+ {
+ "epoch": 2.81,
+ "grad_norm": 1.0234375,
+ "learning_rate": 4.695790918802576e-05,
+ "loss": 2.2835,
+ "step": 90
+ },
+ {
+ "epoch": 2.84,
+ "grad_norm": 0.85546875,
+ "learning_rate": 4.47199934638807e-05,
+ "loss": 2.3145,
+ "step": 91
+ },
+ {
+ "epoch": 2.88,
+ "grad_norm": 0.5546875,
+ "learning_rate": 4.252125897855932e-05,
+ "loss": 2.2521,
+ "step": 92
+ },
+ {
+ "epoch": 2.91,
+ "grad_norm": 0.76953125,
+ "learning_rate": 4.036326414614985e-05,
+ "loss": 2.377,
+ "step": 93
+ },
+ {
+ "epoch": 2.94,
+ "grad_norm": 0.78125,
+ "learning_rate": 3.824753850538082e-05,
+ "loss": 2.4343,
+ "step": 94
+ },
+ {
+ "epoch": 2.97,
+ "grad_norm": 0.8046875,
+ "learning_rate": 3.617558163551802e-05,
+ "loss": 2.2051,
+ "step": 95
+ },
+ {
+ "epoch": 3.0,
+ "grad_norm": 0.88671875,
+ "learning_rate": 3.414886209349615e-05,
+ "loss": 2.2693,
+ "step": 96
+ },
+ {
+ "epoch": 3.0,
+ "eval_loss": 2.3525888919830322,
+ "eval_runtime": 1.2821,
+ "eval_samples_per_second": 10.919,
+ "eval_steps_per_second": 5.46,
+ "step": 96
+ }
+ ],
+ "logging_steps": 1,
+ "max_steps": 128,
+ "num_input_tokens_seen": 0,
+ "num_train_epochs": 4,
+ "save_steps": 32,
+ "total_flos": 2.000199340995379e+16,
+ "train_batch_size": 2,
+ "trial_name": null,
+ "trial_params": null
+}
diff --git a/checkpoint-96/training_args.bin b/checkpoint-96/training_args.bin
new file mode 100644
index 0000000000000000000000000000000000000000..18814c2c9b675fc3be2d153f474e77fb88d5038e
--- /dev/null
+++ b/checkpoint-96/training_args.bin
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:5fca5508434f64a69e54fc8c46f68907814d91f48e751b7a0eeb4050e5ae3225
+size 5816
diff --git a/config.json b/config.json
new file mode 100644
index 0000000000000000000000000000000000000000..90a576b0e1406cefb938f55520f89406e75af264
--- /dev/null
+++ b/config.json
@@ -0,0 +1,43 @@
+{
+ "_name_or_path": "TinyLlama/TinyLlama-1.1B-intermediate-step-1431k-3T",
+ "architectures": [
+ "LlamaForCausalLM"
+ ],
+ "attention_bias": false,
+ "attention_dropout": 0.0,
+ "bos_token_id": 1,
+ "eos_token_id": 2,
+ "hidden_act": "silu",
+ "hidden_size": 2048,
+ "initializer_range": 0.02,
+ "intermediate_size": 5632,
+ "max_position_embeddings": 4096,
+ "model_type": "llama",
+ "num_attention_heads": 32,
+ "num_hidden_layers": 22,
+ "num_key_value_heads": 4,
+ "pretraining_tp": 1,
+ "quantization_config": {
+ "_load_in_4bit": true,
+ "_load_in_8bit": false,
+ "bnb_4bit_compute_dtype": "bfloat16",
+ "bnb_4bit_quant_storage": "bfloat16",
+ "bnb_4bit_quant_type": "nf4",
+ "bnb_4bit_use_double_quant": true,
+ "llm_int8_enable_fp32_cpu_offload": false,
+ "llm_int8_has_fp16_weight": false,
+ "llm_int8_skip_modules": null,
+ "llm_int8_threshold": 6.0,
+ "load_in_4bit": true,
+ "load_in_8bit": false,
+ "quant_method": "bitsandbytes"
+ },
+ "rms_norm_eps": 1e-05,
+ "rope_scaling": null,
+ "rope_theta": 10000.0,
+ "tie_word_embeddings": false,
+ "torch_dtype": "float32",
+ "transformers_version": "4.40.0.dev0",
+ "use_cache": false,
+ "vocab_size": 32000
+}
diff --git a/runs/Apr30_02-51-48_663ec5cd7167/events.out.tfevents.1714445508.663ec5cd7167.5280.0 b/runs/Apr30_02-51-48_663ec5cd7167/events.out.tfevents.1714445508.663ec5cd7167.5280.0
new file mode 100644
index 0000000000000000000000000000000000000000..66828609cbc8a3db07a68014d3322946cab862bc
--- /dev/null
+++ b/runs/Apr30_02-51-48_663ec5cd7167/events.out.tfevents.1714445508.663ec5cd7167.5280.0
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:a3bac99fef0086cdfda4f010adba206b97074810136dff597980b02f3f4736af
+size 37468
diff --git a/special_tokens_map.json b/special_tokens_map.json
new file mode 100644
index 0000000000000000000000000000000000000000..72ecfeeb7e14d244c936169d2ed139eeae235ef1
--- /dev/null
+++ b/special_tokens_map.json
@@ -0,0 +1,24 @@
+{
+ "bos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "eos_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ },
+ "pad_token": "",
+ "unk_token": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false
+ }
+}
diff --git a/tokenizer.model b/tokenizer.model
new file mode 100644
index 0000000000000000000000000000000000000000..6c00c742ce03c627d6cd5b795984876fa49fa899
--- /dev/null
+++ b/tokenizer.model
@@ -0,0 +1,3 @@
+version https://git-lfs.github.com/spec/v1
+oid sha256:9e556afd44213b6bd1be2b850ebbbd98f5481437a8021afaf58ee7fb1818d347
+size 499723
diff --git a/tokenizer_config.json b/tokenizer_config.json
new file mode 100644
index 0000000000000000000000000000000000000000..0773857a13ba5a27453a0b462624fe76e8e82a86
--- /dev/null
+++ b/tokenizer_config.json
@@ -0,0 +1,44 @@
+{
+ "add_bos_token": true,
+ "add_eos_token": false,
+ "add_prefix_space": true,
+ "added_tokens_decoder": {
+ "0": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "1": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ },
+ "2": {
+ "content": "",
+ "lstrip": false,
+ "normalized": false,
+ "rstrip": false,
+ "single_word": false,
+ "special": true
+ }
+ },
+ "bos_token": "",
+ "clean_up_tokenization_spaces": false,
+ "eos_token": "",
+ "legacy": false,
+ "model_max_length": 1000000000000000019884624838656,
+ "pad_token": "",
+ "padding_side": "right",
+ "sp_model_kwargs": {},
+ "spaces_between_special_tokens": false,
+ "tokenizer_class": "LlamaTokenizer",
+ "unk_token": "",
+ "use_default_system_prompt": false,
+ "use_fast": true
+}