pandyamarut commited on
Commit
0fe279d
·
verified ·
1 Parent(s): 07360ae

End of training

Browse files
Files changed (2) hide show
  1. README.md +7 -7
  2. adapter_model.bin +1 -1
README.md CHANGED
@@ -58,7 +58,7 @@ optimizer: adamw_8bit
58
  output_dir: /runpod-volume/fine-tuning/test-run
59
  pad_to_sequence_len: true
60
  run_name: test-run
61
- runpod_job_id: b7693c20-f1ab-4572-ad4f-bf19fa790d82-u1
62
  sample_packing: true
63
  saves_per_epoch: 1
64
  sequence_len: 2048
@@ -82,7 +82,7 @@ weight_decay: 0
82
 
83
  This model is a fine-tuned version of [NousResearch/Llama-3.2-1B](https://huggingface.co/NousResearch/Llama-3.2-1B) on the teknium/GPT4-LLM-Cleaned dataset.
84
  It achieves the following results on the evaluation set:
85
- - Loss: 1.1014
86
 
87
  ## Model description
88
 
@@ -117,15 +117,15 @@ The following hyperparameters were used during training:
117
  | Training Loss | Epoch | Step | Validation Loss |
118
  |:-------------:|:------:|:----:|:---------------:|
119
  | 1.4537 | 0.0009 | 1 | 1.3971 |
120
- | 1.1953 | 0.2503 | 271 | 1.1562 |
121
- | 1.1678 | 0.5007 | 542 | 1.1135 |
122
- | 1.1912 | 0.7510 | 813 | 1.1014 |
123
 
124
 
125
  ### Framework versions
126
 
127
  - PEFT 0.14.0
128
  - Transformers 4.47.1
129
- - Pytorch 2.3.1+cu121
130
- - Datasets 3.1.0
131
  - Tokenizers 0.21.0
 
58
  output_dir: /runpod-volume/fine-tuning/test-run
59
  pad_to_sequence_len: true
60
  run_name: test-run
61
+ runpod_job_id: dd327f42-5f67-4830-b512-4561fa9a3d45-u1
62
  sample_packing: true
63
  saves_per_epoch: 1
64
  sequence_len: 2048
 
82
 
83
  This model is a fine-tuned version of [NousResearch/Llama-3.2-1B](https://huggingface.co/NousResearch/Llama-3.2-1B) on the teknium/GPT4-LLM-Cleaned dataset.
84
  It achieves the following results on the evaluation set:
85
+ - Loss: 1.1018
86
 
87
  ## Model description
88
 
 
117
  | Training Loss | Epoch | Step | Validation Loss |
118
  |:-------------:|:------:|:----:|:---------------:|
119
  | 1.4537 | 0.0009 | 1 | 1.3971 |
120
+ | 1.1978 | 0.2503 | 271 | 1.1561 |
121
+ | 1.1637 | 0.5007 | 542 | 1.1131 |
122
+ | 1.1894 | 0.7510 | 813 | 1.1018 |
123
 
124
 
125
  ### Framework versions
126
 
127
  - PEFT 0.14.0
128
  - Transformers 4.47.1
129
+ - Pytorch 2.5.1+cu124
130
+ - Datasets 3.2.0
131
  - Tokenizers 0.21.0
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:79b8de8c3045ce62cac19cefbf984c4d2db1a67dd6e142abdce6e355a504e44b
3
  size 45169354
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:8e60fa14c8637e48775bc40d495a321bde53978310acf0b5847bb380ec614a2c
3
  size 45169354