pandyamarut
commited on
End of training
Browse files- README.md +7 -7
- adapter_model.bin +1 -1
README.md
CHANGED
@@ -58,7 +58,7 @@ optimizer: adamw_8bit
|
|
58 |
output_dir: /runpod-volume/fine-tuning/test-run
|
59 |
pad_to_sequence_len: true
|
60 |
run_name: test-run
|
61 |
-
runpod_job_id:
|
62 |
sample_packing: true
|
63 |
saves_per_epoch: 1
|
64 |
sequence_len: 2048
|
@@ -82,7 +82,7 @@ weight_decay: 0
|
|
82 |
|
83 |
This model is a fine-tuned version of [NousResearch/Llama-3.2-1B](https://huggingface.co/NousResearch/Llama-3.2-1B) on the teknium/GPT4-LLM-Cleaned dataset.
|
84 |
It achieves the following results on the evaluation set:
|
85 |
-
- Loss: 1.
|
86 |
|
87 |
## Model description
|
88 |
|
@@ -117,15 +117,15 @@ The following hyperparameters were used during training:
|
|
117 |
| Training Loss | Epoch | Step | Validation Loss |
|
118 |
|:-------------:|:------:|:----:|:---------------:|
|
119 |
| 1.4537 | 0.0009 | 1 | 1.3971 |
|
120 |
-
| 1.
|
121 |
-
| 1.
|
122 |
-
| 1.
|
123 |
|
124 |
|
125 |
### Framework versions
|
126 |
|
127 |
- PEFT 0.14.0
|
128 |
- Transformers 4.47.1
|
129 |
-
- Pytorch 2.
|
130 |
-
- Datasets 3.
|
131 |
- Tokenizers 0.21.0
|
|
|
58 |
output_dir: /runpod-volume/fine-tuning/test-run
|
59 |
pad_to_sequence_len: true
|
60 |
run_name: test-run
|
61 |
+
runpod_job_id: dd327f42-5f67-4830-b512-4561fa9a3d45-u1
|
62 |
sample_packing: true
|
63 |
saves_per_epoch: 1
|
64 |
sequence_len: 2048
|
|
|
82 |
|
83 |
This model is a fine-tuned version of [NousResearch/Llama-3.2-1B](https://huggingface.co/NousResearch/Llama-3.2-1B) on the teknium/GPT4-LLM-Cleaned dataset.
|
84 |
It achieves the following results on the evaluation set:
|
85 |
+
- Loss: 1.1018
|
86 |
|
87 |
## Model description
|
88 |
|
|
|
117 |
| Training Loss | Epoch | Step | Validation Loss |
|
118 |
|:-------------:|:------:|:----:|:---------------:|
|
119 |
| 1.4537 | 0.0009 | 1 | 1.3971 |
|
120 |
+
| 1.1978 | 0.2503 | 271 | 1.1561 |
|
121 |
+
| 1.1637 | 0.5007 | 542 | 1.1131 |
|
122 |
+
| 1.1894 | 0.7510 | 813 | 1.1018 |
|
123 |
|
124 |
|
125 |
### Framework versions
|
126 |
|
127 |
- PEFT 0.14.0
|
128 |
- Transformers 4.47.1
|
129 |
+
- Pytorch 2.5.1+cu124
|
130 |
+
- Datasets 3.2.0
|
131 |
- Tokenizers 0.21.0
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 45169354
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8e60fa14c8637e48775bc40d495a321bde53978310acf0b5847bb380ec614a2c
|
3 |
size 45169354
|