End of training
Browse files- README.md +15 -14
- adapter_model.bin +2 -2
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
---
|
2 |
library_name: peft
|
3 |
license: apache-2.0
|
4 |
-
base_model:
|
5 |
tags:
|
6 |
- axolotl
|
7 |
- generated_from_trainer
|
@@ -19,18 +19,17 @@ should probably proofread and complete it, then remove this comment. -->
|
|
19 |
axolotl version: `0.4.1`
|
20 |
```yaml
|
21 |
adapter: lora
|
22 |
-
base_model:
|
23 |
bf16: auto
|
24 |
chat_template: llama3
|
25 |
dataset_prepared_path: null
|
26 |
datasets:
|
27 |
- format: custom
|
28 |
-
path:
|
29 |
type:
|
30 |
-
|
31 |
-
|
32 |
-
|
33 |
-
format: '{instruction} {input}'
|
34 |
no_input_format: '{instruction}'
|
35 |
system_format: '{system}'
|
36 |
system_prompt: ''
|
@@ -65,7 +64,7 @@ lora_target_linear: true
|
|
65 |
lr_scheduler: cosine
|
66 |
max_steps: 10
|
67 |
micro_batch_size: 2
|
68 |
-
mlflow_experiment_name:
|
69 |
model_type: AutoModelForCausalLM
|
70 |
num_epochs: 1
|
71 |
optimizer: adamw_bnb_8bit
|
@@ -76,6 +75,8 @@ s2_attention: null
|
|
76 |
sample_packing: false
|
77 |
saves_per_epoch: 4
|
78 |
sequence_len: 512
|
|
|
|
|
79 |
strict: false
|
80 |
tf32: false
|
81 |
tokenizer_type: AutoTokenizer
|
@@ -98,9 +99,9 @@ xformers_attention: null
|
|
98 |
|
99 |
# e3947cb3-7c20-41c7-b215-9a5e96c2917b
|
100 |
|
101 |
-
This model is a fine-tuned version of [
|
102 |
It achieves the following results on the evaluation set:
|
103 |
-
- Loss:
|
104 |
|
105 |
## Model description
|
106 |
|
@@ -134,10 +135,10 @@ The following hyperparameters were used during training:
|
|
134 |
|
135 |
| Training Loss | Epoch | Step | Validation Loss |
|
136 |
|:-------------:|:------:|:----:|:---------------:|
|
137 |
-
|
|
138 |
-
|
|
139 |
-
|
|
140 |
-
|
|
141 |
|
142 |
|
143 |
### Framework versions
|
|
|
1 |
---
|
2 |
library_name: peft
|
3 |
license: apache-2.0
|
4 |
+
base_model: Maykeye/TinyLLama-v0
|
5 |
tags:
|
6 |
- axolotl
|
7 |
- generated_from_trainer
|
|
|
19 |
axolotl version: `0.4.1`
|
20 |
```yaml
|
21 |
adapter: lora
|
22 |
+
base_model: Maykeye/TinyLLama-v0
|
23 |
bf16: auto
|
24 |
chat_template: llama3
|
25 |
dataset_prepared_path: null
|
26 |
datasets:
|
27 |
- format: custom
|
28 |
+
path: samoline/api-generator
|
29 |
type:
|
30 |
+
field_instruction: query
|
31 |
+
field_output: functions
|
32 |
+
format: '{instruction}'
|
|
|
33 |
no_input_format: '{instruction}'
|
34 |
system_format: '{system}'
|
35 |
system_prompt: ''
|
|
|
64 |
lr_scheduler: cosine
|
65 |
max_steps: 10
|
66 |
micro_batch_size: 2
|
67 |
+
mlflow_experiment_name: samoline/api-generator
|
68 |
model_type: AutoModelForCausalLM
|
69 |
num_epochs: 1
|
70 |
optimizer: adamw_bnb_8bit
|
|
|
75 |
sample_packing: false
|
76 |
saves_per_epoch: 4
|
77 |
sequence_len: 512
|
78 |
+
special_tokens:
|
79 |
+
pad_token: </s>
|
80 |
strict: false
|
81 |
tf32: false
|
82 |
tokenizer_type: AutoTokenizer
|
|
|
99 |
|
100 |
# e3947cb3-7c20-41c7-b215-9a5e96c2917b
|
101 |
|
102 |
+
This model is a fine-tuned version of [Maykeye/TinyLLama-v0](https://huggingface.co/Maykeye/TinyLLama-v0) on the None dataset.
|
103 |
It achieves the following results on the evaluation set:
|
104 |
+
- Loss: 10.0749
|
105 |
|
106 |
## Model description
|
107 |
|
|
|
135 |
|
136 |
| Training Loss | Epoch | Step | Validation Loss |
|
137 |
|:-------------:|:------:|:----:|:---------------:|
|
138 |
+
| 10.0927 | 0.0010 | 1 | 10.5287 |
|
139 |
+
| 10.8992 | 0.0031 | 3 | 10.5203 |
|
140 |
+
| 10.3469 | 0.0062 | 6 | 10.3812 |
|
141 |
+
| 10.3735 | 0.0092 | 9 | 10.0749 |
|
142 |
|
143 |
|
144 |
### Framework versions
|
adapter_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
-
size
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:9712679a4c96a8630eb2bc4dc4c977031474c204fbe23d309a99cda0da01caf1
|
3 |
+
size 416906
|