samoline commited on
Commit
b6bc4e2
·
verified ·
1 Parent(s): a718e08

End of training

Browse files
Files changed (2) hide show
  1. README.md +15 -14
  2. adapter_model.bin +2 -2
README.md CHANGED
@@ -1,7 +1,7 @@
1
  ---
2
  library_name: peft
3
  license: apache-2.0
4
- base_model: Qwen/Qwen2-0.5B
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
@@ -19,18 +19,17 @@ should probably proofread and complete it, then remove this comment. -->
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
- base_model: Qwen/Qwen2-0.5B
23
  bf16: auto
24
  chat_template: llama3
25
  dataset_prepared_path: null
26
  datasets:
27
  - format: custom
28
- path: mhenrichsen/alpaca_2k_test
29
  type:
30
- field_input: input
31
- field_instruction: instruction
32
- field_output: output
33
- format: '{instruction} {input}'
34
  no_input_format: '{instruction}'
35
  system_format: '{system}'
36
  system_prompt: ''
@@ -65,7 +64,7 @@ lora_target_linear: true
65
  lr_scheduler: cosine
66
  max_steps: 10
67
  micro_batch_size: 2
68
- mlflow_experiment_name: mhenrichsen/alpaca_2k_test
69
  model_type: AutoModelForCausalLM
70
  num_epochs: 1
71
  optimizer: adamw_bnb_8bit
@@ -76,6 +75,8 @@ s2_attention: null
76
  sample_packing: false
77
  saves_per_epoch: 4
78
  sequence_len: 512
 
 
79
  strict: false
80
  tf32: false
81
  tokenizer_type: AutoTokenizer
@@ -98,9 +99,9 @@ xformers_attention: null
98
 
99
  # e3947cb3-7c20-41c7-b215-9a5e96c2917b
100
 
101
- This model is a fine-tuned version of [Qwen/Qwen2-0.5B](https://huggingface.co/Qwen/Qwen2-0.5B) on the None dataset.
102
  It achieves the following results on the evaluation set:
103
- - Loss: 1.5238
104
 
105
  ## Model description
106
 
@@ -134,10 +135,10 @@ The following hyperparameters were used during training:
134
 
135
  | Training Loss | Epoch | Step | Validation Loss |
136
  |:-------------:|:------:|:----:|:---------------:|
137
- | 1.5558 | 0.0042 | 1 | 1.5606 |
138
- | 1.402 | 0.0126 | 3 | 1.5598 |
139
- | 1.3486 | 0.0253 | 6 | 1.5497 |
140
- | 1.4565 | 0.0379 | 9 | 1.5238 |
141
 
142
 
143
  ### Framework versions
 
1
  ---
2
  library_name: peft
3
  license: apache-2.0
4
+ base_model: Maykeye/TinyLLama-v0
5
  tags:
6
  - axolotl
7
  - generated_from_trainer
 
19
  axolotl version: `0.4.1`
20
  ```yaml
21
  adapter: lora
22
+ base_model: Maykeye/TinyLLama-v0
23
  bf16: auto
24
  chat_template: llama3
25
  dataset_prepared_path: null
26
  datasets:
27
  - format: custom
28
+ path: samoline/api-generator
29
  type:
30
+ field_instruction: query
31
+ field_output: functions
32
+ format: '{instruction}'
 
33
  no_input_format: '{instruction}'
34
  system_format: '{system}'
35
  system_prompt: ''
 
64
  lr_scheduler: cosine
65
  max_steps: 10
66
  micro_batch_size: 2
67
+ mlflow_experiment_name: samoline/api-generator
68
  model_type: AutoModelForCausalLM
69
  num_epochs: 1
70
  optimizer: adamw_bnb_8bit
 
75
  sample_packing: false
76
  saves_per_epoch: 4
77
  sequence_len: 512
78
+ special_tokens:
79
+ pad_token: </s>
80
  strict: false
81
  tf32: false
82
  tokenizer_type: AutoTokenizer
 
99
 
100
  # e3947cb3-7c20-41c7-b215-9a5e96c2917b
101
 
102
+ This model is a fine-tuned version of [Maykeye/TinyLLama-v0](https://huggingface.co/Maykeye/TinyLLama-v0) on the None dataset.
103
  It achieves the following results on the evaluation set:
104
+ - Loss: 10.0749
105
 
106
  ## Model description
107
 
 
135
 
136
  | Training Loss | Epoch | Step | Validation Loss |
137
  |:-------------:|:------:|:----:|:---------------:|
138
+ | 10.0927 | 0.0010 | 1 | 10.5287 |
139
+ | 10.8992 | 0.0031 | 3 | 10.5203 |
140
+ | 10.3469 | 0.0062 | 6 | 10.3812 |
141
+ | 10.3735 | 0.0092 | 9 | 10.0749 |
142
 
143
 
144
  ### Framework versions
adapter_model.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:fc077b1f1dc59430ba408afb480d511f372e3c86cc07bdd7903fa45a4ebb65af
3
- size 17717130
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:9712679a4c96a8630eb2bc4dc4c977031474c204fbe23d309a99cda0da01caf1
3
+ size 416906