DuongTrongChi commited on
Commit
b450cfe
1 Parent(s): de8e3c0

Model save

Browse files
Files changed (1) hide show
  1. README.md +6 -6
README.md CHANGED
@@ -1,13 +1,13 @@
1
  ---
2
- base_model: Qwen/Qwen2.5-0.5B-Instruct
3
- datasets:
4
- - generator
5
  library_name: peft
6
  license: apache-2.0
 
7
  tags:
8
  - trl
9
  - sft
10
  - generated_from_trainer
 
 
11
  model-index:
12
  - name: trained_models
13
  results: []
@@ -43,7 +43,7 @@ The following hyperparameters were used during training:
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
  - total_train_batch_size: 6
46
- - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
  - lr_scheduler_type: constant
48
  - lr_scheduler_warmup_ratio: 0.03
49
  - num_epochs: 1
@@ -55,7 +55,7 @@ The following hyperparameters were used during training:
55
  ### Framework versions
56
 
57
  - PEFT 0.13.2
58
- - Transformers 4.45.2
59
  - Pytorch 2.2.1+cu121
60
- - Datasets 3.0.1
61
  - Tokenizers 0.20.1
 
1
  ---
 
 
 
2
  library_name: peft
3
  license: apache-2.0
4
+ base_model: Qwen/Qwen2.5-0.5B-Instruct
5
  tags:
6
  - trl
7
  - sft
8
  - generated_from_trainer
9
+ datasets:
10
+ - generator
11
  model-index:
12
  - name: trained_models
13
  results: []
 
43
  - seed: 42
44
  - gradient_accumulation_steps: 2
45
  - total_train_batch_size: 6
46
+ - optimizer: Use adamw_torch_fused with betas=(0.9,0.999) and epsilon=1e-08 and optimizer_args=No additional optimizer arguments
47
  - lr_scheduler_type: constant
48
  - lr_scheduler_warmup_ratio: 0.03
49
  - num_epochs: 1
 
55
  ### Framework versions
56
 
57
  - PEFT 0.13.2
58
+ - Transformers 4.46.0
59
  - Pytorch 2.2.1+cu121
60
+ - Datasets 3.0.2
61
  - Tokenizers 0.20.1