steffygreypaul commited on
Commit
e5a44bf
·
verified ·
1 Parent(s): 1d7976d

End of training

Browse files
README.md CHANGED
@@ -1,5 +1,5 @@
1
  ---
2
- base_model: meta-llama/Llama-3.2-1B-Instruct
3
  datasets:
4
  - generator
5
  library_name: peft
@@ -18,7 +18,7 @@ should probably proofread and complete it, then remove this comment. -->
18
 
19
  # output
20
 
21
- This model is a fine-tuned version of [meta-llama/Llama-3.2-1B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-1B-Instruct) on the generator dataset.
22
 
23
  ## Model description
24
 
@@ -44,7 +44,8 @@ The following hyperparameters were used during training:
44
  - gradient_accumulation_steps: 4
45
  - total_train_batch_size: 4
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
- - lr_scheduler_type: linear
 
48
  - num_epochs: 1
49
  - mixed_precision_training: Native AMP
50
 
 
1
  ---
2
+ base_model: meta-llama/Llama-3.2-3B-Instruct
3
  datasets:
4
  - generator
5
  library_name: peft
 
18
 
19
  # output
20
 
21
+ This model is a fine-tuned version of [meta-llama/Llama-3.2-3B-Instruct](https://huggingface.co/meta-llama/Llama-3.2-3B-Instruct) on the generator dataset.
22
 
23
  ## Model description
24
 
 
44
  - gradient_accumulation_steps: 4
45
  - total_train_batch_size: 4
46
  - optimizer: Adam with betas=(0.9,0.999) and epsilon=1e-08
47
+ - lr_scheduler_type: constant
48
+ - lr_scheduler_warmup_ratio: 0.03
49
  - num_epochs: 1
50
  - mixed_precision_training: Native AMP
51
 
adapter_config.json CHANGED
@@ -1,7 +1,7 @@
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
- "base_model_name_or_path": "meta-llama/Llama-3.2-1B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
 
1
  {
2
  "alpha_pattern": {},
3
  "auto_mapping": null,
4
+ "base_model_name_or_path": "meta-llama/Llama-3.2-3B-Instruct",
5
  "bias": "none",
6
  "fan_in_fan_out": false,
7
  "inference_mode": true,
adapter_model.safetensors CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:8bcc52aedbe039c01ff79bfd7db92da7d5f52546e3a255ab15d2976c902084aa
3
- size 27271552
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:0e5ceb216599113955fb046366e338dd7e36152aa88fd5b332d2e6b3c507beb0
3
+ size 73415408
runs/Oct02_08-11-06_4de3b946ecd5/events.out.tfevents.1727856718.4de3b946ecd5.5062.0 ADDED
@@ -0,0 +1,3 @@
 
 
 
 
1
+ version https://git-lfs.github.com/spec/v1
2
+ oid sha256:f2f9c7bb010d5a8e76a438e8bdfc0922f60d5f1478ce0ffe5591a11d14444819
3
+ size 8806
training_args.bin CHANGED
@@ -1,3 +1,3 @@
1
  version https://git-lfs.github.com/spec/v1
2
- oid sha256:f9da11dc689976993c99065fde89e17f2fad8a21f755456177ead23d26628563
3
  size 5432
 
1
  version https://git-lfs.github.com/spec/v1
2
+ oid sha256:4fcf876aad706f024be5e4df0f9caa8b0cc643c4c59d036ce47ee3e9dc7d781b
3
  size 5432