barneystinson commited on
Commit
356d1b7
1 Parent(s): d770f4b

Upload training_config.yml with huggingface_hub

Browse files
Files changed (1) hide show
  1. training_config.yml +7 -7
training_config.yml CHANGED
@@ -16,10 +16,10 @@ checkpointer:
16
  _component_: torchtune.utils.FullModelMetaCheckpointer
17
  checkpoint_dir: checkpoints/Meta-Llama-3-8B-Instruct/original
18
  checkpoint_files:
19
- - meta_model_0.pt
20
  adapter_checkpoint: null
21
  recipe_checkpoint: null
22
- output_dir: output_checkpoints/experiment_5
23
  model_type: LLAMA3
24
  resume_from_checkpoint: false
25
  interim_checkpoint_steps: 15000
@@ -49,16 +49,16 @@ shuffle: true
49
  batch_size: 6
50
  optimizer:
51
  _component_: torch.optim.AdamW
52
- weight_decay: 0.008
53
- lr: 0.0001
54
  lr_scheduler:
55
  _component_: torchtune.modules.get_cosine_schedule_with_warmup
56
- num_warmup_steps: 100
57
  loss:
58
  _component_: torch.nn.CrossEntropyLoss
59
- epochs: 6
60
  max_steps_per_epoch: null
61
- gradient_accumulation_steps: 16
62
  compile: false
63
  output_dir: /tmp/lora_finetune_output
64
  metric_logger:
 
16
  _component_: torchtune.utils.FullModelMetaCheckpointer
17
  checkpoint_dir: checkpoints/Meta-Llama-3-8B-Instruct/original
18
  checkpoint_files:
19
+ - meta_model_1.pt
20
  adapter_checkpoint: null
21
  recipe_checkpoint: null
22
+ output_dir: output_checkpoints/experiment_6
23
  model_type: LLAMA3
24
  resume_from_checkpoint: false
25
  interim_checkpoint_steps: 15000
 
49
  batch_size: 6
50
  optimizer:
51
  _component_: torch.optim.AdamW
52
+ weight_decay: 0.99
53
+ lr: 2.0e-05
54
  lr_scheduler:
55
  _component_: torchtune.modules.get_cosine_schedule_with_warmup
56
+ num_warmup_steps: 4
57
  loss:
58
  _component_: torch.nn.CrossEntropyLoss
59
+ epochs: 60
60
  max_steps_per_epoch: null
61
+ gradient_accumulation_steps: 260
62
  compile: false
63
  output_dir: /tmp/lora_finetune_output
64
  metric_logger: