winglian commited on
Commit
18d1b3b
1 Parent(s): 4344a03

add new config

Browse files
Files changed (1) hide show
  1. configs/minotaur.yml +9 -6
configs/minotaur.yml CHANGED
@@ -83,6 +83,10 @@ datasets:
83
  type: alpaca_chat
84
  - path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
85
  type: alpaca:chat
 
 
 
 
86
  dataset_prepared_path: last_run_prepared
87
  val_set_size: 0.01
88
  adapter:
@@ -106,9 +110,9 @@ num_epochs: 12
106
  optimizer: adamw_bnb_8bit
107
  torchdistx_path:
108
  lr_scheduler: cosine
109
- learning_rate: 0.00001
110
  train_on_inputs: false
111
- group_by_length: false
112
  bf16: true
113
  fp16: false
114
  tf32: true
@@ -122,13 +126,12 @@ flash_attention:
122
  gptq_groupsize:
123
  gptq_model_v1:
124
  warmup_steps: 100
125
- eval_steps: 27
126
- save_steps: 27
127
  load_best_model_at_end: false
128
- save_total_limit: 10
129
  debug:
130
  deepspeed:
131
- weight_decay: 0.00001
132
  fsdp:
133
  - full_shard
134
  - auto_wrap
 
83
  type: alpaca_chat
84
  - path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
85
  type: alpaca:chat
86
+ - path: ewof/code-alpaca-instruct-unfiltered
87
+ type: alpaca_chat
88
+ - path: teknium/GPT4-LLM-Cleaned
89
+ type: alpaca_chat
90
  dataset_prepared_path: last_run_prepared
91
  val_set_size: 0.01
92
  adapter:
 
110
  optimizer: adamw_bnb_8bit
111
  torchdistx_path:
112
  lr_scheduler: cosine
113
+ learning_rate: 0.00005
114
  train_on_inputs: false
115
+ group_by_length: true
116
  bf16: true
117
  fp16: false
118
  tf32: true
 
126
  gptq_groupsize:
127
  gptq_model_v1:
128
  warmup_steps: 100
129
+ eval_steps: 29
130
+ save_steps: 29
131
  load_best_model_at_end: false
 
132
  debug:
133
  deepspeed:
134
+ weight_decay: 0.0007
135
  fsdp:
136
  - full_shard
137
  - auto_wrap