add new config
Browse files- configs/minotaur.yml +9 -6
configs/minotaur.yml
CHANGED
@@ -83,6 +83,10 @@ datasets:
|
|
83 |
type: alpaca_chat
|
84 |
- path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
|
85 |
type: alpaca:chat
|
|
|
|
|
|
|
|
|
86 |
dataset_prepared_path: last_run_prepared
|
87 |
val_set_size: 0.01
|
88 |
adapter:
|
@@ -106,9 +110,9 @@ num_epochs: 12
|
|
106 |
optimizer: adamw_bnb_8bit
|
107 |
torchdistx_path:
|
108 |
lr_scheduler: cosine
|
109 |
-
learning_rate: 0.
|
110 |
train_on_inputs: false
|
111 |
-
group_by_length:
|
112 |
bf16: true
|
113 |
fp16: false
|
114 |
tf32: true
|
@@ -122,13 +126,12 @@ flash_attention:
|
|
122 |
gptq_groupsize:
|
123 |
gptq_model_v1:
|
124 |
warmup_steps: 100
|
125 |
-
eval_steps:
|
126 |
-
save_steps:
|
127 |
load_best_model_at_end: false
|
128 |
-
save_total_limit: 10
|
129 |
debug:
|
130 |
deepspeed:
|
131 |
-
weight_decay: 0.
|
132 |
fsdp:
|
133 |
- full_shard
|
134 |
- auto_wrap
|
|
|
83 |
type: alpaca_chat
|
84 |
- path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
|
85 |
type: alpaca:chat
|
86 |
+
- path: ewof/code-alpaca-instruct-unfiltered
|
87 |
+
type: alpaca_chat
|
88 |
+
- path: teknium/GPT4-LLM-Cleaned
|
89 |
+
type: alpaca_chat
|
90 |
dataset_prepared_path: last_run_prepared
|
91 |
val_set_size: 0.01
|
92 |
adapter:
|
|
|
110 |
optimizer: adamw_bnb_8bit
|
111 |
torchdistx_path:
|
112 |
lr_scheduler: cosine
|
113 |
+
learning_rate: 0.00005
|
114 |
train_on_inputs: false
|
115 |
+
group_by_length: true
|
116 |
bf16: true
|
117 |
fp16: false
|
118 |
tf32: true
|
|
|
126 |
gptq_groupsize:
|
127 |
gptq_model_v1:
|
128 |
warmup_steps: 100
|
129 |
+
eval_steps: 29
|
130 |
+
save_steps: 29
|
131 |
load_best_model_at_end: false
|
|
|
132 |
debug:
|
133 |
deepspeed:
|
134 |
+
weight_decay: 0.0007
|
135 |
fsdp:
|
136 |
- full_shard
|
137 |
- auto_wrap
|