release v1 epoch 1
Browse files- configs/minotaur.yml +5 -8
- pytorch_model.bin +1 -1
configs/minotaur.yml
CHANGED
@@ -83,10 +83,6 @@ datasets:
|
|
83 |
type: alpaca_chat
|
84 |
- path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
|
85 |
type: alpaca:chat
|
86 |
-
- path: ewof/code-alpaca-instruct-unfiltered
|
87 |
-
type: alpaca_chat
|
88 |
-
- path: teknium/GPT4-LLM-Cleaned
|
89 |
-
type: alpaca_chat
|
90 |
dataset_prepared_path: last_run_prepared
|
91 |
val_set_size: 0.01
|
92 |
adapter:
|
@@ -105,8 +101,8 @@ wandb_run_id:
|
|
105 |
wandb_log_model:
|
106 |
output_dir: ./minotaur-13b
|
107 |
gradient_accumulation_steps: 1
|
108 |
-
micro_batch_size:
|
109 |
-
num_epochs:
|
110 |
optimizer: adamw_bnb_8bit
|
111 |
torchdistx_path:
|
112 |
lr_scheduler: cosine
|
@@ -125,13 +121,13 @@ xformers_attention: true
|
|
125 |
flash_attention:
|
126 |
gptq_groupsize:
|
127 |
gptq_model_v1:
|
128 |
-
warmup_steps:
|
129 |
eval_steps: 29
|
130 |
save_steps: 29
|
131 |
load_best_model_at_end: false
|
132 |
debug:
|
133 |
deepspeed:
|
134 |
-
weight_decay: 0.
|
135 |
fsdp:
|
136 |
- full_shard
|
137 |
- auto_wrap
|
@@ -143,3 +139,4 @@ special_tokens:
|
|
143 |
eos_token: "</s>"
|
144 |
unk_token: "<unk>"
|
145 |
|
|
|
|
83 |
type: alpaca_chat
|
84 |
- path: ehartford/WizardLM_alpaca_evol_instruct_70k_unfiltered
|
85 |
type: alpaca:chat
|
|
|
|
|
|
|
|
|
86 |
dataset_prepared_path: last_run_prepared
|
87 |
val_set_size: 0.01
|
88 |
adapter:
|
|
|
101 |
wandb_log_model:
|
102 |
output_dir: ./minotaur-13b
|
103 |
gradient_accumulation_steps: 1
|
104 |
+
micro_batch_size: 8
|
105 |
+
num_epochs: 16
|
106 |
optimizer: adamw_bnb_8bit
|
107 |
torchdistx_path:
|
108 |
lr_scheduler: cosine
|
|
|
121 |
flash_attention:
|
122 |
gptq_groupsize:
|
123 |
gptq_model_v1:
|
124 |
+
warmup_steps: 300
|
125 |
eval_steps: 29
|
126 |
save_steps: 29
|
127 |
load_best_model_at_end: false
|
128 |
debug:
|
129 |
deepspeed:
|
130 |
+
weight_decay: 0.00003
|
131 |
fsdp:
|
132 |
- full_shard
|
133 |
- auto_wrap
|
|
|
139 |
eos_token: "</s>"
|
140 |
unk_token: "<unk>"
|
141 |
|
142 |
+
|
pytorch_model.bin
CHANGED
@@ -1,3 +1,3 @@
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
-
oid sha256:
|
3 |
size 26031937261
|
|
|
1 |
version https://git-lfs.github.com/spec/v1
|
2 |
+
oid sha256:8faa183c31365889b8f5de5f719c6a463e82833d2a49063f25c5c24755a4bcaf
|
3 |
size 26031937261
|