allow report_to for multiple providers (#1647)
Browse files
src/axolotl/core/trainer_builder.py
CHANGED
@@ -1212,11 +1212,14 @@ class HFCausalTrainerBuilder(TrainerBuilderBase):
|
|
1212 |
)
|
1213 |
training_arguments_kwargs["group_by_length"] = self.cfg.group_by_length
|
1214 |
training_arguments_kwargs["curriculum_sampling"] = self.cfg.curriculum_sampling
|
1215 |
-
report_to =
|
1216 |
if self.cfg.use_wandb:
|
1217 |
-
report_to
|
1218 |
if self.cfg.use_mlflow:
|
1219 |
-
report_to
|
|
|
|
|
|
|
1220 |
training_arguments_kwargs["report_to"] = report_to
|
1221 |
training_arguments_kwargs["run_name"] = (
|
1222 |
self.cfg.wandb_name if self.cfg.use_wandb else None
|
|
|
1212 |
)
|
1213 |
training_arguments_kwargs["group_by_length"] = self.cfg.group_by_length
|
1214 |
training_arguments_kwargs["curriculum_sampling"] = self.cfg.curriculum_sampling
|
1215 |
+
report_to = []
|
1216 |
if self.cfg.use_wandb:
|
1217 |
+
report_to.append("wandb")
|
1218 |
if self.cfg.use_mlflow:
|
1219 |
+
report_to.append("mlflow")
|
1220 |
+
if self.cfg.use_tensorboard:
|
1221 |
+
report_to.append("tensorboard")
|
1222 |
+
|
1223 |
training_arguments_kwargs["report_to"] = report_to
|
1224 |
training_arguments_kwargs["run_name"] = (
|
1225 |
self.cfg.wandb_name if self.cfg.use_wandb else None
|
src/axolotl/utils/config/models/input/v0_4_1/__init__.py
CHANGED
@@ -212,7 +212,7 @@ class LoraConfig(BaseModel):
|
|
212 |
lora_target_modules: Optional[List[str]] = None
|
213 |
lora_target_linear: Optional[bool] = None
|
214 |
lora_modules_to_save: Optional[List[str]] = None
|
215 |
-
lora_dropout: Optional[float] =
|
216 |
peft_layers_to_transform: Optional[List[int]] = None
|
217 |
peft: Optional[PeftConfig] = None
|
218 |
peft_use_dora: Optional[bool] = None
|
@@ -609,6 +609,7 @@ class AxolotlInputConfig(
|
|
609 |
early_stopping_patience: Optional[int] = None
|
610 |
load_best_model_at_end: Optional[bool] = False
|
611 |
save_only_model: Optional[bool] = False
|
|
|
612 |
|
613 |
neftune_noise_alpha: Optional[float] = None
|
614 |
|
|
|
212 |
lora_target_modules: Optional[List[str]] = None
|
213 |
lora_target_linear: Optional[bool] = None
|
214 |
lora_modules_to_save: Optional[List[str]] = None
|
215 |
+
lora_dropout: Optional[float] = 0.0
|
216 |
peft_layers_to_transform: Optional[List[int]] = None
|
217 |
peft: Optional[PeftConfig] = None
|
218 |
peft_use_dora: Optional[bool] = None
|
|
|
609 |
early_stopping_patience: Optional[int] = None
|
610 |
load_best_model_at_end: Optional[bool] = False
|
611 |
save_only_model: Optional[bool] = False
|
612 |
+
use_tensorboard: Optional[bool] = None
|
613 |
|
614 |
neftune_noise_alpha: Optional[float] = None
|
615 |
|