winglian commited on
Commit
6299eb5
·
unverified ·
1 Parent(s): 7c2bf30

allow report_to for multiple providers (#1647)

Browse files
src/axolotl/core/trainer_builder.py CHANGED
@@ -1212,11 +1212,14 @@ class HFCausalTrainerBuilder(TrainerBuilderBase):
1212
  )
1213
  training_arguments_kwargs["group_by_length"] = self.cfg.group_by_length
1214
  training_arguments_kwargs["curriculum_sampling"] = self.cfg.curriculum_sampling
1215
- report_to = None
1216
  if self.cfg.use_wandb:
1217
- report_to = "wandb"
1218
  if self.cfg.use_mlflow:
1219
- report_to = "mlflow"
 
 
 
1220
  training_arguments_kwargs["report_to"] = report_to
1221
  training_arguments_kwargs["run_name"] = (
1222
  self.cfg.wandb_name if self.cfg.use_wandb else None
 
1212
  )
1213
  training_arguments_kwargs["group_by_length"] = self.cfg.group_by_length
1214
  training_arguments_kwargs["curriculum_sampling"] = self.cfg.curriculum_sampling
1215
+ report_to = []
1216
  if self.cfg.use_wandb:
1217
+ report_to.append("wandb")
1218
  if self.cfg.use_mlflow:
1219
+ report_to.append("mlflow")
1220
+ if self.cfg.use_tensorboard:
1221
+ report_to.append("tensorboard")
1222
+
1223
  training_arguments_kwargs["report_to"] = report_to
1224
  training_arguments_kwargs["run_name"] = (
1225
  self.cfg.wandb_name if self.cfg.use_wandb else None
src/axolotl/utils/config/models/input/v0_4_1/__init__.py CHANGED
@@ -212,7 +212,7 @@ class LoraConfig(BaseModel):
212
  lora_target_modules: Optional[List[str]] = None
213
  lora_target_linear: Optional[bool] = None
214
  lora_modules_to_save: Optional[List[str]] = None
215
- lora_dropout: Optional[float] = None
216
  peft_layers_to_transform: Optional[List[int]] = None
217
  peft: Optional[PeftConfig] = None
218
  peft_use_dora: Optional[bool] = None
@@ -609,6 +609,7 @@ class AxolotlInputConfig(
609
  early_stopping_patience: Optional[int] = None
610
  load_best_model_at_end: Optional[bool] = False
611
  save_only_model: Optional[bool] = False
 
612
 
613
  neftune_noise_alpha: Optional[float] = None
614
 
 
212
  lora_target_modules: Optional[List[str]] = None
213
  lora_target_linear: Optional[bool] = None
214
  lora_modules_to_save: Optional[List[str]] = None
215
+ lora_dropout: Optional[float] = 0.0
216
  peft_layers_to_transform: Optional[List[int]] = None
217
  peft: Optional[PeftConfig] = None
218
  peft_use_dora: Optional[bool] = None
 
609
  early_stopping_patience: Optional[int] = None
610
  load_best_model_at_end: Optional[bool] = False
611
  save_only_model: Optional[bool] = False
612
+ use_tensorboard: Optional[bool] = None
613
 
614
  neftune_noise_alpha: Optional[float] = None
615