xhedit commited on
Commit
8608d80
1 Parent(s): 4cb7900

Fix typo (#1231) [skip ci]

Browse files
Files changed (1) hide show
  1. src/axolotl/utils/models.py +2 -2
src/axolotl/utils/models.py CHANGED
@@ -756,7 +756,7 @@ def load_llama_adapter(model, cfg):
756
  )
757
 
758
  if cfg.lora_model_dir:
759
- LOG.debug("Loading pretained PEFT - llama_adapter")
760
  model = PeftModel.from_pretrained(
761
  model,
762
  cfg.lora_model_dir,
@@ -825,7 +825,7 @@ def load_lora(model, cfg, inference=False, config_only=False):
825
  return None, lora_config
826
 
827
  if cfg.lora_model_dir:
828
- LOG.debug("Loading pretained PEFT - LoRA")
829
  model_kwargs: Any = {}
830
  if cfg.lora_on_cpu:
831
  model_kwargs["max_memory"] = {"cpu": "256GiB"}
 
756
  )
757
 
758
  if cfg.lora_model_dir:
759
+ LOG.debug("Loading pretrained PEFT - llama_adapter")
760
  model = PeftModel.from_pretrained(
761
  model,
762
  cfg.lora_model_dir,
 
825
  return None, lora_config
826
 
827
  if cfg.lora_model_dir:
828
+ LOG.debug("Loading pretrained PEFT - LoRA")
829
  model_kwargs: Any = {}
830
  if cfg.lora_on_cpu:
831
  model_kwargs["max_memory"] = {"cpu": "256GiB"}