ngxson HF staff commited on
Commit
ae0dafb
·
1 Parent(s): f11fc50
Files changed (1) hide show
  1. app.py +1 -1
app.py CHANGED
@@ -120,7 +120,7 @@ def process_model(model_id, q_method, use_imatrix, imatrix_q_method, private_rep
120
  config_dir = local_dir/"config.json"
121
  adapter_config_dir = local_dir/"adapter_config.json"
122
  if os.path.exists(adapter_config_dir) and not os.path.exists(config_dir):
123
- raise Exception("adapter_config.json is present.<br/><br/>If you are converting a LoRA adapter to GGUF, please use <a href=\"https://huggingface.co/spaces/ggml-org/gguf-my-lora\" target=\"_blank\">GGUF-my-lora</a>.")
124
 
125
  conversion_script = "convert_hf_to_gguf.py"
126
  fp16_conversion = f"python llama.cpp/{conversion_script} {local_dir} --outtype f16 --outfile {fp16}"
 
120
  config_dir = local_dir/"config.json"
121
  adapter_config_dir = local_dir/"adapter_config.json"
122
  if os.path.exists(adapter_config_dir) and not os.path.exists(config_dir):
123
+ raise Exception('adapter_config.json is present.<br/><br/>If you are converting a LoRA adapter to GGUF, please use <a href="https://huggingface.co/spaces/ggml-org/gguf-my-lora" target="_blank" style="text-decoration:underline">GGUF-my-lora</a>.')
124
 
125
  conversion_script = "convert_hf_to_gguf.py"
126
  fp16_conversion = f"python llama.cpp/{conversion_script} {local_dir} --outtype f16 --outfile {fp16}"