Spaces:
Runtime error
Runtime error
# export-lora | |
Apply LORA adapters to base model and export the resulting model. | |
``` | |
usage: llama-export-lora [options] | |
options: | |
-m, --model model path from which to load base model (default '') | |
--lora FNAME path to LoRA adapter (can be repeated to use multiple adapters) | |
--lora-scaled FNAME S path to LoRA adapter with user defined scaling S (can be repeated to use multiple adapters) | |
-t, --threads N number of threads to use during computation (default: 4) | |
-o, --output FNAME output file (default: 'ggml-lora-merged-f16.gguf') | |
``` | |
For example: | |
```bash | |
./bin/llama-export-lora \ | |
-m open-llama-3b-v2.gguf \ | |
-o open-llama-3b-v2-english2tokipona-chat.gguf \ | |
--lora lora-open-llama-3b-v2-english2tokipona-chat-LATEST.gguf | |
``` | |
Multiple LORA adapters can be applied by passing multiple `--lora FNAME` or `--lora-scaled FNAME S` command line parameters: | |
```bash | |
./bin/llama-export-lora \ | |
-m your_base_model.gguf \ | |
-o your_merged_model.gguf \ | |
--lora-scaled lora_task_A.gguf 0.5 \ | |
--lora-scaled lora_task_B.gguf 0.5 | |
``` | |