|
autoGPTQ QLora finetune sample data and script. |
|
|
|
``` |
|
git lfs install |
|
git clone https://huggingface.co/dahara1/weblab-10b-instruction-sft-GPTQ |
|
cd weblab-10b-instruction-sft-GPTQ/finetune_sample |
|
python3 -m venv gptq_finetune |
|
source gptq_finetune/bin/activate |
|
|
|
|
|
pip install transformers==4.34.1 |
|
pip install datasets |
|
pip install peft==0.5.0 |
|
pip install trl |
|
pip install auto-gptq |
|
pip install optimum |
|
pip install torch==2.0.1 |
|
|
|
# finetune qlora |
|
python3 finetune.py |
|
|
|
# use qlora sample |
|
python3 lora_test.py |
|
``` |
|
|
|
Version is very very important. |
|
For example if you get something like |
|
``` |
|
ValueError: Target module QuantLinear() is not supported. Currently, only `torch.nn.Linear` and `Conv1D` are supported. |
|
``` |
|
It's because peft old version. |
|
|
|
I don't know if it's required, but the version of my running environment. |
|
* auto-gptq 0.4.2 |
|
* trl 0.7.2 |
|
* optimum 1.13.2 |
|
* datasets 2.14.6 |
|
|
|
The official documentation says to install from source, but sometimes that causes errors. |
|
If you can't get it to work, it might be better to wait until the stable version comes out. |
|
Good luck! |
|
|
|
If you encounter ```RuntimeError: Unrecognized tensor type ID: AutocastCUDA```, check your torch version. |
|
auto-gptq 0.4.2 with torch 2.1.0 can't work for me. |
|
|
|
- finetune.py gptq finetune sample file. |
|
- jawiki3.csv sample data.(Japanese) |
|
- lora_test.py after finetune, you can use lora with this script. |
|
- checkpoint-700 Created sample LoRA for test. |
|
|
|
The model.safetensors is ../gptq_model-4bit-128g.safetensors. |
|
It's samefile. I can't find how to change script defaults model name, So I copied it. |