Update README.md
Browse files
README.md
CHANGED
@@ -20,7 +20,7 @@ Command to create was:
|
|
20 |
cd gptq-safe && CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
|
21 |
```
|
22 |
|
23 |
-
Note that
|
24 |
|
25 |
Command to clone the correct GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
|
26 |
```
|
|
|
20 |
cd gptq-safe && CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
|
21 |
```
|
22 |
|
23 |
+
Note that as `--act-order` was used, this will not work with ooba's fork of GPTQ. You must use the qwopqwop repo as of April 13th.
|
24 |
|
25 |
Command to clone the correct GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
|
26 |
```
|