TheBloke commited on
Commit
39409e1
1 Parent(s): 7aa76f9

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -20,7 +20,7 @@ Command to create was:
20
  cd gptq-safe && CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
21
  ```
22
 
23
- Note that only as `--act-order` was used, this will not work with ooba's fork of GPTQ. You must use the qwopqwop repo as of April 13th.
24
 
25
  Command to clone the correct GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
26
  ```
 
20
  cd gptq-safe && CUDA_VISIBLE_DEVICES=0 python3 llama.py /content/gpt4-alpaca-lora-13B-HF c4 --wbits 4 --true-sequential --act-order --groupsize 128 --save_safetensors /content/gpt4-alpaca-lora-13B-GPTQ-4bit-128g.safetensors
21
  ```
22
 
23
+ Note that as `--act-order` was used, this will not work with ooba's fork of GPTQ. You must use the qwopqwop repo as of April 13th.
24
 
25
  Command to clone the correct GPTQ-for-LLaMa repo for inference using `llama_inference.py`, or in `text-generation-webui`:
26
  ```