Update README.md
Browse files
README.md
CHANGED
@@ -12,6 +12,7 @@ The size is smaller and the execution speed is faster, but the inference perform
|
|
12 |
|
13 |
### sample code
|
14 |
At least one GPU is currently required due to a limitation of the Accelerate library.
|
|
|
15 |
|
16 |
```
|
17 |
pip install auto-gptq
|
@@ -40,13 +41,16 @@ output = model.generate(input_ids=tokens, max_new_tokens=100, do_sample=True, te
|
|
40 |
print(tokenizer.decode(output[0]))
|
41 |
```
|
42 |
|
43 |
-
###
|
44 |
https://github.com/PanQiWei/AutoGPTQ/blob/main/docs/tutorial/01-Quick-Start.md
|
45 |
|
|
|
|
|
46 |
|
47 |
### Benchmark
|
48 |
|
49 |
-
The results below are preliminary. The blank part is under measurement.
|
|
|
50 |
|
51 |
* **Japanese benchmark**
|
52 |
|
|
|
12 |
|
13 |
### sample code
|
14 |
At least one GPU is currently required due to a limitation of the Accelerate library.
|
15 |
+
So this model cannot be run with the huggingface space free version.
|
16 |
|
17 |
```
|
18 |
pip install auto-gptq
|
|
|
41 |
print(tokenizer.decode(output[0]))
|
42 |
```
|
43 |
|
44 |
+
### Other documents
|
45 |
https://github.com/PanQiWei/AutoGPTQ/blob/main/docs/tutorial/01-Quick-Start.md
|
46 |
|
47 |
+
### Original Authors
|
48 |
+
Takeshi Kojima
|
49 |
|
50 |
### Benchmark
|
51 |
|
52 |
+
The results below are preliminary. The blank part is under measurement.
|
53 |
+
Also, the score may change as a result of tuning after this.
|
54 |
|
55 |
* **Japanese benchmark**
|
56 |
|