Update README.md
Browse files
README.md
CHANGED
@@ -2,7 +2,7 @@
|
|
2 |
license: other
|
3 |
license_name: tongyi-qianwen-research
|
4 |
license_link: >-
|
5 |
-
https://huggingface.co/Qwen/Qwen1.5-
|
6 |
language:
|
7 |
- en
|
8 |
pipeline_tag: text-generation
|
@@ -47,10 +47,10 @@ from transformers import AutoModelForCausalLM, AutoTokenizer
|
|
47 |
device = "cuda" # the device to load the model onto
|
48 |
|
49 |
model = AutoModelForCausalLM.from_pretrained(
|
50 |
-
"Qwen/Qwen1.5-
|
51 |
device_map="auto"
|
52 |
)
|
53 |
-
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen1.5-
|
54 |
|
55 |
prompt = "Give me a short introduction to large language model."
|
56 |
messages = [
|
|
|
2 |
license: other
|
3 |
license_name: tongyi-qianwen-research
|
4 |
license_link: >-
|
5 |
+
https://huggingface.co/Qwen/Qwen1.5-0.5B-Chat-GPTQ-Int8/blob/main/LICENSE
|
6 |
language:
|
7 |
- en
|
8 |
pipeline_tag: text-generation
|
|
|
47 |
device = "cuda" # the device to load the model onto
|
48 |
|
49 |
model = AutoModelForCausalLM.from_pretrained(
|
50 |
+
"Qwen/Qwen1.5-0.5B-Chat-GPTQ-Int8",
|
51 |
device_map="auto"
|
52 |
)
|
53 |
+
tokenizer = AutoTokenizer.from_pretrained("Qwen/Qwen1.5-0.5B-Chat-GPTQ-Int8")
|
54 |
|
55 |
prompt = "Give me a short introduction to large language model."
|
56 |
messages = [
|