|
# GLM-Edge-4B-Chat |
|
|
|
## 使用 transformers 库进行推理 |
|
|
|
### 安装 |
|
|
|
请安装源代码的transformers库。 |
|
|
|
```shell |
|
pip install git+https://github.com/huggingface/transformers.git |
|
``` |
|
### 推理 |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
MODEL_PATH = "THUDM/glm-edge-4b-chat" |
|
|
|
tokenizer = AutoTokenizer.from_pretrained(MODEL_PATH) |
|
model = AutoModelForCausalLM.from_pretrained(MODEL_PATH, device_map="auto") |
|
|
|
message = [{"role": "user", "content": "hello!"}] |
|
|
|
inputs = tokenizer.apply_chat_template( |
|
message, |
|
return_tensors="pt", |
|
add_generation_prompt=True, |
|
return_dict=True, |
|
).to(model.device) |
|
|
|
generate_kwargs = { |
|
"input_ids": inputs["input_ids"], |
|
"attention_mask": inputs["attention_mask"], |
|
"max_new_tokens": 128, |
|
"do_sample": False, |
|
} |
|
out = model.generate(**generate_kwargs) |
|
print(tokenizer.decode(out[0][inputs["input_ids"].shape[1]:], skip_special_tokens=True)) |
|
|
|
``` |
|
|
|
## 协议 |
|
|
|
本模型的权重的使用则需要遵循 [LICENSE](LICENSE)。 |