|
--- |
|
language: |
|
- zh |
|
license: apache-2.0 |
|
tasks: |
|
- text-generation |
|
--- |
|
|
|
<!-- markdownlint-disable first-line-h1 --> |
|
<!-- markdownlint-disable html --> |
|
<div align="center"> |
|
<h1> |
|
HuatuoGPT2-13B |
|
</h1> |
|
</div> |
|
|
|
<div align="center"> |
|
<a href="https://github.com/FreedomIntelligence/HuatuoGPT-II" target="_blank">GitHub</a> | <a href="https://arxiv.org/pdf/2311.09774.pdf" target="_blank">Our Paper</a> |
|
</div> |
|
|
|
# <span id="Start">Quick Start</span> |
|
|
|
```Python |
|
import torch |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
from transformers.generation.utils import GenerationConfig |
|
tokenizer = AutoTokenizer.from_pretrained("FreedomIntelligence/HuatuoGPT2-13B", use_fast=True, trust_remote_code=True) |
|
model = AutoModelForCausalLM.from_pretrained("FreedomIntelligence/HuatuoGPT2-13B", device_map="auto", torch_dtype='auto', trust_remote_code=True) |
|
model.generation_config = GenerationConfig.from_pretrained("FreedomIntelligence/HuatuoGPT2-13B") |
|
messages = [] |
|
messages.append({"role": "user", "content": "肚子疼怎么办?"}) |
|
response = model.HuatuoChat(tokenizer, messages) |
|
print(response) |
|
``` |