Example:

from transformers import AutoModelForCausalLM, AutoTokenizer, GenerationConfig

tokenizer = AutoTokenizer.from_pretrained("Qwen-1_8B-m4-LDJnr-combined", trust_remote_code=True)
model = AutoModelForCausalLM.from_pretrained("Qwen-1_8B-m4-LDJnr-combined", device_map="auto", trust_remote_code=True).eval()
response, _ = model.chat(tokenizer, "What kind of a noise annoys a noisy oyster?", history=None)
print(response)
Downloads last month
61
Safetensors
Model size
1.84B params
Tensor type
FP16
·
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The HF Inference API does not support model that require custom code execution.

Collection including aloobun/Qwen-1_8B-m4-LDJnr-combined