Model Card for Model ID
Model Details
Model Description
GPTQ-4bit quantization version (use exllamav2) of https://huggingface.co/T3LS/stella-mrl-large-zh-v3.5-1792d-1024
Uses
model = AutoModelForSequenceClassification.from_pretrained(
'T3LS/stella-mrl-large-zh-v3.5-1792d-1024-gptq-4bit',
device_map='cuda' # Exllamav2 backend requires all the modules to be on GPU
)
- Downloads last month
- 5
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.