File size: 519 Bytes
610b178 |
1 2 3 4 5 6 7 8 9 10 11 12 13 14 15 16 |
# Quantized Qwen Model
This repository contains a quantized version of the Qwen model for causal language modeling.
## Model Details
- **Model Type**: Qwen2ForCausalLM
- **Quantization**: Dynamic Quantization
## Usage
You can load this model using the Hugging Face Transformers library:
```python
from transformers import AutoModelForCausalLM, AutoTokenizer
model = AutoModelForCausalLM.from_pretrained("ShubhaLabs/quantized_qwen_model")
tokenizer = AutoTokenizer.from_pretrained("ShubhaLabs/quantized_qwen_model") |