|
# Quantized Qwen Model |
|
|
|
This repository contains a quantized version of the Qwen model for causal language modeling. |
|
|
|
## Model Details |
|
- **Model Type**: Qwen2ForCausalLM |
|
- **Quantization**: Dynamic Quantization |
|
|
|
## Usage |
|
You can load this model using the Hugging Face Transformers library: |
|
|
|
```python |
|
from transformers import AutoModelForCausalLM, AutoTokenizer |
|
|
|
model = AutoModelForCausalLM.from_pretrained("ShubhaLabs/quantized_qwen_model") |
|
tokenizer = AutoTokenizer.from_pretrained("ShubhaLabs/quantized_qwen_model") |