File size: 519 Bytes
610b178
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
# Quantized Qwen Model

This repository contains a quantized version of the Qwen model for causal language modeling.

## Model Details
- **Model Type**: Qwen2ForCausalLM
- **Quantization**: Dynamic Quantization

## Usage
You can load this model using the Hugging Face Transformers library:

```python
from transformers import AutoModelForCausalLM, AutoTokenizer

model = AutoModelForCausalLM.from_pretrained("ShubhaLabs/quantized_qwen_model")
tokenizer = AutoTokenizer.from_pretrained("ShubhaLabs/quantized_qwen_model")