# Quantized Qwen Model This repository contains a quantized version of the Qwen model for causal language modeling. ## Model Details - **Model Type**: Qwen2ForCausalLM - **Quantization**: Dynamic Quantization ## Usage You can load this model using the Hugging Face Transformers library: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("ShubhaLabs/quantized_qwen_model") tokenizer = AutoTokenizer.from_pretrained("ShubhaLabs/quantized_qwen_model")