--- license: apache-2.0 language: - en base_model: - Qwen/Qwen2.5-0.5B-Instruct new_version: ShubhaLabs/quantized_qwen_model --- # Quantized Qwen Model This repository contains a quantized version of the Qwen model for causal language modeling. ## Model Details - **Model Type**: Qwen2ForCausalLM - **Quantization**: Dynamic Quantization ## Usage You can load this model using the Hugging Face Transformers library: ```python from transformers import AutoModelForCausalLM, AutoTokenizer model = AutoModelForCausalLM.from_pretrained("ShubhaLabs/quantized_qwen_model") tokenizer = AutoTokenizer.from_pretrained("ShubhaLabs/quantized_qwen_model")