--- tags: - quantized - 4-bit - AWQ - autotrain_compatible - endpoints_compatible - text-generation-inference license: apache-2.0 language: - en base_model: mistral-community/Mixtral-8x22B-v0.1 model_creator: Vezora model_name: Mistral-22B-v0.1 model_type: mistral pipeline_tag: text-generation inference: false --- # Vezora/Mistral-22B-v0.1 AWQ - Model creator: [Vezora](https://huggingface.co/Vezora) - Original model: [Mistral-22B-v0.1](https://huggingface.co/Vezora/Mistral-22B-v0.1) ## Model Summary This model is not an moe, it is infact a 22B parameter dense model! Just one day after the release of **Mixtral-8x-22b**, we are excited to introduce our handcrafted experimental model, **Mistral-22b-V.01**. This model is a culmination of equal knowledge distilled from all experts into a single, dense 22b model. This model is not a single trained expert, rather its a compressed MOE model, turning it into a dense 22b mode. This is the first working MOE to Dense model conversion.