Edit model card

Vezora/Mistral-22B-v0.1 AWQ

Model Summary

This model is not an moe, it is infact a 22B parameter dense model!

Just one day after the release of Mixtral-8x-22b, we are excited to introduce our handcrafted experimental model, Mistral-22b-V.01. This model is a culmination of equal knowledge distilled from all experts into a single, dense 22b model. This model is not a single trained expert, rather its a compressed MOE model, turning it into a dense 22b mode. This is the first working MOE to Dense model conversion.

How to use

GUANACO PROMPT FORMAT YOU MUST USE THE GUANACO PROMPT FORMAT SHOWN BELOW. Not using this prompt format will lead to sub optimal results.

  • This model requires a specific chat template, as the training format was Guanaco this is what it looks like:
  • "### System: You are a helpful assistant. ### Human###: Give me the best chili recipe you can ###Assistant: Here is the best chili recipe..."
Downloads last month
18
Safetensors
Model size
3.32B params
Tensor type
I32
·
FP16
·
Inference Examples
Inference API (serverless) has been turned off for this model.

Model tree for solidrust/Mistral-22B-v0.1-AWQ

Quantized
(2)
this model

Collection including solidrust/Mistral-22B-v0.1-AWQ