Miquliz-120b-v2.0-FP8-dynamic
This quant was made for infermatic.ai
Dynamic FP8 quant of Miquliz 120B v2.0 made with AutoFP8.
Model Details
- Max Context: 32768 tokens
- Layers: 140
Prompt template: Mistral
<s>[INST] {prompt} [/INST]
- Downloads last month
- 13
This model does not have enough activity to be deployed to Inference API (serverless) yet. Increase its social
visibility and check back later, or deploy to Inference Endpoints (dedicated)
instead.