Edit model card

Model Card for Mistral-Large-Instruct-2411-MLX

This repository serves as the parent directory for the MLX quantized versions of the Mistral Large Instruct 2411 model. The quantized versions were created for MLX (Apple Silicon) using the mlx-lm library.

Quantized Versions

Each version is optimized for specific memory and performance trade-offs. See the individual repositories for details on the quantization methods.

Original Model

The original Mistral-Large-Instruct-2411 model is available here. Mistral model usage is governed by the Mistral Research License.

License

This model family is governed by the Mistral Research License. Please review the license terms before use.

Table of Contents

Model Details

Model Description

The Mistral-Large-Instruct-2411-MLX family includes quantized versions of the Mistral Large Instruct 2411 model, optimized for deployment on MLX (Apple Silicon). The quantization reduces memory usage and inference latency, enabling efficient deployment on resource-constrained systems.

  • Developed by: Mistral AI
  • Model type: Large language model
  • Language(s): English, French, German, Spanish, Italian, Portuguese, Chinese, Japanese, Russian, Korean
  • Quantization levels: 2-bit (Q2), 4-bit (Q4)

Technical Specifications

How to Get Started

Visit the individual quantized repositories for details and usage instructions:

Model Card Contact

For inquiries, contact Zach Landes.

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference API
Unable to determine this model’s pipeline type. Check the docs .

Model tree for zachlandes/Mistral-Large-Instruct-2411-MLX

Finetuned
(3)
this model