supra-nexus-o1-instruct-mlx-4bit
4-bit quantized MLX format of the instruction-tuned Supra Nexus O1 model for efficient inference on Apple Silicon.
π Model Collection
Base Models
- π€ supra-nexus-o1-instruct - Instruction-following model
- π supra-nexus-o1-thinking - Chain-of-thought reasoning model
Available Formats
Instruction Model
Thinking Model
Training Data
- π supra-nexus-o1-training - Complete training dataset
π‘ Key Features
- Transparent Reasoning: Shows thought process using
<thinking>tags - Chain-of-Thought: Step-by-step problem solving approach
- Self-Improvement: Trained with recursive improvement examples
- Multi-Format: Available in multiple formats for different platforms
π Quick Start
Using with MLX (4-bit Quantized)
from mlx_lm import load, generate
# Load 4-bit quantized model (75% smaller)
model, tokenizer = load("Supra-Nexus/supra-nexus-o1-instruct-mlx-4bit")
# Generate with chain-of-thought
prompt = "Solve step by step: What is 25% of 480?"
response = generate(model, tokenizer, prompt=prompt, max_tokens=500)
print(response)
Benefits of 4-bit Quantization
- π 75% smaller model size
- β‘ Faster inference on M1/M2/M3 Macs
- πΎ Lower memory requirements
- β¨ Minimal quality loss
π Performance
The O1 models excel at:
- Complex reasoning tasks
- Step-by-step problem solving
- Mathematical computations
- Code generation and debugging
- Creative writing with logical structure
ποΈ Architecture
Based on Qwen2.5 architecture with:
- Custom fine-tuning for reasoning
- Chain-of-thought training
- Self-improvement capabilities
- Identity preservation techniques
π¬ Training Details
- Base Model: Qwen/Qwen2.5-7B-Instruct
- Training Framework: Zoo Gym
- Dataset: supra-nexus-o1-training
- Training Duration: Multiple iterations with self-improvement
- Hardware: NVIDIA A100 GPUs
π Resources
- π GitHub Repository - Source code and documentation
- π’ Supra Foundation - Organization behind O1
- π¦ Twitter - Latest updates
- π¬ Discord - Community support
π Citation
@software{supra_nexus_o1_2025,
title = {Supra Nexus O1: Advanced Reasoning Models},
author = {Supra Foundation},
year = {2025},
url = {https://github.com/Supra-Nexus/o1}
}
π License
Apache 2.0 - See LICENSE for details.
Building transparent AI reasoning systems π§ β¨
Inference Providers
NEW
This model isn't deployed by any Inference Provider.
π
Ask for provider support
Model tree for Supra-Nexus/supra-nexus-o1-instruct-mlx-4bit
Base model
Supra-Nexus/supra-nexus-o1-instruct