supra-nexus-o1-instruct-mlx-4bit

4-bit quantized MLX format of the instruction-tuned Supra Nexus O1 model for efficient inference on Apple Silicon.

πŸ”— Model Collection

Base Models

Available Formats

Instruction Model

Thinking Model

Training Data

πŸ’‘ Key Features

  • Transparent Reasoning: Shows thought process using <thinking> tags
  • Chain-of-Thought: Step-by-step problem solving approach
  • Self-Improvement: Trained with recursive improvement examples
  • Multi-Format: Available in multiple formats for different platforms

πŸš€ Quick Start

Using with MLX (4-bit Quantized)

from mlx_lm import load, generate

# Load 4-bit quantized model (75% smaller)
model, tokenizer = load("Supra-Nexus/supra-nexus-o1-instruct-mlx-4bit")

# Generate with chain-of-thought
prompt = "Solve step by step: What is 25% of 480?"
response = generate(model, tokenizer, prompt=prompt, max_tokens=500)
print(response)

Benefits of 4-bit Quantization

  • πŸš€ 75% smaller model size
  • ⚑ Faster inference on M1/M2/M3 Macs
  • πŸ’Ύ Lower memory requirements
  • ✨ Minimal quality loss

πŸ“ˆ Performance

The O1 models excel at:

  • Complex reasoning tasks
  • Step-by-step problem solving
  • Mathematical computations
  • Code generation and debugging
  • Creative writing with logical structure

πŸ—οΈ Architecture

Based on Qwen2.5 architecture with:

  • Custom fine-tuning for reasoning
  • Chain-of-thought training
  • Self-improvement capabilities
  • Identity preservation techniques

πŸ”¬ Training Details

  • Base Model: Qwen/Qwen2.5-7B-Instruct
  • Training Framework: Zoo Gym
  • Dataset: supra-nexus-o1-training
  • Training Duration: Multiple iterations with self-improvement
  • Hardware: NVIDIA A100 GPUs

πŸ“š Resources

πŸ“„ Citation

@software{supra_nexus_o1_2025,
  title = {Supra Nexus O1: Advanced Reasoning Models},
  author = {Supra Foundation},
  year = {2025},
  url = {https://github.com/Supra-Nexus/o1}
}

πŸ“ License

Apache 2.0 - See LICENSE for details.


Building transparent AI reasoning systems 🧠✨

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model isn't deployed by any Inference Provider. πŸ™‹ Ask for provider support

Model tree for Supra-Nexus/supra-nexus-o1-instruct-mlx-4bit

Finetuned
(3)
this model