CoreML conversion of Llama-3.2-3B-Instruct with a 512 context length. Optimized for Apple Neural Engine.

Use this CLI to download and run inference. macOS 14 (Sonoma) is required.

This model will likley run slowly or not at all on M1 Macs and phones. Consider trying the 1B model for those devices: smpanaro/Llama-3.2-1B-Instruct-CoreML

Downloads last month

-

Downloads are not tracked for this model. How to track
Inference Providers NEW
This model is not currently available via any of the supported Inference Providers.
The model cannot be deployed to the HF Inference API: The model has no library tag.

Model tree for smpanaro/Llama-3.2-3B-Instruct-CoreML

Finetuned
(246)
this model

Collection including smpanaro/Llama-3.2-3B-Instruct-CoreML