This is a 2.4-bit quantization of Aurelian v0.1alpha 70B 32K for testing & feedback. See that page for more details.

This quantization fits in a single 24GB using Exllamav2 & 8-bit cache @ 10K context. It uses the newer experimental quantization method from turboderp.

Downloads last month
-
Inference Providers NEW
This model isn't deployed by any Inference Provider. ๐Ÿ™‹ Ask for provider support