Granite 4.0 H-Small (GGUF)

This repository contains models that have been converted to the GGUF format with various quantizations from an IBM Granite base model.

Please reference the base model's full model card here: https://huggingface.co/ibm-granite/granite-4.0-h-small

Downloads last month
6,953
GGUF
Model size
32B params
Architecture
granitehybrid
Hardware compatibility
Log In to view the estimation

2-bit

3-bit

4-bit

5-bit

6-bit

8-bit

16-bit

Inference Providers NEW
This model isn't deployed by any Inference Provider. 🙋 Ask for provider support

Model tree for ibm-granite/granite-4.0-h-small-GGUF

Quantized
(25)
this model

Collection including ibm-granite/granite-4.0-h-small-GGUF