--- tags: - quantized - 8-bit - GGUF license: apache-2.0 language: - en base_model: - ibm-granite/granite-guardian-3.1-2b pipeline_tag: text-generation --- A quantized version of [Granite Guardian 3.1 2B](https://huggingface.co/ibm-granite/granite-guardian-3.1-2b) model from IBM. Quantization is done by [llama.cpp](https://github.com/ggerganov/llama.cpp). ## Model Summary (from original repository) **Granite Guardian 3.1 2B** is a fine-tuned Granite 3.1 2B Instruct model designed to detect risks in prompts and responses. It can help with risk detection along many key dimensions catalogued in the [IBM AI Risk Atlas](https://www.ibm.com/docs/en/watsonx/saas?topic=ai-risk-atlas). It is trained on unique data comprising human annotations and synthetic data informed by internal red-teaming. It outperforms other open-source models in the same space on standard benchmarks. - **Developers:** IBM Research - **GitHub Repository:** [ibm-granite/granite-guardian](https://github.com/ibm-granite/granite-guardian) - **Cookbook:** [Granite Guardian Recipes](https://github.com/ibm-granite/granite-guardian/tree/main/cookbooks/granite-guardian-3.1) - **Website**: [Granite Guardian Docs](https://www.ibm.com/granite/docs/models/guardian/) - **Paper:** [Granite Guardian](https://arxiv.org/abs/2412.07724) - **Release Date**: December 18, 2024 - **License:** [Apache 2.0](https://www.apache.org/licenses/LICENSE-2.0)