--- license: gemma base_model: - ifable/gemma-2-Ifable-9B pipeline_tag: text-generation --- ## Llama.cpp imatrix quants of gemma-2-Ifable-9B Using llama.cpp release b3804 for quantization. Original model: https://huggingface.co/ifable/gemma-2-Ifable-9B Both quants were made using the imatrix option. The imatrix was generated with the dataset from [here](https://gist.github.com/tristandruyen/9e207a95c7d75ddf37525d353e00659c), using the BF16 GGUF with a context size of 8192 tokens (default is 512 but higher/same as model context size should improve quality) and 13 chunks.