|
--- |
|
pipeline_tag: text-generation |
|
tags: |
|
- nlp |
|
- code |
|
- exllama |
|
- gguf |
|
- phi-4 |
|
- phi |
|
- microsoft |
|
- exl2 |
|
- code |
|
- math |
|
- chat |
|
- chat |
|
- conversational |
|
- roleplay |
|
- text-generation |
|
- safetensors |
|
library_name: transformers |
|
base_model: microsoft/phi-4 |
|
--- |
|
# phi-4-4.0bpw-exl2 |
|
|
|
**Original Model:** `/microsoft/phi-4` |
|
**Quantization Method:** `EXL2` |
|
|
|
## Overview |
|
This is an EXL2 4.0bpw quantized version of [phi-4](https://huggingface.co/microsoft/phi-4). |
|
|
|
## Quantization By |
|
I often have idle A100 GPUs while building/testing and training the RP app, so I put them to use quantizing models. |
|
|
|
I hope the community finds these quantizations useful. |
|
|
|
Andrew Webby @ [RolePlai](https://roleplai.app/) |
|
|
|
|