--- license: mit language: - en base_model: PhilipMay/Phi-3-mini-4k-instruct-LLaMAfied-8xMoE-raw --- # Phi-3 MoE mini 4k instruct raw GGUF This is a GGUF version of https://huggingface.co/PhilipMay/Phi-3-mini-4k-instruct-LLaMAfied-8xMoE-raw The source model is an 8x MoE version of microsoft/Phi-3-mini-4k-instruct. It is based on the Llamafied version vonjack/Phi-3-mini-4k-instruct-LLaMAfied of Gan Feng. It was created with the help of mergekit. I have included the gguf-imat.py script and imatrix\imatrix.txt configuration used for the conversion. This is based on FantasiaFoundry/GGUF-Quantization-Script, and tweaked to pad vocab to allow operation with this model. This model has been tested to be functional with LlamaSharp, so should be compatible with any llama.cpp based solutions.