This is a more recent quantized GGUF version of OpenOrca-Platypus-13B quantized to 4_K-M respecting the new format from (https://github.com/ggerganov/llama.cpp).

(link to the original model : https://huggingface.co/Open-Orca/OpenOrca-Platypus2-13B)

Instruction template :

### Instruction:

<prompt> (without the <>)

### Response:
Downloads last month
8
GGUF
Model size
13B params
Architecture
llama

4-bit

Inference Providers NEW
This model is not currently available via any of the supported third-party Inference Providers, and HF Inference API was unable to determine this model's library.

Dataset used to train kroonen/OpenOrca-Platypus2-13B-GGUF