File size: 6,252 Bytes
f08255e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
---
base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
inference: false
language:
  - en
license: cc-by-nc-4.0
model_creator: SanjiWatsuki
model_name: Kunoichi-DPO-v2-7B
model_type: mistral
pipeline_tag: text-generation
prompt_template: "{{system_message}}

  

  ### Instruction:

  {{prompt}}

  

  ### Response:

  "
quantized_by: brittlewis12
---

# Kunoichi-DPO-v2-7B GGUF

![Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B/resolve/main/assets/kunoichi.png)

Original model: [Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
Model creator: [SanjiWatsuki](https://huggingface.co/SanjiWatsuki)

This repo contains GGUF format model files for SanjiWatsuki’s Kunoichi-DPO-v2-7B.

### What is GGUF?

GGUF is a file format for representing AI models. It is the third version of the format, introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
Converted using llama.cpp build 1879 (revision [3e5ca79](https://github.com/ggerganov/llama.cpp/commit/3e5ca7931c68152e4ec18d126e9c832dd84914c8))

### Prompt template: Unknown (Alpaca)

[Alpaca-style](https://huggingface.co/SanjiWatsuki/Kunoichi-7B#prompt-template-custom-format-or-alpaca) was the prompt format for the original [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B).

```
Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{{prompt}}

### Response:

```

---

## Download & run with [cnvrs](https://testflight.apple.com/join/sFWReS7K) on iPhone, iPad, and Mac!

[cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
- create & save **Characters** with custom system prompts & temperature settings
- download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
- make it your own with custom **Theme colors**
- powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming!
- **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!

---

## Original Model Evaluations:

| Model                | MT Bench | EQ Bench | MMLU   | Logic Test |
|----------------------|----------|----------|---------|-------------|
| GPT-4-Turbo         | 9.32     | -        | -       | -           |
| GPT-4               | 8.99     | 62.52    | 86.4    | 0.86        |
| **Kunoichi-DPO-v2-7B** | **8.51**     | **42.18**    | -    | **0.58**        |
| Mixtral-8x7B-Instruct| 8.30     | 44.81    | 70.6    | 0.75        |
| **Kunoichi-DPO-7B** | **8.29**     | **41.60**    | **64.83**    | **0.59**        |
| **Kunoichi-7B**     | **8.14**     | **44.32**    | **64.9**    | **0.58**            |
| Starling-7B         | 8.09     | -        | 63.9    | 0.51        |
| Claude-2            | 8.06     | 52.14    | 78.5    | -           |
| Silicon-Maid-7B     | 7.96     | 40.44    | 64.7    | 0.54           |
| Loyal-Macaroni-Maid-7B | 7.95     | 38.66    | 64.9   | 0.57        |
| GPT-3.5-Turbo       | 7.94     | 50.28    | 70      | 0.57        |
| Claude-1            | 7.9       | -        | 77      | -           |
| Openchat-3.5        | 7.81     | 37.08    | 64.3    | 0.39        |
| Dolphin-2.6-DPO     | 7.74     | 42.88    | 61.9    | 0.53        |
| Zephyr-7B-beta      | 7.34     | 38.71    | 61.4    | 0.30        |
| Llama-2-70b-chat-hf | 6.86     | 51.56    | 63      | -           |
| Neural-chat-7b-v3-1 | 6.84     | 43.61    | 62.4    | 0.30        |

| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|---|---:|---:|---:|---:|---:|
| **Kunoichi-DPO-7B**|**58.4**|  45.08 |  74|     66.99|   47.52|
| **Kunoichi-DPO-v2-7B**|**58.31**|  44.85|  75.05|     65.69|   47.65|
| [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B)|57.54|  44.99|  74.86|     63.72|   46.58|
| [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)| 56.85 | 44.74 | 75.6 | 59.89 | 47.17 |
| [Silicon-Maid-7B](https://huggingface.co/SanjiWatsuki/Silicon-Maid-7B) | 56.45|  44.74|  74.26|      61.5|   45.32|
| [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) | 53.51 | 43.67 | 73.24 | 55.37 | 41.76 |
| [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B)  | 52.42 | 42.75 | 72.99 | 52.99 | 40.94 |
| [openchat/openchat_3.5](https://huggingface.co/openchat/openchat_3.5) | 51.34 | 42.67 | 72.92 | 47.27 | 42.51 |
| [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha) | 51.16 | 42.06 | 72.72 | 47.33 | 42.53 |
| [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) | 50.99 | 37.33 | 71.83 | 55.1 | 39.7 |

| Model                       | AlpacaEval2 | Length |
| --------------------------- | ----------- | ------ |
| GPT-4                       | 23.58%      | 1365   |
| GPT-4 0314                  | 22.07%      | 1371   |
| Mistral Medium              | 21.86%      | 1500   |
| Mixtral 8x7B v0.1           | 18.26%      | 1465   |
| **Kunoichi-DPO-v2**         | **17.19%**  | 1785   |
| Claude 2                    | 17.19%      | 1069   |
| Claude                      | 16.99%      | 1082   |
| Gemini Pro                  | 16.85%      | 1315   |
| GPT-4 0613                  | 15.76%      | 1140   |
| Claude 2.1                  | 15.73%      | 1096   |
| Mistral 7B v0.2             | 14.72%      | 1676   |
| GPT 3.5 Turbo 0613          | 14.13%      | 1328   |
| LLaMA2 Chat 70B             | 13.87%      | 1790   |
| LMCocktail-10.7B-v1         | 13.15%      | 1203   |
| WizardLM 13B V1.1           | 11.23%      | 1525   |
| Zephyr 7B Beta              | 10.99%      | 1444   |
| OpenHermes-2.5-Mistral (7B) | 10.34%      | 1107   |
| GPT 3.5 Turbo 0301          | 9.62%       | 827    |
| **Kunoichi-7B**             | **9.38%**   | 1492   |
| GPT 3.5 Turbo 1106          | 9.18%       | 796    |
| GPT-3.5                     | 8.56%       | 1018   |
| Phi-2 DPO                   | 7.76%       | 1687   |
| LLaMA2 Chat 13B             | 7.70%       | 1513   |