File size: 6,405 Bytes
f08255e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
8367fb3
 
 
f08255e
 
 
 
 
 
 
8367fb3
f08255e
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
83
84
85
86
87
88
89
90
91
92
93
94
95
96
97
98
99
100
101
102
103
104
105
106
107
108
109
110
111
112
113
114
115
116
117
118
119
120
121
122
123
124
125
126
127
128
129
130
---
base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
inference: false
language:
  - en
license: cc-by-nc-4.0
model_creator: SanjiWatsuki
model_name: Kunoichi-DPO-v2-7B
model_type: mistral
pipeline_tag: text-generation
prompt_template: "{{system_message}}

  

  ### Instruction:

  {{prompt}}

  

  ### Response:

  "
quantized_by: brittlewis12
---

# Kunoichi-DPO-v2-7B GGUF

![Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B/resolve/main/assets/kunoichi.png)

Original model: [Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
Model creator: [SanjiWatsuki](https://huggingface.co/SanjiWatsuki)

This repo contains GGUF format model files for SanjiWatsuki’s Kunoichi-DPO-v2-7B.

### What is GGUF?

GGUF is a file format for representing AI models. It is the third version of the format, introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
Converted using llama.cpp build 1879 (revision [3e5ca79](https://github.com/ggerganov/llama.cpp/commit/3e5ca7931c68152e4ec18d126e9c832dd84914c8))

### Prompt template: Unknown (Alpaca)

[Alpaca-style](https://huggingface.co/SanjiWatsuki/Kunoichi-7B#prompt-template-custom-format-or-alpaca) was the prompt format for the original [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B).

```
Below is an instruction that describes a task. Write a response that appropriately completes the request.

### Instruction:
{{prompt}}

### Response:

```

---

## Download & run with [cnvrs](https://twitter.com/cnvrsai) on iPhone, iPad, and Mac!

![cnvrs.ai](https://pbs.twimg.com/profile_images/1744049151241797632/0mIP-P9e_400x400.jpg)

[cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
- create & save **Characters** with custom system prompts & temperature settings
- download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
- make it your own with custom **Theme colors**
- powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming!
- **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
- follow [cnvrs on twitter](https://twitter.com/cnvrsai) to stay up to date

---

## Original Model Evaluations:

| Model                | MT Bench | EQ Bench | MMLU   | Logic Test |
|----------------------|----------|----------|---------|-------------|
| GPT-4-Turbo         | 9.32     | -        | -       | -           |
| GPT-4               | 8.99     | 62.52    | 86.4    | 0.86        |
| **Kunoichi-DPO-v2-7B** | **8.51**     | **42.18**    | -    | **0.58**        |
| Mixtral-8x7B-Instruct| 8.30     | 44.81    | 70.6    | 0.75        |
| **Kunoichi-DPO-7B** | **8.29**     | **41.60**    | **64.83**    | **0.59**        |
| **Kunoichi-7B**     | **8.14**     | **44.32**    | **64.9**    | **0.58**            |
| Starling-7B         | 8.09     | -        | 63.9    | 0.51        |
| Claude-2            | 8.06     | 52.14    | 78.5    | -           |
| Silicon-Maid-7B     | 7.96     | 40.44    | 64.7    | 0.54           |
| Loyal-Macaroni-Maid-7B | 7.95     | 38.66    | 64.9   | 0.57        |
| GPT-3.5-Turbo       | 7.94     | 50.28    | 70      | 0.57        |
| Claude-1            | 7.9       | -        | 77      | -           |
| Openchat-3.5        | 7.81     | 37.08    | 64.3    | 0.39        |
| Dolphin-2.6-DPO     | 7.74     | 42.88    | 61.9    | 0.53        |
| Zephyr-7B-beta      | 7.34     | 38.71    | 61.4    | 0.30        |
| Llama-2-70b-chat-hf | 6.86     | 51.56    | 63      | -           |
| Neural-chat-7b-v3-1 | 6.84     | 43.61    | 62.4    | 0.30        |

| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|---|---:|---:|---:|---:|---:|
| **Kunoichi-DPO-7B**|**58.4**|  45.08 |  74|     66.99|   47.52|
| **Kunoichi-DPO-v2-7B**|**58.31**|  44.85|  75.05|     65.69|   47.65|
| [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B)|57.54|  44.99|  74.86|     63.72|   46.58|
| [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)| 56.85 | 44.74 | 75.6 | 59.89 | 47.17 |
| [Silicon-Maid-7B](https://huggingface.co/SanjiWatsuki/Silicon-Maid-7B) | 56.45|  44.74|  74.26|      61.5|   45.32|
| [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) | 53.51 | 43.67 | 73.24 | 55.37 | 41.76 |
| [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B)  | 52.42 | 42.75 | 72.99 | 52.99 | 40.94 |
| [openchat/openchat_3.5](https://huggingface.co/openchat/openchat_3.5) | 51.34 | 42.67 | 72.92 | 47.27 | 42.51 |
| [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha) | 51.16 | 42.06 | 72.72 | 47.33 | 42.53 |
| [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) | 50.99 | 37.33 | 71.83 | 55.1 | 39.7 |

| Model                       | AlpacaEval2 | Length |
| --------------------------- | ----------- | ------ |
| GPT-4                       | 23.58%      | 1365   |
| GPT-4 0314                  | 22.07%      | 1371   |
| Mistral Medium              | 21.86%      | 1500   |
| Mixtral 8x7B v0.1           | 18.26%      | 1465   |
| **Kunoichi-DPO-v2**         | **17.19%**  | 1785   |
| Claude 2                    | 17.19%      | 1069   |
| Claude                      | 16.99%      | 1082   |
| Gemini Pro                  | 16.85%      | 1315   |
| GPT-4 0613                  | 15.76%      | 1140   |
| Claude 2.1                  | 15.73%      | 1096   |
| Mistral 7B v0.2             | 14.72%      | 1676   |
| GPT 3.5 Turbo 0613          | 14.13%      | 1328   |
| LLaMA2 Chat 70B             | 13.87%      | 1790   |
| LMCocktail-10.7B-v1         | 13.15%      | 1203   |
| WizardLM 13B V1.1           | 11.23%      | 1525   |
| Zephyr 7B Beta              | 10.99%      | 1444   |
| OpenHermes-2.5-Mistral (7B) | 10.34%      | 1107   |
| GPT 3.5 Turbo 0301          | 9.62%       | 827    |
| **Kunoichi-7B**             | **9.38%**   | 1492   |
| GPT 3.5 Turbo 1106          | 9.18%       | 796    |
| GPT-3.5                     | 8.56%       | 1018   |
| Phi-2 DPO                   | 7.76%       | 1687   |
| LLaMA2 Chat 13B             | 7.70%       | 1513   |