brittlewis12 commited on
Commit
f08255e
1 Parent(s): 7509ece

Create README.md

Browse files
Files changed (1) hide show
  1. README.md +127 -0
README.md ADDED
@@ -0,0 +1,127 @@
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
 
1
+ ---
2
+ base_model: SanjiWatsuki/Kunoichi-DPO-v2-7B
3
+ inference: false
4
+ language:
5
+ - en
6
+ license: cc-by-nc-4.0
7
+ model_creator: SanjiWatsuki
8
+ model_name: Kunoichi-DPO-v2-7B
9
+ model_type: mistral
10
+ pipeline_tag: text-generation
11
+ prompt_template: "{{system_message}}
12
+
13
+
14
+
15
+ ### Instruction:
16
+
17
+ {{prompt}}
18
+
19
+
20
+
21
+ ### Response:
22
+
23
+ "
24
+ quantized_by: brittlewis12
25
+ ---
26
+
27
+ # Kunoichi-DPO-v2-7B GGUF
28
+
29
+ ![Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B/resolve/main/assets/kunoichi.png)
30
+
31
+ Original model: [Kunoichi-DPO-v2-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-DPO-v2-7B)
32
+ Model creator: [SanjiWatsuki](https://huggingface.co/SanjiWatsuki)
33
+
34
+ This repo contains GGUF format model files for SanjiWatsuki’s Kunoichi-DPO-v2-7B.
35
+
36
+ ### What is GGUF?
37
+
38
+ GGUF is a file format for representing AI models. It is the third version of the format, introduced by the llama.cpp team on August 21st 2023. It is a replacement for GGML, which is no longer supported by llama.cpp.
39
+ Converted using llama.cpp build 1879 (revision [3e5ca79](https://github.com/ggerganov/llama.cpp/commit/3e5ca7931c68152e4ec18d126e9c832dd84914c8))
40
+
41
+ ### Prompt template: Unknown (Alpaca)
42
+
43
+ [Alpaca-style](https://huggingface.co/SanjiWatsuki/Kunoichi-7B#prompt-template-custom-format-or-alpaca) was the prompt format for the original [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B).
44
+
45
+ ```
46
+ Below is an instruction that describes a task. Write a response that appropriately completes the request.
47
+
48
+ ### Instruction:
49
+ {{prompt}}
50
+
51
+ ### Response:
52
+
53
+ ```
54
+
55
+ ---
56
+
57
+ ## Download & run with [cnvrs](https://testflight.apple.com/join/sFWReS7K) on iPhone, iPad, and Mac!
58
+
59
+ [cnvrs](https://testflight.apple.com/join/sFWReS7K) is the best app for private, local AI on your device:
60
+ - create & save **Characters** with custom system prompts & temperature settings
61
+ - download and experiment with any **GGUF model** you can [find on HuggingFace](https://huggingface.co/models?library=gguf)!
62
+ - make it your own with custom **Theme colors**
63
+ - powered by Metal ⚡️ & [Llama.cpp](https://github.com/ggerganov/llama.cpp), with **haptics** during response streaming!
64
+ - **try it out** yourself today, on [Testflight](https://testflight.apple.com/join/sFWReS7K)!
65
+
66
+ ---
67
+
68
+ ## Original Model Evaluations:
69
+
70
+ | Model | MT Bench | EQ Bench | MMLU | Logic Test |
71
+ |----------------------|----------|----------|---------|-------------|
72
+ | GPT-4-Turbo | 9.32 | - | - | - |
73
+ | GPT-4 | 8.99 | 62.52 | 86.4 | 0.86 |
74
+ | **Kunoichi-DPO-v2-7B** | **8.51** | **42.18** | - | **0.58** |
75
+ | Mixtral-8x7B-Instruct| 8.30 | 44.81 | 70.6 | 0.75 |
76
+ | **Kunoichi-DPO-7B** | **8.29** | **41.60** | **64.83** | **0.59** |
77
+ | **Kunoichi-7B** | **8.14** | **44.32** | **64.9** | **0.58** |
78
+ | Starling-7B | 8.09 | - | 63.9 | 0.51 |
79
+ | Claude-2 | 8.06 | 52.14 | 78.5 | - |
80
+ | Silicon-Maid-7B | 7.96 | 40.44 | 64.7 | 0.54 |
81
+ | Loyal-Macaroni-Maid-7B | 7.95 | 38.66 | 64.9 | 0.57 |
82
+ | GPT-3.5-Turbo | 7.94 | 50.28 | 70 | 0.57 |
83
+ | Claude-1 | 7.9 | - | 77 | - |
84
+ | Openchat-3.5 | 7.81 | 37.08 | 64.3 | 0.39 |
85
+ | Dolphin-2.6-DPO | 7.74 | 42.88 | 61.9 | 0.53 |
86
+ | Zephyr-7B-beta | 7.34 | 38.71 | 61.4 | 0.30 |
87
+ | Llama-2-70b-chat-hf | 6.86 | 51.56 | 63 | - |
88
+ | Neural-chat-7b-v3-1 | 6.84 | 43.61 | 62.4 | 0.30 |
89
+
90
+ | Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
91
+ |---|---:|---:|---:|---:|---:|
92
+ | **Kunoichi-DPO-7B**|**58.4**| 45.08 | 74| 66.99| 47.52|
93
+ | **Kunoichi-DPO-v2-7B**|**58.31**| 44.85| 75.05| 65.69| 47.65|
94
+ | [Kunoichi-7B](https://huggingface.co/SanjiWatsuki/Kunoichi-7B)|57.54| 44.99| 74.86| 63.72| 46.58|
95
+ | [OpenPipe/mistral-ft-optimized-1218](https://huggingface.co/OpenPipe/mistral-ft-optimized-1218)| 56.85 | 44.74 | 75.6 | 59.89 | 47.17 |
96
+ | [Silicon-Maid-7B](https://huggingface.co/SanjiWatsuki/Silicon-Maid-7B) | 56.45| 44.74| 74.26| 61.5| 45.32|
97
+ | [mlabonne/NeuralHermes-2.5-Mistral-7B](https://huggingface.co/mlabonne/NeuralHermes-2.5-Mistral-7B) | 53.51 | 43.67 | 73.24 | 55.37 | 41.76 |
98
+ | [teknium/OpenHermes-2.5-Mistral-7B](https://huggingface.co/teknium/OpenHermes-2.5-Mistral-7B) | 52.42 | 42.75 | 72.99 | 52.99 | 40.94 |
99
+ | [openchat/openchat_3.5](https://huggingface.co/openchat/openchat_3.5) | 51.34 | 42.67 | 72.92 | 47.27 | 42.51 |
100
+ | [berkeley-nest/Starling-LM-7B-alpha](https://huggingface.co/berkeley-nest/Starling-LM-7B-alpha) | 51.16 | 42.06 | 72.72 | 47.33 | 42.53 |
101
+ | [HuggingFaceH4/zephyr-7b-beta](https://huggingface.co/HuggingFaceH4/zephyr-7b-beta) | 50.99 | 37.33 | 71.83 | 55.1 | 39.7 |
102
+
103
+ | Model | AlpacaEval2 | Length |
104
+ | --------------------------- | ----------- | ------ |
105
+ | GPT-4 | 23.58% | 1365 |
106
+ | GPT-4 0314 | 22.07% | 1371 |
107
+ | Mistral Medium | 21.86% | 1500 |
108
+ | Mixtral 8x7B v0.1 | 18.26% | 1465 |
109
+ | **Kunoichi-DPO-v2** | **17.19%** | 1785 |
110
+ | Claude 2 | 17.19% | 1069 |
111
+ | Claude | 16.99% | 1082 |
112
+ | Gemini Pro | 16.85% | 1315 |
113
+ | GPT-4 0613 | 15.76% | 1140 |
114
+ | Claude 2.1 | 15.73% | 1096 |
115
+ | Mistral 7B v0.2 | 14.72% | 1676 |
116
+ | GPT 3.5 Turbo 0613 | 14.13% | 1328 |
117
+ | LLaMA2 Chat 70B | 13.87% | 1790 |
118
+ | LMCocktail-10.7B-v1 | 13.15% | 1203 |
119
+ | WizardLM 13B V1.1 | 11.23% | 1525 |
120
+ | Zephyr 7B Beta | 10.99% | 1444 |
121
+ | OpenHermes-2.5-Mistral (7B) | 10.34% | 1107 |
122
+ | GPT 3.5 Turbo 0301 | 9.62% | 827 |
123
+ | **Kunoichi-7B** | **9.38%** | 1492 |
124
+ | GPT 3.5 Turbo 1106 | 9.18% | 796 |
125
+ | GPT-3.5 | 8.56% | 1018 |
126
+ | Phi-2 DPO | 7.76% | 1687 |
127
+ | LLaMA2 Chat 13B | 7.70% | 1513 |