Update README.md
Browse files
README.md
CHANGED
@@ -112,20 +112,26 @@ model-index:
|
|
112 |
|
113 |
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/gFEhcIDSKa3AWpkNfH91q.jpeg)
|
114 |
|
115 |
-
This is a DPO fine-tune of [mlabonne/Daredevil-8-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated) trained on one epoch of [mlabonne/orpo-dpo-mix-40k](https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k).
|
116 |
-
|
117 |
|
118 |
## π Applications
|
119 |
|
120 |
-
|
121 |
|
122 |
-
Tested on LM Studio using the "Llama 3" preset.
|
|
|
|
|
|
|
|
|
123 |
|
124 |
## π Evaluation
|
125 |
|
126 |
### Open LLM Leaderboard
|
127 |
|
128 |
-
|
|
|
|
|
129 |
|
130 |
### Nous
|
131 |
|
@@ -141,20 +147,31 @@ Evaluation performed using [LLM AutoEval](https://github.com/mlabonne/llm-autoev
|
|
141 |
| [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) [π](https://gist.github.com/mlabonne/8329284d86035e6019edb11eb0933628) | 51.34 | 41.22 | 69.86 | 51.65 | 42.64 |
|
142 |
| [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) [π](https://gist.github.com/mlabonne/616b6245137a9cfc4ea80e4c6e55d847) | 45.42 | 31.1 | 69.95 | 43.91 | 36.7 |
|
143 |
|
144 |
-
|
145 |
## π³ Model family tree
|
146 |
|
147 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/ekwRGgnjzEOyprT8sEBFt.png)
|
148 |
-
# [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
149 |
-
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_mlabonne__Daredevil-8B-abliterated-dpomix)
|
150 |
-
|
151 |
-
| Metric |Value|
|
152 |
-
|---------------------------------|----:|
|
153 |
-
|Avg. |72.32|
|
154 |
-
|AI2 Reasoning Challenge (25-Shot)|69.28|
|
155 |
-
|HellaSwag (10-Shot) |85.05|
|
156 |
-
|MMLU (5-Shot) |69.10|
|
157 |
-
|TruthfulQA (0-shot) |60.00|
|
158 |
-
|Winogrande (5-shot) |78.69|
|
159 |
-
|GSM8k (5-shot) |71.80|
|
160 |
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
112 |
|
113 |
![image/jpeg](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/gFEhcIDSKa3AWpkNfH91q.jpeg)
|
114 |
|
115 |
+
This is a DPO fine-tune of [mlabonne/Daredevil-8-abliterated](https://huggingface.co/mlabonne/Daredevil-8B-abliterated), trained on one epoch of [mlabonne/orpo-dpo-mix-40k](https://huggingface.co/datasets/mlabonne/orpo-dpo-mix-40k).
|
116 |
+
The DPO fine-tuning successfully recovers the performance loss due to the abliteration process, making it an excellent uncensored model.
|
117 |
|
118 |
## π Applications
|
119 |
|
120 |
+
NeuralDaredevil-8B-abliterated performs better than the Instruct model on my tests.
|
121 |
|
122 |
+
You can use it for any application that doesn't require alignment, like role-playing. Tested on LM Studio using the "Llama 3" preset.
|
123 |
+
|
124 |
+
## β‘ Quantization
|
125 |
+
|
126 |
+
* **GGUF**: https://huggingface.co/mlabonne/NeuralDaredevil-8B-abliterated-GGUF
|
127 |
|
128 |
## π Evaluation
|
129 |
|
130 |
### Open LLM Leaderboard
|
131 |
|
132 |
+
NeuralDaredevil-8B is the best-performing uncensored 8B model on the Open LLM Leaderboard (MMLU score).
|
133 |
+
|
134 |
+
![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/TqTEbSQ7HAE7QTQBSyKoi.png)
|
135 |
|
136 |
### Nous
|
137 |
|
|
|
147 |
| [meta-llama/Meta-Llama-3-8B-Instruct](https://huggingface.co/meta-llama/Meta-Llama-3-8B-Instruct) [π](https://gist.github.com/mlabonne/8329284d86035e6019edb11eb0933628) | 51.34 | 41.22 | 69.86 | 51.65 | 42.64 |
|
148 |
| [meta-llama/Meta-Llama-3-8B](https://huggingface.co/meta-llama/Meta-Llama-3-8B) [π](https://gist.github.com/mlabonne/616b6245137a9cfc4ea80e4c6e55d847) | 45.42 | 31.1 | 69.95 | 43.91 | 36.7 |
|
149 |
|
|
|
150 |
## π³ Model family tree
|
151 |
|
152 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/61b8e2ba285851687028d395/ekwRGgnjzEOyprT8sEBFt.png)
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
|
153 |
|
154 |
+
## π» Usage
|
155 |
+
|
156 |
+
```python
|
157 |
+
!pip install -qU transformers accelerate
|
158 |
+
|
159 |
+
from transformers import AutoTokenizer
|
160 |
+
import transformers
|
161 |
+
import torch
|
162 |
+
|
163 |
+
model = "mlabonne/Daredevil-8B"
|
164 |
+
messages = [{"role": "user", "content": "What is a large language model?"}]
|
165 |
+
|
166 |
+
tokenizer = AutoTokenizer.from_pretrained(model)
|
167 |
+
prompt = tokenizer.apply_chat_template(messages, tokenize=False, add_generation_prompt=True)
|
168 |
+
pipeline = transformers.pipeline(
|
169 |
+
"text-generation",
|
170 |
+
model=model,
|
171 |
+
torch_dtype=torch.float16,
|
172 |
+
device_map="auto",
|
173 |
+
)
|
174 |
+
|
175 |
+
outputs = pipeline(prompt, max_new_tokens=256, do_sample=True, temperature=0.7, top_k=50, top_p=0.95)
|
176 |
+
print(outputs[0]["generated_text"])
|
177 |
+
```
|