Update README.md
Browse files
README.md
CHANGED
@@ -10,9 +10,11 @@ tags:
|
|
10 |
- rlhf
|
11 |
---
|
12 |
|
|
|
|
|
13 |
# NeuralBeagle14-7B
|
14 |
|
15 |
-
NeuralBeagle14-7B is a DPO fine-tune of [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) using [argilla/distilabel-intel-orca-dpo-pairs](https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs).
|
16 |
|
17 |
Thanks [Argilla](https://huggingface.co/argilla) for providing the dataset and the training recipe [here](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp). πͺ
|
18 |
|
@@ -22,7 +24,7 @@ The evaluation was performed using [LLM AutoEval](https://github.com/mlabonne/ll
|
|
22 |
|
23 |
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|
24 |
|---|---:|---:|---:|---:|---:|
|
25 |
-
|[**Beagle14-7B**](https://huggingface.co/mlabonne/Beagle14-7B)| ****| ****| ****| ****|
|
26 |
| [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) [π](https://gist.github.com/mlabonne/f5a5bf8c0827bbec2f05b97cc62d642c) | 59.4 | 44.38 | 76.53 | 69.44 | 47.25 |
|
27 |
| [mlabonne/NeuralDaredevil-7B](https://huggingface.co/mlabonne/NeuralDaredevil-7B) [π](https://gist.github.com/mlabonne/cbeb077d1df71cb81c78f742f19f4155) | 59.39 | 45.23 | 76.2 | 67.61 | 48.52 |
|
28 |
| [argilla/distilabeled-Marcoro14-7B-slerp](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp) [π](https://gist.github.com/mlabonne/9082c4e59f4d3f3543c5eda3f4807040) | 58.93 | 45.38 | 76.48 | 65.68 | 48.18 |
|
|
|
10 |
- rlhf
|
11 |
---
|
12 |
|
13 |
+
![](https://i.imgur.com/89ZAKcn.png)
|
14 |
+
|
15 |
# NeuralBeagle14-7B
|
16 |
|
17 |
+
NeuralBeagle14-7B is a DPO fine-tune of [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) using the [argilla/distilabel-intel-orca-dpo-pairs](https://huggingface.co/datasets/argilla/distilabel-intel-orca-dpo-pairs) preference dataset and my DPO notebook from [this article](https://towardsdatascience.com/fine-tune-a-mistral-7b-model-with-direct-preference-optimization-708042745aac).
|
18 |
|
19 |
Thanks [Argilla](https://huggingface.co/argilla) for providing the dataset and the training recipe [here](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp). πͺ
|
20 |
|
|
|
24 |
|
25 |
| Model | Average | AGIEval | GPT4All | TruthfulQA | Bigbench |
|
26 |
|---|---:|---:|---:|---:|---:|
|
27 |
+
|[**Beagle14-7B**](https://huggingface.co/mlabonne/Beagle14-7B)| ****| ****| ****| ****| ****|
|
28 |
| [mlabonne/Beagle14-7B](https://huggingface.co/mlabonne/Beagle14-7B) [π](https://gist.github.com/mlabonne/f5a5bf8c0827bbec2f05b97cc62d642c) | 59.4 | 44.38 | 76.53 | 69.44 | 47.25 |
|
29 |
| [mlabonne/NeuralDaredevil-7B](https://huggingface.co/mlabonne/NeuralDaredevil-7B) [π](https://gist.github.com/mlabonne/cbeb077d1df71cb81c78f742f19f4155) | 59.39 | 45.23 | 76.2 | 67.61 | 48.52 |
|
30 |
| [argilla/distilabeled-Marcoro14-7B-slerp](https://huggingface.co/argilla/distilabeled-Marcoro14-7B-slerp) [π](https://gist.github.com/mlabonne/9082c4e59f4d3f3543c5eda3f4807040) | 58.93 | 45.38 | 76.48 | 65.68 | 48.18 |
|