InferenceIllusionist
commited on
Commit
•
f318713
1
Parent(s):
13c4766
Update README.md
Browse filesUpdated merlinite-7b metrics from Open LLM leaderboard
README.md
CHANGED
@@ -129,9 +129,9 @@ These three models showed excellent acumen in technical topics so I wanted to se
|
|
129 |
### Benchmark Performance
|
130 |
| Name | Avg. | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
|
131 |
| ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |
|
132 |
-
| <b>Magic-Dolphin-7b</b> | <u><b>67.48</b></u> | 65.78 | 85.61 | 64.64 | 58.01 |
|
133 |
| dolphin-2.6-mistral-7b-dpo-laser | 67.28 | 66.3 | 85.73 | 63.16 | 61.71 | 79.16 | 47.61 |
|
134 |
-
| merlinite-7b |
|
135 |
| Hyperion-1.5-Mistral-7B | 61.43 | 60.49 | 83.64 | 63.57 | 41.78 | 78.61 | 40.49 |
|
136 |
|
137 |
This was my first experiment with merging models so any feedback is greatly appreciated.
|
|
|
129 |
### Benchmark Performance
|
130 |
| Name | Avg. | ARC | HellaSwag | MMLU | TruthfulQA | Winogrande | GSM8K |
|
131 |
| ---- | ---- | ---- | ---- | ---- | ---- | ---- | ---- |
|
132 |
+
| <b>Magic-Dolphin-7b</b> | <u><b>67.48</b></u> | 65.78 | 85.61 | 64.64 | 58.01 | 79.64 | <u><b>51.18</b></u> |
|
133 |
| dolphin-2.6-mistral-7b-dpo-laser | 67.28 | 66.3 | 85.73 | 63.16 | 61.71 | 79.16 | 47.61 |
|
134 |
+
| merlinite-7b | 64 | 63.65 | 84.52 | 64.91 | 50.15 | 79.72 | 41.09 |
|
135 |
| Hyperion-1.5-Mistral-7B | 61.43 | 60.49 | 83.64 | 63.57 | 41.78 | 78.61 | 40.49 |
|
136 |
|
137 |
This was my first experiment with merging models so any feedback is greatly appreciated.
|