ChuckMcSneed
commited on
Commit
•
26b3ed1
1
Parent(s):
b54fd6d
Update README.md
Browse files
README.md
CHANGED
@@ -128,7 +128,7 @@ Thanks for GGUF, [@Nexesenex](https://huggingface.co/Nexesenex)!
|
|
128 |
| P | 5.5 | 2.25 |
|
129 |
| Total | 14.75 | 8.25 |
|
130 |
|
131 |
-
### Open LLM
|
132 |
[Leaderboard on Huggingface](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
133 |
|Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
|
134 |
|---------------------------------------|-------|-----|---------|-----|----------|----------|-----|
|
@@ -137,7 +137,7 @@ Thanks for GGUF, [@Nexesenex](https://huggingface.co/Nexesenex)!
|
|
137 |
|Difference |3.83 |1.62 |0.99 |4.7 |1.89 |2.44 |11.3 |
|
138 |
|
139 |
Here the losses seem far less brutal than on my bench. It seems that extending with LongLORA kills MMLU and GSM8K performance.
|
140 |
-
|
141 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ChuckMcSneed__WinterGoddess-1.4x-70b-32k)
|
142 |
|
143 |
| Metric |Value|
|
|
|
128 |
| P | 5.5 | 2.25 |
|
129 |
| Total | 14.75 | 8.25 |
|
130 |
|
131 |
+
### [Open LLM Leaderboard Evaluation Results](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
132 |
[Leaderboard on Huggingface](https://huggingface.co/spaces/HuggingFaceH4/open_llm_leaderboard)
|
133 |
|Model |Average|ARC |HellaSwag|MMLU |TruthfulQA|Winogrande|GSM8K|
|
134 |
|---------------------------------------|-------|-----|---------|-----|----------|----------|-----|
|
|
|
137 |
|Difference |3.83 |1.62 |0.99 |4.7 |1.89 |2.44 |11.3 |
|
138 |
|
139 |
Here the losses seem far less brutal than on my bench. It seems that extending with LongLORA kills MMLU and GSM8K performance.
|
140 |
+
|
141 |
Detailed results can be found [here](https://huggingface.co/datasets/open-llm-leaderboard/details_ChuckMcSneed__WinterGoddess-1.4x-70b-32k)
|
142 |
|
143 |
| Metric |Value|
|