Update README.md
Browse files
README.md
CHANGED
@@ -54,7 +54,7 @@ pipeline_tag: text-generation
|
|
54 |
</div>
|
55 |
|
56 |
<div style="display: flex; justify-content: center; align-items: center">
|
57 |
-
<img src="https://
|
58 |
</div>
|
59 |
|
60 |
|
@@ -183,7 +183,7 @@ Score 5: {orig_score5_description}
|
|
183 |
| **OpenChat-3.5-0106** | **7B** | **64.5** | 7.8 | **71.3** | **51.5** | **49.1** | 61.0 | 65.8 | **77.4** | 62.2 |
|
184 |
| OpenChat-3.5-1210 | **7B** | 63.8 | 7.76 | 68.9 | 49.5 | 48.0 | **61.8** | 65.3 | 77.3 | 61.8 |
|
185 |
| OpenChat-3.5 | **7B** | 61.6 | 7.81 | 55.5 | 47.6 | 47.4 | 59.1 | 64.3 | 77.3 | 63.5 |
|
186 |
-
| ChatGPT (March)* |
|
187 |
| | | | | | | | | | | |
|
188 |
| OpenHermes 2.5 | 7B | 59.3 | 7.54 | 48.2 | 49.4 | 46.5 | 57.5 | 63.8 | 73.5 | 59.9 |
|
189 |
| OpenOrca Mistral | 7B | 52.7 | 6.86 | 38.4 | 49.4 | 42.9 | 45.9 | 59.3 | 59.1 | 58.1 |
|
@@ -209,7 +209,7 @@ All models are evaluated in chat mode (e.g. with the respective conversation tem
|
|
209 |
| Model | Size | HumanEval+ pass@1 |
|
210 |
|-----------------------------|--------|-------------------|
|
211 |
| **OpenChat-3.5-0106** | **7B** | **65.9** |
|
212 |
-
| ChatGPT (December 12, 2023) |
|
213 |
| WizardCoder-Python-34B-V1.0 | 34B | 64.6 |
|
214 |
| OpenChat 3.5 1210 | 7B | 63.4 |
|
215 |
| OpenHermes 2.5 | 7B | 41.5 |
|
|
|
54 |
</div>
|
55 |
|
56 |
<div style="display: flex; justify-content: center; align-items: center">
|
57 |
+
<img src="https://raw.githubusercontent.com/imoneoi/openchat/master/assets/openchat-bench-0106.png" style="width: 100%; border-radius: 1em">
|
58 |
</div>
|
59 |
|
60 |
|
|
|
183 |
| **OpenChat-3.5-0106** | **7B** | **64.5** | 7.8 | **71.3** | **51.5** | **49.1** | 61.0 | 65.8 | **77.4** | 62.2 |
|
184 |
| OpenChat-3.5-1210 | **7B** | 63.8 | 7.76 | 68.9 | 49.5 | 48.0 | **61.8** | 65.3 | 77.3 | 61.8 |
|
185 |
| OpenChat-3.5 | **7B** | 61.6 | 7.81 | 55.5 | 47.6 | 47.4 | 59.1 | 64.3 | 77.3 | 63.5 |
|
186 |
+
| ChatGPT (March)* | ???B | 61.5 | **7.94** | 48.1 | 47.6 | 47.1 | 57.7 | **67.3** | 74.9 | **70.1** |
|
187 |
| | | | | | | | | | | |
|
188 |
| OpenHermes 2.5 | 7B | 59.3 | 7.54 | 48.2 | 49.4 | 46.5 | 57.5 | 63.8 | 73.5 | 59.9 |
|
189 |
| OpenOrca Mistral | 7B | 52.7 | 6.86 | 38.4 | 49.4 | 42.9 | 45.9 | 59.3 | 59.1 | 58.1 |
|
|
|
209 |
| Model | Size | HumanEval+ pass@1 |
|
210 |
|-----------------------------|--------|-------------------|
|
211 |
| **OpenChat-3.5-0106** | **7B** | **65.9** |
|
212 |
+
| ChatGPT (December 12, 2023) | ???B | 64.6 |
|
213 |
| WizardCoder-Python-34B-V1.0 | 34B | 64.6 |
|
214 |
| OpenChat 3.5 1210 | 7B | 63.4 |
|
215 |
| OpenHermes 2.5 | 7B | 41.5 |
|