qanthony-z
commited on
Commit
•
5fc1404
1
Parent(s):
c9ea486
update bar chart
Browse files
README.md
CHANGED
@@ -49,13 +49,12 @@ print((tokenizer.decode(outputs[0])))
|
|
49 |
Zamba2-2.7B-Instruct punches dramatically above its weight, achieving extremely strong instruction-following benchmark scores, significantly outperforming Gemma2-2B-Instruct of the same size and outperforming Mistral-7B-Instruct in most metrics.
|
50 |
|
51 |
<center>
|
52 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/
|
53 |
</center>
|
54 |
|
55 |
|
56 |
-
|
57 |
-
|
58 |
-
|---------------------------|-----:|---------:|---------:|
|
59 |
| **Zamba2-2.7B-Instruct** | 2.7B | **72.40**| **48.02**|
|
60 |
| Mistral-7B-Instruct | 7B| 66.4 | 45.3 |
|
61 |
| Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
|
@@ -66,7 +65,7 @@ Zamba2-2.7B-Instruct punches dramatically above its weight, achieving extremely
|
|
66 |
Moreover, due to its unique hybrid SSM architecture, Zamba2-2.7B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
|
67 |
|
68 |
<center>
|
69 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/
|
70 |
</center>
|
71 |
|
72 |
|
|
|
49 |
Zamba2-2.7B-Instruct punches dramatically above its weight, achieving extremely strong instruction-following benchmark scores, significantly outperforming Gemma2-2B-Instruct of the same size and outperforming Mistral-7B-Instruct in most metrics.
|
50 |
|
51 |
<center>
|
52 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/QnudHrMeMx_NuRc2evwRG.png" width="900"/>
|
53 |
</center>
|
54 |
|
55 |
|
56 |
+
| Model | Size | Aggregate MT-Bench | IFEval |
|
57 |
+
|:---------------------------:|:-----:|:------------------:|:---------:|
|
|
|
58 |
| **Zamba2-2.7B-Instruct** | 2.7B | **72.40**| **48.02**|
|
59 |
| Mistral-7B-Instruct | 7B| 66.4 | 45.3 |
|
60 |
| Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
|
|
|
65 |
Moreover, due to its unique hybrid SSM architecture, Zamba2-2.7B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
|
66 |
|
67 |
<center>
|
68 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/WKTcYkhDgJCHyze4TDpLa.png" width="700" alt="Zamba performance">
|
69 |
</center>
|
70 |
|
71 |
|