qanthony-z commited on
Commit
5fc1404
1 Parent(s): c9ea486

update bar chart

Browse files
Files changed (1) hide show
  1. README.md +4 -5
README.md CHANGED
@@ -49,13 +49,12 @@ print((tokenizer.decode(outputs[0])))
49
  Zamba2-2.7B-Instruct punches dramatically above its weight, achieving extremely strong instruction-following benchmark scores, significantly outperforming Gemma2-2B-Instruct of the same size and outperforming Mistral-7B-Instruct in most metrics.
50
 
51
  <center>
52
- <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/ml2i8lHK4ojqBh_xV50L9.png" width="900"/>
53
  </center>
54
 
55
 
56
-
57
- | Model | Size | MT-Bench | IFEval |
58
- |---------------------------|-----:|---------:|---------:|
59
  | **Zamba2-2.7B-Instruct** | 2.7B | **72.40**| **48.02**|
60
  | Mistral-7B-Instruct | 7B| 66.4 | 45.3 |
61
  | Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
@@ -66,7 +65,7 @@ Zamba2-2.7B-Instruct punches dramatically above its weight, achieving extremely
66
  Moreover, due to its unique hybrid SSM architecture, Zamba2-2.7B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
67
 
68
  <center>
69
- <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/GyojH0mFCaAaAHBAXlm4T.png" width="700" alt="Zamba performance">
70
  </center>
71
 
72
 
 
49
  Zamba2-2.7B-Instruct punches dramatically above its weight, achieving extremely strong instruction-following benchmark scores, significantly outperforming Gemma2-2B-Instruct of the same size and outperforming Mistral-7B-Instruct in most metrics.
50
 
51
  <center>
52
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/QnudHrMeMx_NuRc2evwRG.png" width="900"/>
53
  </center>
54
 
55
 
56
+ | Model | Size | Aggregate MT-Bench | IFEval |
57
+ |:---------------------------:|:-----:|:------------------:|:---------:|
 
58
  | **Zamba2-2.7B-Instruct** | 2.7B | **72.40**| **48.02**|
59
  | Mistral-7B-Instruct | 7B| 66.4 | 45.3 |
60
  | Gemma2-2B-Instruct | 2.7B | 51.69 | 42.20 |
 
65
  Moreover, due to its unique hybrid SSM architecture, Zamba2-2.7B-Instruct achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer-based models.
66
 
67
  <center>
68
+ <img src="https://cdn-uploads.huggingface.co/production/uploads/65bc13717c6ad1994b6619e9/WKTcYkhDgJCHyze4TDpLa.png" width="700" alt="Zamba performance">
69
  </center>
70
 
71