Update README.md
Browse files
README.md
CHANGED
@@ -82,6 +82,9 @@ Zamba2-1.2B utilizes and extends our original Zamba hybrid SSM-attention archite
|
|
82 |
|
83 |
Zamba2-1.2B achieves leading and state-of-the-art performance among models of <2B parameters and is competitive with some models of significantly greater size. Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer based models.
|
84 |
|
|
|
|
|
|
|
85 |
Zamba2-1.2B's high performance and small inference compute and memory footprint renders it an ideal generalist model for on-device applications.
|
86 |
|
87 |
<center>
|
@@ -89,10 +92,6 @@ Zamba2-1.2B's high performance and small inference compute and memory footprint
|
|
89 |
</center>
|
90 |
|
91 |
|
92 |
-
<center>
|
93 |
-
<img src="https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/JVZUvVMPIpIJy9RDyohMJ.png" width="800" alt="Zamba performance">
|
94 |
-
</center>
|
95 |
-
|
96 |
Time to First Token (TTFT) | Output Generation
|
97 |
:-------------------------:|:-------------------------:
|
98 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/5lpWDLdtPPVAk8COJq7gZ.png) | ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/V2tS6eCOGbpKybEoZmOB7.png)
|
@@ -103,8 +102,3 @@ And memory overhead
|
|
103 |
<center>
|
104 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/m0YUmAmiVnRg6l9m10CEt.png" width="400" alt="Zamba inference and memory cost">
|
105 |
</center>
|
106 |
-
|
107 |
-
|
108 |
-
## Notice
|
109 |
-
|
110 |
-
Zamba2-1.2B is a pretrained base model and therefore does not have any moderation mechanism and may output toxic or otherwise harmful language. In addition, one should not expect good instruct or chat performance, as this model was not fine-tuned for instruction following or chat.
|
|
|
82 |
|
83 |
Zamba2-1.2B achieves leading and state-of-the-art performance among models of <2B parameters and is competitive with some models of significantly greater size. Moreover, due to its unique hybrid SSM architecture, Zamba2-1.2B achieves extremely low inference latency and rapid generation with a significantly smaller memory footprint than comparable transformer based models.
|
84 |
|
85 |
+
<img src="https://cdn-uploads.huggingface.co/production/uploads/64e40335c0edca443ef8af3e/t7et3jazHNvxKSkeorZuo.png" width="600"/>
|
86 |
+
|
87 |
+
|
88 |
Zamba2-1.2B's high performance and small inference compute and memory footprint renders it an ideal generalist model for on-device applications.
|
89 |
|
90 |
<center>
|
|
|
92 |
</center>
|
93 |
|
94 |
|
|
|
|
|
|
|
|
|
95 |
Time to First Token (TTFT) | Output Generation
|
96 |
:-------------------------:|:-------------------------:
|
97 |
![image/png](https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/5lpWDLdtPPVAk8COJq7gZ.png) | ![image/png](https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/V2tS6eCOGbpKybEoZmOB7.png)
|
|
|
102 |
<center>
|
103 |
<img src="https://cdn-uploads.huggingface.co/production/uploads/65c05e75c084467acab2f84a/m0YUmAmiVnRg6l9m10CEt.png" width="400" alt="Zamba inference and memory cost">
|
104 |
</center>
|
|
|
|
|
|
|
|
|
|