Update README.md
Browse files
README.md
CHANGED
@@ -1,7 +1,7 @@
|
|
1 |
-
---
|
2 |
-
library_name: transformers
|
3 |
-
license: other
|
4 |
-
---
|
5 |
|
6 |
This is an early checkpoint of sarvam-2b, a small, yet powerful language model pre-trained from scratch on 4 trillion tokens. It is trained to be good at 10 Indic languages + English. Officially, the Indic languages supported are: Bengali, Gujarati, Hindi, Kannada, Malayalam, Marathi, Oriya, Punjabi, Tamil, and Telugu.
|
7 |
|
@@ -10,7 +10,7 @@ sarvam-2b will be trained on a data mixture containing equal parts English (2T)
|
|
10 |
Getting started:
|
11 |
```
|
12 |
from transformers import pipeline
|
13 |
-
pipe = pipeline(model='
|
14 |
pipe('भारत के प्रथम प्रधानमंत्री', max_new_tokens=15, temperature=0.1, repetition_penalty=1.2)[0]['generated_text']
|
15 |
# 'भारत के प्रथम प्रधानमंत्री जवाहरलाल नेहरू की बेटी इंदिरा गांधी थीं।\n\n'
|
16 |
```
|
|
|
1 |
+
---
|
2 |
+
library_name: transformers
|
3 |
+
license: other
|
4 |
+
---
|
5 |
|
6 |
This is an early checkpoint of sarvam-2b, a small, yet powerful language model pre-trained from scratch on 4 trillion tokens. It is trained to be good at 10 Indic languages + English. Officially, the Indic languages supported are: Bengali, Gujarati, Hindi, Kannada, Malayalam, Marathi, Oriya, Punjabi, Tamil, and Telugu.
|
7 |
|
|
|
10 |
Getting started:
|
11 |
```
|
12 |
from transformers import pipeline
|
13 |
+
pipe = pipeline(model='sarvamai/sarvam-2b-v0.5', device=0)
|
14 |
pipe('भारत के प्रथम प्रधानमंत्री', max_new_tokens=15, temperature=0.1, repetition_penalty=1.2)[0]['generated_text']
|
15 |
# 'भारत के प्रथम प्रधानमंत्री जवाहरलाल नेहरू की बेटी इंदिरा गांधी थीं।\n\n'
|
16 |
```
|