amanrangapur
commited on
Commit
•
640eb46
1
Parent(s):
bb91cbe
Update README.md
Browse files
README.md
CHANGED
@@ -49,7 +49,7 @@ inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
|
|
49 |
# olmo = olmo.to('cuda')
|
50 |
response = olmo.generate(**inputs, max_new_tokens=100, do_sample=True, top_k=50, top_p=0.95)
|
51 |
print(tokenizer.batch_decode(response, skip_special_tokens=True)[0])
|
52 |
-
>> 'Language modeling is
|
53 |
```
|
54 |
|
55 |
For faster performance, you can quantize the model using the following method:
|
|
|
49 |
# olmo = olmo.to('cuda')
|
50 |
response = olmo.generate(**inputs, max_new_tokens=100, do_sample=True, top_k=50, top_p=0.95)
|
51 |
print(tokenizer.batch_decode(response, skip_special_tokens=True)[0])
|
52 |
+
>> 'Language modeling is a key component of any text-based application, but its effectiveness...'
|
53 |
```
|
54 |
|
55 |
For faster performance, you can quantize the model using the following method:
|