Safetensors
English
olmo2
amanrangapur commited on
Commit
640eb46
1 Parent(s): bb91cbe

Update README.md

Browse files
Files changed (1) hide show
  1. README.md +1 -1
README.md CHANGED
@@ -49,7 +49,7 @@ inputs = tokenizer(message, return_tensors='pt', return_token_type_ids=False)
49
  # olmo = olmo.to('cuda')
50
  response = olmo.generate(**inputs, max_new_tokens=100, do_sample=True, top_k=50, top_p=0.95)
51
  print(tokenizer.batch_decode(response, skip_special_tokens=True)[0])
52
- >> 'Language modeling is the first step to build natural language generation...'
53
  ```
54
 
55
  For faster performance, you can quantize the model using the following method:
 
49
  # olmo = olmo.to('cuda')
50
  response = olmo.generate(**inputs, max_new_tokens=100, do_sample=True, top_k=50, top_p=0.95)
51
  print(tokenizer.batch_decode(response, skip_special_tokens=True)[0])
52
+ >> 'Language modeling is a key component of any text-based application, but its effectiveness...'
53
  ```
54
 
55
  For faster performance, you can quantize the model using the following method: