Update README.md
Browse files
README.md
CHANGED
@@ -151,7 +151,7 @@ tokenizer = AutoTokenizer.from_pretrained(model_id)
|
|
151 |
|
152 |
inputs = tokenizer("Bagaimanakah sebuah negara dapat terbentuk?", return_tensors="pt").to("cuda")
|
153 |
outputs = model.generate(inputs.input_ids,
|
154 |
-
max_new_tokens =
|
155 |
pad_token_id=tokenizer.pad_token_id,
|
156 |
eos_token_id=tokenizer.eos_token_id,
|
157 |
temperature=0.7,
|
|
|
151 |
|
152 |
inputs = tokenizer("Bagaimanakah sebuah negara dapat terbentuk?", return_tensors="pt").to("cuda")
|
153 |
outputs = model.generate(inputs.input_ids,
|
154 |
+
max_new_tokens = 2048,
|
155 |
pad_token_id=tokenizer.pad_token_id,
|
156 |
eos_token_id=tokenizer.eos_token_id,
|
157 |
temperature=0.7,
|