Update README.md
Browse files
README.md
CHANGED
@@ -265,7 +265,8 @@ Command template:
|
|
265 |
The maximum context size of this model is 8192 tokens. These llamafiles
|
266 |
use a default context size of 512 tokens. Whenever you need the maximum
|
267 |
context size to be available with llamafile for any given model, you can
|
268 |
-
pass the `-c 0` flag.
|
|
|
269 |
|
270 |
## About Quantization
|
271 |
|
|
|
265 |
The maximum context size of this model is 8192 tokens. These llamafiles
|
266 |
use a default context size of 512 tokens. Whenever you need the maximum
|
267 |
context size to be available with llamafile for any given model, you can
|
268 |
+
pass the `-c 0` flag. The default temperature for these llamafiles is 0.
|
269 |
+
It can be changed, e.g. `--temp 0.8`.
|
270 |
|
271 |
## About Quantization
|
272 |
|