Update README.md
Browse files
README.md
CHANGED
|
@@ -265,7 +265,8 @@ Command template:
|
|
| 265 |
The maximum context size of this model is 8192 tokens. These llamafiles
|
| 266 |
use a default context size of 512 tokens. Whenever you need the maximum
|
| 267 |
context size to be available with llamafile for any given model, you can
|
| 268 |
-
pass the `-c 0` flag.
|
|
|
|
| 269 |
|
| 270 |
## About Quantization
|
| 271 |
|
|
|
|
| 265 |
The maximum context size of this model is 8192 tokens. These llamafiles
|
| 266 |
use a default context size of 512 tokens. Whenever you need the maximum
|
| 267 |
context size to be available with llamafile for any given model, you can
|
| 268 |
+
pass the `-c 0` flag. The default temperature for these llamafiles is 0.
|
| 269 |
+
It can be changed, e.g. `--temp 0.8`.
|
| 270 |
|
| 271 |
## About Quantization
|
| 272 |
|