diff options
author | Richard Kiss <him@richardkiss.com> | 2023-11-11 22:04:58 -0800 |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-11-11 23:04:58 -0700 |
commit | 532dd74e38c29e16ea1cfc4e7eedb4f2fab3f3cd (patch) | |
tree | 649147d4e7ac415c2ba030d7eb5df33ae1b60067 /examples/main/README.md | |
parent | e86fc56f7521ca4b18d1d9939e82abd40c2f1c01 (diff) |
Fix some documentation typos/grammar mistakes (#4032)
* typos
* Update examples/parallel/README.md
Co-authored-by: Kerfuffle <44031344+KerfuffleV2@users.noreply.github.com>
---------
Co-authored-by: Kerfuffle <44031344+KerfuffleV2@users.noreply.github.com>
Diffstat (limited to 'examples/main/README.md')
-rw-r--r-- | examples/main/README.md | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/examples/main/README.md b/examples/main/README.md index a3428b48..c7997f66 100644 --- a/examples/main/README.md +++ b/examples/main/README.md @@ -142,7 +142,7 @@ The `--ctx-size` option allows you to set the size of the prompt context used by ### Extended Context Size -Some fine-tuned models have extened the context length by scaling RoPE. For example, if the original pretrained model have a context length (max sequence length) of 4096 (4k) and the fine-tuned model have 32k. That is a scaling factor of 8, and should work by setting the above `--ctx-size` to 32768 (32k) and `--rope-scale` to 8. +Some fine-tuned models have extended the context length by scaling RoPE. For example, if the original pre-trained model have a context length (max sequence length) of 4096 (4k) and the fine-tuned model have 32k. That is a scaling factor of 8, and should work by setting the above `--ctx-size` to 32768 (32k) and `--rope-scale` to 8. - `--rope-scale N`: Where N is the linear scaling factor used by the fine-tuned model. |