diff options
author | Olivier Chafik <ochafik@users.noreply.github.com> | 2024-04-30 00:52:50 +0100 |
---|---|---|
committer | GitHub <noreply@github.com> | 2024-04-30 00:52:50 +0100 |
commit | 8843a98c2ba97a25e93319a104f9ddfaf83ce4c4 (patch) | |
tree | 82d73687b9dd42033a388d83c3b491925a0444b9 /examples/main | |
parent | b8c1476e44cc1f3a1811613f65251cf779067636 (diff) |
Improve usability of --model-url & related flags (#6930)
* args: default --model to models/ + filename from --model-url or --hf-file (or else legacy models/7B/ggml-model-f16.gguf)
* args: main & server now call gpt_params_handle_model_default
* args: define DEFAULT_MODEL_PATH + update cli docs
* curl: check url of previous download (.json metadata w/ url, etag & lastModified)
* args: fix update to quantize-stats.cpp
* curl: support legacy .etag / .lastModified companion files
* curl: rm legacy .etag file support
* curl: reuse regex across headers callback calls
* curl: unique_ptr to manage lifecycle of curl & outfile
* curl: nit: no need for multiline regex flag
* curl: update failed test (model file collision) + gitignore *.gguf.json
Diffstat (limited to 'examples/main')
-rw-r--r-- | examples/main/README.md | 2 |
1 files changed, 1 insertions, 1 deletions
diff --git a/examples/main/README.md b/examples/main/README.md index 649f4e0f..e7a38743 100644 --- a/examples/main/README.md +++ b/examples/main/README.md @@ -66,7 +66,7 @@ main.exe -m models\7B\ggml-model.bin --ignore-eos -n -1 --random-prompt In this section, we cover the most commonly used options for running the `main` program with the LLaMA models: -- `-m FNAME, --model FNAME`: Specify the path to the LLaMA model file (e.g., `models/7B/ggml-model.bin`). +- `-m FNAME, --model FNAME`: Specify the path to the LLaMA model file (e.g., `models/7B/ggml-model.gguf`; inferred from `--model-url` if set). - `-mu MODEL_URL --model-url MODEL_URL`: Specify a remote http url to download the file (e.g https://huggingface.co/ggml-org/models/resolve/main/phi-2/ggml-model-q4_0.gguf). - `-i, --interactive`: Run the program in interactive mode, allowing you to provide input directly and receive real-time responses. - `-ins, --instruct`: Run the program in instruction mode, which is particularly useful when working with Alpaca models. |