diff options
author | Cebtenzzre <cebtenzzre@gmail.com> | 2023-09-28 17:41:44 -0400 |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-09-28 17:41:44 -0400 |
commit | bc39553c901a91cfcb757863586250838c83eeab (patch) | |
tree | 507d1aedf8ad63e4ed84e37154de9abf31ba358a /examples/quantize | |
parent | 0ccfc62a96a6b59a8faa14d1b350493f4cd51ae2 (diff) |
build : enable more non-default compiler warnings (#3200)
Diffstat (limited to 'examples/quantize')
-rw-r--r-- | examples/quantize/quantize.cpp | 1 |
1 files changed, 1 insertions, 0 deletions
diff --git a/examples/quantize/quantize.cpp b/examples/quantize/quantize.cpp index 1c1d957e..c7dd0d89 100644 --- a/examples/quantize/quantize.cpp +++ b/examples/quantize/quantize.cpp @@ -72,6 +72,7 @@ static bool try_parse_ftype(const std::string & ftype_str_in, llama_ftype & ftyp // usage: // ./quantize [--allow-requantize] [--leave-output-tensor] models/llama/ggml-model.gguf [models/llama/ggml-model-quant.gguf] type [nthreads] // +[[noreturn]] static void usage(const char * executable) { printf("usage: %s [--help] [--allow-requantize] [--leave-output-tensor] model-f32.gguf [model-quant.gguf] type [nthreads]\n\n", executable); printf(" --allow-requantize: Allows requantizing tensors that have already been quantized. Warning: This can severely reduce quality compared to quantizing from 16bit or 32bit\n"); |