diff options
author | Georgi Gerganov <ggerganov@gmail.com> | 2023-10-18 21:44:43 +0300 |
---|---|---|
committer | GitHub <noreply@github.com> | 2023-10-18 21:44:43 +0300 |
commit | 004797f6ac135383f8c1d1f5bd415ddee2f79318 (patch) | |
tree | 24f70dbe6d117b5e258a9dacb9e9a8bc31a2102e | |
parent | 4e82b2ea3fa6482915d147bc9f46e70b9ada7700 (diff) |
readme : update hot topics
-rw-r--r-- | README.md | 8 |
1 files changed, 2 insertions, 6 deletions
@@ -10,13 +10,9 @@ Inference of [LLaMA](https://arxiv.org/abs/2302.13971) model in pure C/C++ ### Hot topics -- ‼️ BPE tokenizer update: existing Falcon and Starcoder `.gguf` models will need to be reconverted: [#3252](https://github.com/ggerganov/llama.cpp/pull/3252) -- ‼️ Breaking change: `rope_freq_base` and `rope_freq_scale` must be set to zero to use the model default values: [#3401](https://github.com/ggerganov/llama.cpp/pull/3401) -- Parallel decoding + continuous batching support added: [#3228](https://github.com/ggerganov/llama.cpp/pull/3228) \ - **Devs should become familiar with the new API** -- Local Falcon 180B inference on Mac Studio - https://github.com/ggerganov/llama.cpp/assets/1991296/98abd4e8-7077-464c-ae89-aebabca7757e +- LLaVA support: https://github.com/ggerganov/llama.cpp/pull/3436 +- ‼️ BPE tokenizer update: existing Falcon and Starcoder `.gguf` models will need to be reconverted: [#3252](https://github.com/ggerganov/llama.cpp/pull/3252) ---- |