summaryrefslogtreecommitdiff
diff options
context:
space:
mode:
authorGeorgi Gerganov <ggerganov@gmail.com>2023-10-18 21:44:43 +0300
committerGitHub <noreply@github.com>2023-10-18 21:44:43 +0300
commit004797f6ac135383f8c1d1f5bd415ddee2f79318 (patch)
tree24f70dbe6d117b5e258a9dacb9e9a8bc31a2102e
parent4e82b2ea3fa6482915d147bc9f46e70b9ada7700 (diff)
readme : update hot topics
-rw-r--r--README.md8
1 files changed, 2 insertions, 6 deletions
diff --git a/README.md b/README.md
index 4fd4bd42..ce63c6f0 100644
--- a/README.md
+++ b/README.md
@@ -10,13 +10,9 @@
Inference of [LLaMA](https://arxiv.org/abs/2302.13971) model in pure C/C++
### Hot topics
-- ‼️ BPE tokenizer update: existing Falcon and Starcoder `.gguf` models will need to be reconverted: [#3252](https://github.com/ggerganov/llama.cpp/pull/3252)
-- ‼️ Breaking change: `rope_freq_base` and `rope_freq_scale` must be set to zero to use the model default values: [#3401](https://github.com/ggerganov/llama.cpp/pull/3401)
-- Parallel decoding + continuous batching support added: [#3228](https://github.com/ggerganov/llama.cpp/pull/3228) \
- **Devs should become familiar with the new API**
-- Local Falcon 180B inference on Mac Studio
- https://github.com/ggerganov/llama.cpp/assets/1991296/98abd4e8-7077-464c-ae89-aebabca7757e
+- LLaVA support: https://github.com/ggerganov/llama.cpp/pull/3436
+- ‼️ BPE tokenizer update: existing Falcon and Starcoder `.gguf` models will need to be reconverted: [#3252](https://github.com/ggerganov/llama.cpp/pull/3252)
----