diff options
author | Kawrakow <iwankawrakow@gmail.com> | 2025-05-09 10:13:25 +0300 |
---|---|---|
committer | GitHub <noreply@github.com> | 2025-05-09 10:13:25 +0300 |
commit | 496451a1d4c41300ebdb102f12401b8ffa5b1d4b (patch) | |
tree | 1ccffb9d1efb20f095584e0ac846569494f90b1c | |
parent | bc6ae515ceb14eeaf198e00251a9689539cea176 (diff) |
Update README.md
-rw-r--r-- | README.md | 1 |
1 files changed, 1 insertions, 0 deletions
@@ -14,6 +14,7 @@ This repository is a fork of [llama.cpp](https://github.com/ggerganov/llama.cpp) ## Latest News +* May 9 2025: Support for LlaMA-3-Nmotron models added, see [PR 377](https://github.com/ikawrakow/ik_llama.cpp/pull/377) * May 7 2025: 🚀 Faster TG for DeepSeek models with GPU or hybrid GPU/CPU inference. See [PR 386](https://github.com/ikawrakow/ik_llama.cpp/pull/386) for details. Caveat: Ampere or newer Nvidia GPU required * May 4 2025: 🚀 Significant token generation performance improvement on CUDA with Flash Attention for GQA models. For details and benchmarks see [PR #370](https://github.com/ikawrakow/ik_llama.cpp/pull/370) * April 29 2025: Qwen3 support added |