| Age | Commit message (Expand) | Author |
|---|---|---|
| 2024-09-08 | Adding fused rms_norm (#42) | Kawrakow |
| 2024-08-27 | Faster Gemma2 (#27) | Kawrakow |
| 2024-08-21 | softcap: minor improvement (#24) | Kawrakow |
| 2024-08-20 | Fused soft cap and SIMD-ified GeLU (#9) | Kawrakow |
| 2024-08-20 | iq4_k: use iq5_k also when n_gqa = 2 (#23) | Kawrakow |
| 2024-08-19 | iq2_k: slightly better bpw - accuracy compromise (#20) | Kawrakow |
| 2024-08-12 | Merge mainline - Aug 12 2024 (#17) | Kawrakow |
| 2024-08-09 | iq6_k: WIP (quantize/dequantize) | Iwan Kawrakow |
| 2024-08-07 | Adding IQ2_TN for use with ternary models (#13) | Kawrakow |
| 2024-08-05 | q2_K: allow it to detect ternary nets and quantize accordingly | Iwan Kawrakow |
| 2024-08-01 | iq3_k: Basics | Iwan Kawrakow |
| 2024-08-01 | iq5_k: Basics | Iwan Kawrakow |
| 2024-08-01 | iq2_k: Basics | Iwan Kawrakow |
| 2024-07-28 | IQ4_K: SOTA 4-bit quantization (#6) | Kawrakow |
| 2024-07-27 | Merge mainline llama.cpp (#3) | Kawrakow |
