summaryrefslogtreecommitdiff
path: root/ggml/src/ggml-cuda
AgeCommit message (Expand)Author
2025-02-07cuda: non-contiguous rms norm (#190)Kawrakow
2024-11-21MMQ for Q6_0 (#115)Kawrakow
2024-10-31Faster MoE inference (#112)Kawrakow
2024-10-26Bitnet CUDA improvements (#109)Kawrakow
2024-10-25Bitnet changes (#106)Kawrakow
2024-10-24Fix quantized k-cache without FA (#105)Kawrakow
2024-10-22Enable q6_0 for flash attention (#101)Kawrakow
2024-10-21Enable IQ4_NL for KV-cache in token generation using Flash Attention (#99)Kawrakow
2024-10-16Adding IQ4_KSS: 4.0 bpw quants (#89)Kawrakow
2024-10-13IQ2_KS: 2.1875 bpw non-linear quantization (#85)Kawrakow
2024-10-09New SOTA quantization: 4.25 bpw IQ4_KS (#83)Kawrakow
2024-10-04Move scale fudge factors to quantization (#81)Kawrakow
2024-10-02Fused unary(x)*y (#70)Kawrakow
2024-10-02Adding Q6_0 (#77)Kawrakow
2024-10-01CUDA: faster float -> iq4_nl conversion (#73)Kawrakow
2024-09-29Allow bf16 kv-cache (#69)Kawrakow
2024-09-28CUDA non-contiguous RoPE (#66)Kawrakow
2024-09-28Adding SWIGLU unary op (#65)Kawrakow
2024-09-27Adding ability to have meta data per tensor row (#61)Kawrakow
2024-09-14Adding bf16 support to CUDA (#40)Kawrakow
2024-09-09Add CUDA support for IQ1_TN (#45)Kawrakow
2024-09-08Adding fused rms_norm (#42)Kawrakow
2024-08-27Faster Gemma2 (#27)Kawrakow
2024-08-20Fused soft cap and SIMD-ified GeLU (#9)Kawrakow
2024-08-12Merge mainline - Aug 12 2024 (#17)Kawrakow
2024-08-09Fix Zen4 implementation of iq3_k, iq4_k, iq5_kIwan Kawrakow
2024-08-09iq6_k: CUDA dot productIwan Kawrakow
2024-08-09iq6_k: CUDA dequantizeIwan Kawrakow
2024-08-09iq6_k: WIP (nothing works)Iwan Kawrakow
2024-08-07Adding IQ2_TN for use with ternary models (#13)Kawrakow
2024-08-01Add copyright noticeIwan Kawrakow
2024-08-01iq3_k: faster CUDA dot productIwan Kawrakow
2024-08-01iq3_k: CUDA dot productIwan Kawrakow
2024-08-01iq3_k: BasicsIwan Kawrakow
2024-08-01iq2_k: very slightly better CUDA dot productIwan Kawrakow
2024-08-01iq2_k: better CUDA dot productIwan Kawrakow
2024-08-01iq2_k: CUDA dot product finally worksIwan Kawrakow
2024-08-01iq5_k: CUDA dot product finally worksIwan Kawrakow
2024-08-01Factor out iqk CUDA dot productsIwan Kawrakow
2024-08-01iq5_k: CUDA dot product still not workingIwan Kawrakow
2024-08-01iq5_k: BasicsIwan Kawrakow
2024-08-01iq2_k: BasicsIwan Kawrakow
2024-07-28IQ4_K: SOTA 4-bit quantization (#6)Kawrakow
2024-07-27Merge mainline llama.cpp (#3)Kawrakow