summaryrefslogtreecommitdiff
path: root/ggml-cuda/template-instances
diff options
context:
space:
mode:
authorIwan Kawrakow <iwan.kawrakow@gmail.com>2024-06-25 11:32:48 +0300
committerIwan Kawrakow <iwan.kawrakow@gmail.com>2024-06-25 11:32:48 +0300
commitaa14a06b44ff12be7e4461a6e169a657275a5b20 (patch)
treec0ab2e1cd51a778594f0dd226d3e54c102c81b39 /ggml-cuda/template-instances
parentcc44d4a5c3368801f1de0d68096619a6746d47a4 (diff)
Bitnet: trying an alternative iq1_bn grid
Faster on CUDA. The scalar version is faster too. The issue with CUDA is that now I see wild performance fluctuations. Running llama-bench I can get 220 t/s for TG-128 one time, and 190 t/s another time, with uncertaintiers of 1-2 t/s. Same for PP, results are jumping back-and-fort between ~9500 t/s and ~8900 t/s. So, basically no reliable measurement at this point, but for sure faster than the previous version, which was at around 170-180 t/s.
Diffstat (limited to 'ggml-cuda/template-instances')
0 files changed, 0 insertions, 0 deletions