index
:
ik_llama.cpp.git
main
Unnamed repository; edit this file 'description' to name the repository.
summary
refs
log
tree
commit
diff
log msg
author
committer
range
path:
root
/
ggml-backend.c
Age
Commit message (
Expand
)
Author
2024-05-15
ggml : tag ggml_tensor::backend as deprecated (#7290)
slaren
2024-05-11
build: fix and ignore msvc warnings (ggml/805)
Borislav Stanimirov
2024-04-26
Reset schedule earlier to allow overlap with ggml graph computation on device...
agray3
2024-04-22
ggml : fix calloc argument ordering. (#6820)
Dave Airlie
2024-04-21
ggml : fix ggml_backend_cpu_supports_op() for CPY (#0)
Georgi Gerganov
2024-03-26
cuda : rename build flag to LLAMA_CUDA (#6299)
slaren
2024-03-18
backend : set max split inputs to GGML_MAX_SRC (#6137)
slaren
2024-03-18
backend : offload large batches to GPU (#6083)
slaren
2024-03-13
llama : add pipeline parallelism support (#6017)
slaren
2024-03-04
ggml : introduce ggml_status (ggml/750)
Michael Podvitskiy
2024-02-28
Introduce backend GUIDs (ggml/743)
UEXTM.com
2024-02-18
1.5 bit quantization (#5453)
Kawrakow
2024-02-17
ggml : add ALiBi support for ggml_soft_max_ext (#5488)
Georgi Gerganov
2024-02-17
ci : add an option to fail on compile warning (#3952)
Ananta Bastola
2024-02-13
Early return for zero size calls to get_tensor. (#5482)
AT
2024-02-12
sync : ggml (#5452)
Georgi Gerganov
2024-02-10
ggml : add abort_callback for cpu backend (ggml/725)
Michael Podvitskiy
2024-01-29
Nomic Vulkan backend (#4456)
Jared Van Bortel
2024-01-28
ggml : add Vulkan backend (#2059)
0cc4m
2024-01-28
ggml : add unified SYCL backend for Intel GPUs (#2690)
Abhilash Majumder
2024-01-26
cuda : fix tensor size calculation for non-split buffer (#5145)
slaren
2024-01-20
llama : run all KQV ops on the CPU with no KV offload (#5049)
slaren
2024-01-17
ggml : add IQ2 to test-backend-ops + refactoring (#4990)
Georgi Gerganov
2024-01-17
backend : add eval callback (#4935)
Georgi Gerganov
2024-01-16
ggml : introduce GGML_CALL function annotation (#4850)
Justine Tunney
2024-01-12
backend_sched : fix assignments
slaren
2024-01-12
llama : ggml-backend integration (#4766)
slaren
2024-01-05
ggml : add error handling to graph_compute (whisper/1714)
Finn Voorhees
2023-12-29
ggml : fix some mul mat cases + add tests for src1 F16 (ggml/669)
bssrdf
2023-12-24
cuda : improve cuda pool efficiency using virtual memory (#4606)
slaren
2023-12-21
llama : initial ggml-backend integration (#4520)
slaren
2023-12-07
sync : ggml (new ops, tests, backend, etc.) (#4359)
Georgi Gerganov
2023-11-13
sync : ggml (backend v2) (#3912)
Georgi Gerganov
2023-10-08
sync : ggml (ggml-backend) (#3548)
Georgi Gerganov