summaryrefslogtreecommitdiff
path: root/ggml.h
AgeCommit message (Expand)Author
2023-10-24sync : ggml (conv ops + cuda MSVC fixes) (#3765)Georgi Gerganov
2023-10-20gguf : support big endian platform (#3552)Qin Yue Chen
2023-10-13ggml : add context enumeration functions (#3605)slaren
2023-10-08sync : ggml (ggml-backend) (#3548)Georgi Gerganov
2023-10-04sync : ggml (conv 1d + 2d updates, UB fixes) (#3468)Georgi Gerganov
2023-09-28build : enable more non-default compiler warnings (#3200)Cebtenzzre
2023-09-28ggml_tensor: update the structure comments. (#3283)Hua Jiang
2023-09-28train : finetune LORA (#2632)xaedes
2023-09-28gguf : basic type checking in gguf_get_* (#3346)Cebtenzzre
2023-09-28llama : custom attention mask + parallel decoding + no context swaps (#3228)Georgi Gerganov
2023-09-27metal : reusing llama.cpp logging (#3152)Rickard Hallerbäck
2023-09-15sync : ggml (Metal F32 support + reduce ggml-alloc size) (#3192)Georgi Gerganov
2023-09-12arm64 support for windows (#3007)Eric Sommerlade
2023-08-29ggml : add view_src and view_offs to ggml_tensor for views (#2874)slaren
2023-08-28train : mem usage and other improvements (#2439)xaedes
2023-08-28ggml : sync (mem align to header + conv_transpose_2d fixes + ggml_alloc) (#2852)Georgi Gerganov
2023-08-27gguf : add 64-bit support (GGUF v2) (#2821)Georgi Gerganov
2023-08-27ggml : detect SSSE3 (#2825)Przemysław Pawełczyk
2023-08-23llm : add Falcon support (#2717)Georgi Gerganov
2023-08-22ggml : sync latest (SAM + SD operators, CUDA alibi) (#2709)Georgi Gerganov
2023-08-22ggml : support CUDA's half type for aarch64(#1455) (#2670)Kylin
2023-08-21gguf : new file format with flexible meta data (beta) (#2398)Georgi Gerganov
2023-08-20ggml : move all type info to ggml_type_traits (#2663)slaren
2023-08-07ggml : sync (custom ops) (#2537)Georgi Gerganov
2023-07-30ggml : add graph tensor allocator (#2411)slaren
2023-07-26ggml : allocate graphs in a context (#2392)slaren
2023-07-25ggml : improve graph build time via hash table lookup (#2329)slaren
2023-07-24make rms_norm_eps a parameter (#2374)slaren
2023-07-24ggml : sync (unary ops refactor, static-correctness) (#2370)Georgi Gerganov
2023-07-23ggml: move op parameters from tensors to ggml_tensor::op_params (#2333)slaren
2023-07-21ggml : fix rope args order + assert (#2054)Georgi Gerganov
2023-07-15llama : add custom RoPE (#2054)Xiao-Yong Jin
2023-07-12ggml : add ggml_pool_1d and ggml_pool_2dGeorgi Gerganov
2023-07-11ggml : sync (abort callback, mul / add broadcast, fix alibi) (#2183)Georgi Gerganov
2023-07-11ggml : remove src0 and src1 from ggml_tensor and rename opt to src (#2178)Spencer Sutton
2023-07-07ggml : change ggml_graph_compute() API to not require context (#1999)Qingyou Meng
2023-07-06ggml : fix restrict usageGeorgi Gerganov
2023-07-05ggml : generalize `quantize_fns` for simpler FP16 handling (#1237)Stephan Walter
2023-07-04ggml : sync latest (new ops, macros, refactoring) (#2106)Georgi Gerganov
2023-07-01ggml : disable GGML_TASK_INIT and GGML_TASK_FINALIZE by default (#1995)Qingyou Meng
2023-06-27ggml : add support for ChatGLM RoPEGeorgi Gerganov
2023-06-26ggml : increase max tensor name + clean up compiler warnings in train-text (#...David Yang
2023-06-26ggml : add NUMA support (#1556)zrm
2023-06-25ggml : sync latest ggml (custom operators)Georgi Gerganov
2023-06-24ggml : improve ggml_graph_dump_dot, add ggml_format_name (#1978)slaren
2023-06-19ggml : sync latest ggml repo (#1924)Georgi Gerganov
2023-06-18metal : handle buffers larger than device's maxBufferLength (#1826)Georgi Gerganov
2023-06-14CUDA full GPU acceleration, KV cache in VRAM (#1827)Johannes Gäßler
2023-06-13train : improved training-from-scratch example (#1652)xaedes
2023-06-06Multi GPU support, CUDA refactor, CUDA scratch buffer (#1703)Johannes Gäßler