summaryrefslogtreecommitdiff
path: root/examples/train-text-from-scratch/train-text-from-scratch.cpp
AgeCommit message (Expand)Author
2024-07-27Merge mainline llama.cpp (#3)Kawrakow
2024-06-05ggml : refactor rope norm/neox (#7634)Georgi Gerganov
2024-05-23ggml : remove ggml_flash_attn and ggml_flash_ff (#7463)Georgi Gerganov
2024-05-21llama : add phi3 128K model support (#7225)liuwei-git
2024-04-19train : add general name (#6752)Austin
2024-03-14gguf : fix resource leaks (#6061)Steve Grubb
2024-02-25code : normalize enum names (#5697)Georgi Gerganov
2024-02-18ggml, common, examples, tests : fixed type arguments in printf (#5528)Herman Semenov
2024-02-13finetune : rename feed-forward tensors (w1/w2/w3) (#4839)Daniel Bevenius
2024-02-12sync : ggml (#5452)Georgi Gerganov
2024-01-19finetune : fix ggml_allocr lifetimes (tmp workaround) (#5033)Uzo Nweke
2023-12-21ggml : change ggml_scale to take a float instead of tensor (#4573)Georgi Gerganov
2023-12-07train : fix #4227 (double free in examples/train-text-from-scratch/train-text...Hongyu Ouyang
2023-11-13sync : ggml (backend v2) (#3912)Georgi Gerganov
2023-11-01llama : implement YaRN RoPE scaling (#2268)cebtenzzre
2023-10-17train-text-from-scratch : fix assert failure in ggml-alloc (#3618)slaren
2023-09-29train : fix KQ_pos allocation (#3392)Georgi Gerganov
2023-09-28build : enable more non-default compiler warnings (#3200)Cebtenzzre
2023-09-28llama.cpp : split llama_context_params into model and context params (#3301)slaren
2023-09-28train : finetune LORA (#2632)xaedes
2023-09-28llama : custom attention mask + parallel decoding + no context swaps (#3228)Georgi Gerganov
2023-09-16Fixing the last deviations from sentencepiece indicated by test-tokenizer-1 (...goerch
2023-09-07fix some warnings from gcc and clang-tidy (#3038)Cebtenzzre
2023-08-28train : mem usage and other improvements (#2439)xaedes
2023-08-27llama : more tokenizer fixes (#2810)Georgi Gerganov
2023-08-22ggml : sync latest (SAM + SD operators, CUDA alibi) (#2709)Georgi Gerganov
2023-08-21gguf : new file format with flexible meta data (beta) (#2398)Georgi Gerganov
2023-07-25Add LLAMA_DEFAULT_RMS_EPS so we can change the default (#2384)Kawrakow
2023-07-24make rms_norm_eps a parameter (#2374)slaren
2023-07-21ggml : fix rope args order + assert (#2054)Georgi Gerganov
2023-07-11ggml : remove src0 and src1 from ggml_tensor and rename opt to src (#2178)Spencer Sutton
2023-07-07ggml : change ggml_graph_compute() API to not require context (#1999)Qingyou Meng
2023-07-01train : fix compile warningGeorgi Gerganov
2023-06-29Use unsigned for random seed (#2006)Howard Su
2023-06-27llama : fix rope usage after ChatGLM changeGeorgi Gerganov
2023-06-26ggml : increase max tensor name + clean up compiler warnings in train-text (#...David Yang
2023-06-24llama : make model stateless and context stateful (llama_state) (#1797)Didzis Gosko
2023-06-16build : fix and ignore MSVC warnings (#1889)Borislav Stanimirov
2023-06-13train : improved training-from-scratch example (#1652)xaedes