summaryrefslogtreecommitdiff
path: root/examples/train-text-from-scratch/README.md
diff options
context:
space:
mode:
authorKawrakow <48489457+ikawrakow@users.noreply.github.com>2024-07-27 07:55:01 +0200
committerGitHub <noreply@github.com>2024-07-27 07:55:01 +0200
commit154e0d75fccf1784fe9ff6fd76a630b66563da3d (patch)
tree81ce6dbb5b1900c1aa78a879f0593c694cab9d27 /examples/train-text-from-scratch/README.md
parent0684c3e9c70d49323b4fc517128cbe222cab7f96 (diff)
Merge mainline llama.cpp (#3)
* Merging mainline - WIP * Merging mainline - WIP AVX2 and CUDA appear to work. CUDA performance seems slightly (~1-2%) lower as it is so often the case with llama.cpp/ggml after some "improvements" have been made. * Merging mainline - fix Metal * Remove check --------- Co-authored-by: Iwan Kawrakow <iwan.kawrakow@gmail.com>
Diffstat (limited to 'examples/train-text-from-scratch/README.md')
-rw-r--r--examples/train-text-from-scratch/README.md27
1 files changed, 0 insertions, 27 deletions
diff --git a/examples/train-text-from-scratch/README.md b/examples/train-text-from-scratch/README.md
deleted file mode 100644
index 3abae238..00000000
--- a/examples/train-text-from-scratch/README.md
+++ /dev/null
@@ -1,27 +0,0 @@
-# train-text-from-scratch
-
-Basic usage instructions:
-
-```bash
-# get training data
-wget https://raw.githubusercontent.com/brunoklein99/deep-learning-notes/master/shakespeare.txt
-
-# train
-./bin/llama-train-text-from-scratch \
- --vocab-model ../models/ggml-vocab-llama.gguf \
- --ctx 64 --embd 256 --head 8 --layer 16 \
- --checkpoint-in chk-shakespeare-256x16-LATEST.gguf \
- --checkpoint-out chk-shakespeare-256x16-ITERATION.gguf \
- --model-out ggml-shakespeare-256x16-f32-ITERATION.gguf \
- --train-data "shakespeare.txt" \
- -t 6 -b 16 --seed 1 --adam-iter 256 \
- --no-checkpointing
-
-# predict
-./bin/llama-cli -m ggml-shakespeare-256x16-f32.gguf
-```
-
-Output files will be saved every N iterations (config with `--save-every N`).
-The pattern "ITERATION" in the output filenames will be replaced with the iteration number and "LATEST" for the latest output.
-
-To train GGUF models just pass them to `--checkpoint-in FN`.