summaryrefslogtreecommitdiff
path: root/examples/lookup
diff options
context:
space:
mode:
Diffstat (limited to 'examples/lookup')
-rw-r--r--examples/lookup/lookup-create.cpp4
-rw-r--r--examples/lookup/lookup-stats.cpp5
-rw-r--r--examples/lookup/lookup.cpp5
3 files changed, 3 insertions, 11 deletions
diff --git a/examples/lookup/lookup-create.cpp b/examples/lookup/lookup-create.cpp
index 46a6bed0..1c230c96 100644
--- a/examples/lookup/lookup-create.cpp
+++ b/examples/lookup/lookup-create.cpp
@@ -28,10 +28,8 @@ int main(int argc, char ** argv){
GGML_ASSERT(model != nullptr);
// tokenize the prompt
- const bool add_bos = llama_should_add_bos_token(model);
-
std::vector<llama_token> inp;
- inp = ::llama_tokenize(ctx, params.prompt, add_bos, true);
+ inp = ::llama_tokenize(ctx, params.prompt, true, true);
fprintf(stderr, "%s: tokenization done\n", __func__);
diff --git a/examples/lookup/lookup-stats.cpp b/examples/lookup/lookup-stats.cpp
index 31f22777..41b62c2f 100644
--- a/examples/lookup/lookup-stats.cpp
+++ b/examples/lookup/lookup-stats.cpp
@@ -34,11 +34,8 @@ int main(int argc, char ** argv){
GGML_ASSERT(llama_n_vocab(model) < (1 << 16));
// tokenize the prompt
- const bool add_bos = llama_should_add_bos_token(model);
- LOG("add_bos tgt: %d\n", add_bos);
-
std::vector<llama_token> inp;
- inp = ::llama_tokenize(ctx, params.prompt, add_bos, true);
+ inp = ::llama_tokenize(ctx, params.prompt, true, true);
llama_ngram_cache ngram_cache_context;
llama_ngram_cache ngram_cache_dynamic;
diff --git a/examples/lookup/lookup.cpp b/examples/lookup/lookup.cpp
index 2e8c35de..65ed408a 100644
--- a/examples/lookup/lookup.cpp
+++ b/examples/lookup/lookup.cpp
@@ -42,11 +42,8 @@ int main(int argc, char ** argv){
GGML_ASSERT(llama_n_vocab(model) < (1 << 16));
// tokenize the prompt
- const bool add_bos = llama_should_add_bos_token(model);
- LOG("add_bos tgt: %d\n", add_bos);
-
std::vector<llama_token> inp;
- inp = ::llama_tokenize(ctx, params.prompt, add_bos, true);
+ inp = ::llama_tokenize(ctx, params.prompt, true, true);
llama_ngram_cache ngram_cache_context;
llama_ngram_cache ngram_cache_dynamic;