From 9e1707218a24ff758c7b623594f8c0ce5e12eb6c Mon Sep 17 00:00:00 2001 From: Georgi Gerganov Date: Sun, 19 Mar 2023 18:37:02 +0200 Subject: Add "--instruct" argument for usage with Alpaca (#240) Also start adding prompts in "./prompts" --- utils.h | 8 ++++---- 1 file changed, 4 insertions(+), 4 deletions(-) (limited to 'utils.h') diff --git a/utils.h b/utils.h index c1a8498a..e329ba16 100644 --- a/utils.h +++ b/utils.h @@ -27,14 +27,14 @@ struct gpt_params { int32_t n_batch = 8; // batch size for prompt processing - std::string model = "models/lamma-7B/ggml-model.bin"; // model path - std::string prompt; + std::string model = "models/lamma-7B/ggml-model.bin"; // model path + std::string prompt = ""; + std::string antiprompt = ""; // string upon seeing which more user input is prompted bool use_color = false; // use color to distinguish generations and inputs bool interactive = false; // interactive mode - bool interactive_start = false; // reverse prompt immediately - std::string antiprompt = ""; // string upon seeing which more user input is prompted + bool instruct = false; // instruction mode (used for Alpaca models) }; bool gpt_params_parse(int argc, char ** argv, gpt_params & params); -- cgit v1.2.3