From 04fbc5f23e9708136ff03ebe194c7a7e965a7ca4 Mon Sep 17 00:00:00 2001 From: Chao Jiang Date: Mon, 15 Apr 2024 00:16:34 +0800 Subject: Add Command R chat template (#6650) * Add chat template for command-r model series * Fix indentation * Add chat template test for command-r models and update the implementation to trim whitespaces * Remove debug print --- llama.cpp | 15 +++++++++++++++ 1 file changed, 15 insertions(+) (limited to 'llama.cpp') diff --git a/llama.cpp b/llama.cpp index b93c1abc..e3205e58 100644 --- a/llama.cpp +++ b/llama.cpp @@ -16625,6 +16625,21 @@ static int32_t llama_chat_apply_template_internal( if (add_ass) { ss << "### Response:\n"; } + } else if (tmpl == "command-r" || (tmpl.find("<|START_OF_TURN_TOKEN|>") != std::string::npos && tmpl.find("<|USER_TOKEN|>") != std::string::npos)) { + // CohereForAI/c4ai-command-r-plus + for (auto message : chat) { + std::string role(message->role); + if (role == "system") { + ss << "<|START_OF_TURN_TOKEN|><|SYSTEM_TOKEN|>" << trim(message->content) << "<|END_OF_TURN_TOKEN|>"; + } else if (role == "user") { + ss << "<|START_OF_TURN_TOKEN|><|USER_TOKEN|>" << trim(message->content) << "<|END_OF_TURN_TOKEN|>"; + } else if (role == "assistant") { + ss << "<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>" << trim(message->content) << "<|END_OF_TURN_TOKEN|>"; + } + } + if (add_ass) { + ss << "<|START_OF_TURN_TOKEN|><|CHATBOT_TOKEN|>"; + } } else { // template not supported return -1; -- cgit v1.2.3