Skip to content
This repository was archived by the owner on Jul 4, 2025. It is now read-only.

Commit 2602b4f

Browse files
committed
enable cache prompt
1 parent 7c0be9f commit 2602b4f

File tree

1 file changed

+1
-1
lines changed

1 file changed

+1
-1
lines changed

controllers/llamaCPP.cc

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -164,7 +164,7 @@ void llamaCPP::chatCompletion(
164164
data["frequency_penalty"] =
165165
(*jsonBody).get("frequency_penalty", 0).asFloat();
166166
data["presence_penalty"] = (*jsonBody).get("presence_penalty", 0).asFloat();
167-
167+
data["cache_prompt"] = true;
168168
const Json::Value &messages = (*jsonBody)["messages"];
169169
for (const auto &message : messages) {
170170
std::string input_role = message["role"].asString();

0 commit comments

Comments
 (0)