From 791bc769c1f712c9832dcfcd6c10f73334d48eec Mon Sep 17 00:00:00 2001 From: Ettore Di Giacinto Date: Fri, 10 Oct 2025 09:51:22 +0200 Subject: [PATCH] chore(deps): bump llama.cpp to '1deee0f8d494981c32597dca8b5f8696d399b0f2' (#6421) Signed-off-by: Ettore Di Giacinto --- backend/cpp/llama-cpp/Makefile | 2 +- backend/cpp/llama-cpp/grpc-server.cpp | 12 ++++++------ 2 files changed, 7 insertions(+), 7 deletions(-) diff --git a/backend/cpp/llama-cpp/Makefile b/backend/cpp/llama-cpp/Makefile index 42baf5993..ab9f5b4c6 100644 --- a/backend/cpp/llama-cpp/Makefile +++ b/backend/cpp/llama-cpp/Makefile @@ -1,5 +1,5 @@ -LLAMA_VERSION?=9d0882840e6c3fb62965d03af0e22880ea90e012 +LLAMA_VERSION?=1deee0f8d494981c32597dca8b5f8696d399b0f2 LLAMA_REPO?=https://github.com/ggerganov/llama.cpp CMAKE_ARGS?= diff --git a/backend/cpp/llama-cpp/grpc-server.cpp b/backend/cpp/llama-cpp/grpc-server.cpp index 36565abf1..4d5ca6fb6 100644 --- a/backend/cpp/llama-cpp/grpc-server.cpp +++ b/backend/cpp/llama-cpp/grpc-server.cpp @@ -469,12 +469,12 @@ public: task.id = ctx_server.queue_tasks.get_new_id(); task.index = i; - task.prompt_tokens = std::move(inputs[i]); + task.tokens = std::move(inputs[i]); task.params = server_task::params_from_json_cmpl( ctx_server.ctx, ctx_server.params_base, data); - task.id_selected_slot = json_value(data, "id_slot", -1); + task.id_slot = json_value(data, "id_slot", -1); // OAI-compat task.params.oaicompat = OAICOMPAT_TYPE_NONE; @@ -624,12 +624,12 @@ public: task.id = ctx_server.queue_tasks.get_new_id(); task.index = i; - task.prompt_tokens = std::move(inputs[i]); + task.tokens = std::move(inputs[i]); task.params = server_task::params_from_json_cmpl( ctx_server.ctx, ctx_server.params_base, data); - task.id_selected_slot = json_value(data, "id_slot", -1); + task.id_slot = json_value(data, "id_slot", -1); // OAI-compat task.params.oaicompat = OAICOMPAT_TYPE_NONE; @@ -725,7 +725,7 @@ public: task.id = ctx_server.queue_tasks.get_new_id(); task.index = i; - task.prompt_tokens = std::move(tokenized_prompts[i]); + task.tokens = std::move(tokenized_prompts[i]); task.params.oaicompat = OAICOMPAT_TYPE_NONE; task.params.embd_normalize = embd_normalize; @@ -819,7 +819,7 @@ public: server_task task = server_task(SERVER_TASK_TYPE_RERANK); task.id = ctx_server.queue_tasks.get_new_id(); task.index = i; - task.prompt_tokens = std::move(tmp); + task.tokens = std::move(tmp); tasks.push_back(std::move(task)); }