mirror of
https://github.com/mudler/LocalAI.git
synced 2026-01-16 23:49:44 -06:00
chore(deps): bump llama.cpp to '0e1ccf15c7b6d05c720551b537857ecf6194d420' (#7684)
Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
This commit is contained in:
committed by
GitHub
parent
8b3e0ebf8a
commit
fc6057a952
@@ -78,9 +78,6 @@ static void start_llama_server(server_context& ctx_server) {
|
||||
std::this_thread::sleep_for(std::chrono::milliseconds(100));
|
||||
}
|
||||
|
||||
ctx_server.init();
|
||||
//state.store(SERVER_STATE_READY);
|
||||
|
||||
LOG_INF("%s: model loaded\n", __func__);
|
||||
|
||||
// print sample chat example to make it clear which template is used
|
||||
|
||||
Reference in New Issue
Block a user