Ettore Di Giacinto
7a94d237c4
chore(deps): bump llama.cpp to '583cb83416467e8abf9b37349dcf1f6a0083745a ( #7358 )
...
chore(deps): bump llama.cpp to '583cb83416467e8abf9b37349dcf1f6a0083745a'
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-26 08:23:21 +01:00
dependabot[bot]
7e01aa8faa
chore(deps): bump protobuf from 6.32.0 to 6.33.1 in /backend/python/transformers ( #7340 )
...
chore(deps): bump protobuf in /backend/python/transformers
Bumps [protobuf](https://github.com/protocolbuffers/protobuf ) from 6.32.0 to 6.33.1.
- [Release notes](https://github.com/protocolbuffers/protobuf/releases )
- [Changelog](https://github.com/protocolbuffers/protobuf/blob/main/protobuf_release.bzl )
- [Commits](https://github.com/protocolbuffers/protobuf/commits )
---
updated-dependencies:
- dependency-name: protobuf
dependency-version: 6.33.1
dependency-type: direct:production
update-type: version-update:semver-minor
...
Signed-off-by: dependabot[bot] <support@github.com >
Co-authored-by: dependabot[bot] <49699333+dependabot[bot]@users.noreply.github.com>
2025-11-24 20:12:17 +00:00
LocalAI [bot]
f6d2a52cd5
chore: ⬆️ Update ggml-org/llama.cpp to 0c7220db56525d40177fcce3baa0d083448ec813 ( #7337 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-24 09:11:38 +01:00
LocalAI [bot]
05a00b2399
chore: ⬆️ Update ggml-org/llama.cpp to 3f3a4fb9c3b907c68598363b204e6f58f4757c8c ( #7336 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-22 21:53:40 +00:00
Ettore Di Giacinto
3a232446e0
Revert "chore(chatterbox): bump l4t index to support more recent pytorch" ( #7333 )
...
Revert "chore(chatterbox): bump l4t index to support more recent pytorch (#7332 )"
This reverts commit 55607a5aac .
2025-11-22 10:10:27 +01:00
LocalAI [bot]
bdfe8431fa
chore: ⬆️ Update ggml-org/llama.cpp to 23bc779a6e58762ea892eca1801b2ea1b9050c00 ( #7331 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-22 08:44:01 +01:00
Ettore Di Giacinto
55607a5aac
chore(chatterbox): bump l4t index to support more recent pytorch ( #7332 )
...
This should add support for devices like the DGX Spark
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-11-21 22:24:46 +01:00
Ettore Di Giacinto
ec492a4c56
fix(typo): environment variable name for max jobs
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-11-21 18:37:22 +01:00
Ettore Di Giacinto
2defe98df8
fix(vllm): Update flash-attn to specific wheel URL
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-11-21 18:06:46 +01:00
Ettore Di Giacinto
6261c87b1b
Add NVCC_THREADS and MAX_JOB environment variables
...
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-11-21 16:14:13 +01:00
Ettore Di Giacinto
e88db7d142
fix(llama.cpp): handle corner cases with tool content ( #7324 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-21 09:21:49 +01:00
LocalAI [bot]
b7b8a0a748
chore: ⬆️ Update ggml-org/llama.cpp to dd0f3219419b24740864b5343958a97e1b3e4b26 ( #7322 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-21 08:11:47 +01:00
LocalAI [bot]
b8011f49f2
chore: ⬆️ Update ggml-org/whisper.cpp to 19ceec8eac980403b714d603e5ca31653cd42a3f ( #7321 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-20 23:07:22 +01:00
Ettore Di Giacinto
daf39e1efd
chore(vllm/ci): set maximum number of jobs
...
Also added comments to clarify CPU usage during build.
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-11-20 15:53:32 +01:00
LocalAI [bot]
bfa07df7cd
chore: ⬆️ Update ggml-org/llama.cpp to 7d77f07325985c03a91fa371d0a68ef88a91ec7f ( #7314 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-20 07:58:42 +01:00
Ettore Di Giacinto
3152611184
chore(deps): bump llama.cpp to '10e9780154365b191fb43ca4830659ef12def80f ( #7311 )
...
chore(deps): bump llama.cpp to '10e9780154365b191fb43ca4830659ef12def80f'
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-19 14:42:11 +01:00
LocalAI [bot]
4278506876
chore: ⬆️ Update ggml-org/llama.cpp to cb623de3fc61011e5062522b4d05721a22f2e916 ( #7301 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-18 07:43:57 +01:00
LocalAI [bot]
1dd1d12da1
chore: ⬆️ Update ggml-org/whisper.cpp to b12abefa9be2abae39a73fa903322af135024a36 ( #7300 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-18 07:43:33 +01:00
LocalAI [bot]
fb834805db
chore: ⬆️ Update ggml-org/llama.cpp to 80deff3648b93727422461c41c7279ef1dac7452 ( #7287 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-17 07:51:08 +01:00
Ettore Di Giacinto
d7f9f3ac93
feat: add support to logitbias and logprobs ( #7283 )
...
* feat: add support to logprobs in results
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat: add support to logitbias
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-16 13:27:36 +01:00
LocalAI [bot]
d1a0dd10e6
chore: ⬆️ Update ggml-org/llama.cpp to 662192e1dcd224bc25759aadd0190577524c6a66 ( #7277 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-16 08:41:12 +01:00
LocalAI [bot]
a09d49da43
chore: ⬆️ Update ggml-org/llama.cpp to 9b17d74ab7d31cb7d15ee7eec1616c3d825a84c0 ( #7273 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-15 00:05:39 +01:00
Ettore Di Giacinto
03e9f4b140
fix: handle tool errors ( #7271 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-14 17:23:56 +01:00
Ettore Di Giacinto
7129409bf6
chore(deps): bump llama.cpp to c4abcb2457217198efdd67d02675f5fddb7071c2 ( #7266 )
...
* chore(deps): bump llama.cpp to '92bb442ad999a0d52df0af2730cd861012e8ac5c'
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* DEBUG
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Bump
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* test/debug
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Revert "DEBUG"
This reverts commit 2501ca3ff2 .
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-14 12:16:52 +01:00
LocalAI [bot]
d9e9ec6825
chore: ⬆️ Update ggml-org/whisper.cpp to d9b7613b34a343848af572cc14467fc5e82fc788 ( #7268 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-13 23:05:06 +01:00
Ettore Di Giacinto
3728552e94
feat: import models via URI ( #7245 )
...
* feat: initial hook to install elements directly
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* WIP: ui changes
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Move HF api client to pkg
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add simple importer for gguf files
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add opcache
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* wire importers to CLI
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add omitempty to config fields
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fix tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add MLX importer
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Small refactors to star to use HF for discovery
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add tests
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Common preferences
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Add support to bare HF repos
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(importer/llama.cpp): add support for mmproj files
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* add mmproj quants to common preferences
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fix vlm usage in tokenizer mode with llama.cpp
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-12 20:48:56 +01:00
Mikhail Khludnev
04fe0b0da8
fix(reranker): llama-cpp sort score desc, crop top_n ( #7211 )
...
Signed-off-by: Mikhail Khludnev <mkhl@apache.org >
2025-11-12 09:13:01 +01:00
LocalAI [bot]
fae93e5ba2
chore: ⬆️ Update ggml-org/llama.cpp to 7d019cff744b73084b15ca81ba9916f3efab1223 ( #7247 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-11 21:31:01 +00:00
LocalAI [bot]
5f4663252d
chore: ⬆️ Update ggml-org/llama.cpp to 13730c183b9e1a32c09bf132b5367697d6c55048 ( #7232 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-11 00:03:01 +01:00
LocalAI [bot]
80bb7c5f67
chore: ⬆️ Update ggml-org/whisper.cpp to a1867e0dad0b21b35afa43fc815dae60c9a139d6 ( #7231 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-10 22:51:49 +01:00
LocalAI [bot]
e42f0f7e79
chore: ⬆️ Update ggml-org/llama.cpp to b8595b16e69e3029e06be3b8f6635f9812b2bc3f ( #7210 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-09 23:56:27 +01:00
Mikhail Khludnev
01cd58a739
fix(reranker): support omitting top_n ( #7199 )
...
* fix(reranker): support omitting top_n
Signed-off-by: Mikhail Khludnev <mkhl@apache.org >
* fix(reranker): support omitting top_n
Signed-off-by: Mikhail Khludnev <mkhl@apache.org >
* pass 0 explicitly
Signed-off-by: Mikhail Khludnev <mkhludnev@users.noreply.github.com >
---------
Signed-off-by: Mikhail Khludnev <mkhl@apache.org >
Signed-off-by: Mikhail Khludnev <mkhludnev@users.noreply.github.com >
2025-11-09 18:40:32 +01:00
Ettore Di Giacinto
679d43c2f5
feat: respect context and add request cancellation ( #7187 )
...
* feat: respect context
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* workaround fasthttp
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(ui): allow to abort call
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Refactor
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* chore: improving error
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Respect context also with MCP
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Tie to both contexts
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Make detection more robust
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-09 18:19:19 +01:00
LocalAI [bot]
f678c6b0a9
chore: ⬆️ Update ggml-org/llama.cpp to 333f2595a3e0e4c0abf233f2f29ef1710acd134d ( #7201 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-08 21:06:17 +00:00
Ettore Di Giacinto
2f2f9beee7
fix(chatterbox): pin numpy ( #7198 )
...
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-08 16:52:22 +01:00
LocalAI [bot]
8ac7e28c12
chore: ⬆️ Update ggml-org/llama.cpp to 65156105069fa86a4a81b6cb0e8cb583f6420677 ( #7184 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-08 09:07:44 +01:00
Ettore Di Giacinto
02cc8cbcaa
feat(llama.cpp): consolidate options and respect tokenizer template when enabled ( #7120 )
...
* feat(llama.cpp): expose env vars as options for consistency
This allows to configure everything in the YAML file of the model rather
than have global configurations
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* feat(llama.cpp): respect usetokenizertemplate and use llama.cpp templating system to process messages
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* WIP
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Detect template exists if use tokenizer template is enabled
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Better recognization of chat
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fixes to support tool calls while using templates from tokenizer
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Fixups
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Drop template guessing, fix passing tools to tokenizer
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Extract grammar and other options from chat template, add schema struct
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* WIP
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* WIP
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Automatically set use_jinja
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Cleanups, identify by default gguf models for chat
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Update docs
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
2025-11-07 21:23:50 +01:00
LocalAI [bot]
8f7c499f17
chore: ⬆️ Update ggml-org/llama.cpp to 7f09a680af6e0ef612de81018e1d19c19b8651e8 ( #7156 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-07 08:38:56 +01:00
Mikhail Khludnev
122e4c7094
fix(reranker): reproduce ignoring top_n ( #7025 )
...
* fix(reranker): reproduce ignoring top_n
Signed-off-by: Mikhail Khludnev <mkhl@apache.org >
* fix(reranker): ignoring top_n
Signed-off-by: Mikhail Khludnev <mkhl@apache.org >
---------
Signed-off-by: Mikhail Khludnev <mkhl@apache.org >
2025-11-06 10:03:05 +00:00
LocalAI [bot]
db9957b94e
chore: ⬆️ Update ggml-org/llama.cpp to a44d77126c911d105f7f800c17da21b2a5b112d1 ( #7125 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-05 21:22:04 +00:00
LocalAI [bot]
98158881c2
chore: ⬆️ Update ggml-org/llama.cpp to ad51c0a720062a04349c779aae301ad65ca4c856 ( #7098 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-04 21:19:58 +00:00
LocalAI [bot]
e2cb44ef37
chore: ⬆️ Update ggml-org/llama.cpp to c5023daf607c578d6344c628eb7da18ac3d92d32 ( #7069 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-04 09:26:10 +01:00
LocalAI [bot]
2cad2c8591
chore: ⬆️ Update ggml-org/llama.cpp to cd5e3b57541ecc52421130742f4d89acbcf77cd4 ( #7023 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-02 21:24:19 +00:00
Ettore Di Giacinto
424acd66ad
feat(llama.cpp): allow to set cache-ram and ctx_shift ( #7009 )
...
* feat(llama.cpp): allow to set cache-ram and ctx_shift
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
* Apply suggestion from @mudler
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
---------
Signed-off-by: Ettore Di Giacinto <mudler@localai.io >
Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com >
2025-11-02 17:33:29 +01:00
LocalAI [bot]
3cd8234550
chore: ⬆️ Update ggml-org/whisper.cpp to 999a7e0cbf8484dc2cea1e9f855d6b39f34f7ae9 ( #6997 )
...
⬆️ Update ggml-org/whisper.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-02 10:39:11 +01:00
LocalAI [bot]
f85e2dd1b8
chore: ⬆️ Update ggml-org/llama.cpp to 2f68ce7cfd20e9e7098514bf730e5389b7bba908 ( #6998 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-11-02 09:44:37 +01:00
LocalAI [bot]
9ecfdc5938
chore: ⬆️ Update ggml-org/llama.cpp to 31c511a968348281e11d590446bb815048a1e912 ( #6970 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-10-31 21:04:53 +00:00
LocalAI [bot]
0ddb2e8dcf
chore: ⬆️ Update ggml-org/llama.cpp to 4146d6a1a6228711a487a1e3e9ddd120f8d027d7 ( #6945 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-10-31 14:51:03 +00:00
LocalAI [bot]
1e5b9135df
chore: ⬆️ Update ggml-org/llama.cpp to 16724b5b6836a2d4b8936a5824d2ff27c52b4517 ( #6925 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-10-30 21:07:33 +00:00
LocalAI [bot]
dd21a0d2f9
chore: ⬆️ Update ggml-org/llama.cpp to 3464bdac37027c5e9661621fc75ffcef3c19c6ef ( #6896 )
...
⬆️ Update ggml-org/llama.cpp
Signed-off-by: github-actions[bot] <41898282+github-actions[bot]@users.noreply.github.com>
Co-authored-by: mudler <2420543+mudler@users.noreply.github.com >
2025-10-30 14:17:58 +01:00