feat(alias): alias llama to llama-cpp, update docs (#1448)

Signed-off-by: Ettore Di Giacinto <mudler@users.noreply.github.com>
This commit is contained in:
Ettore Di Giacinto
2023-12-16 12:22:45 -05:00
committed by GitHub
parent 1c286c3c2f
commit 3d83128f16
3 changed files with 15 additions and 4 deletions
@@ -9,7 +9,7 @@ weight = 1
{{% notice note %}}
The `ggml` file format has been deprecated. If you are using `ggml` models and you are configuring your model with a YAML file, specify, use the `llama-stable` backend instead. If you are relying in automatic detection of the model, you should be fine. For `gguf` models, use the `llama` backend.
The `ggml` file format has been deprecated. If you are using `ggml` models and you are configuring your model with a YAML file, specify, use the `llama-ggml` backend instead. If you are relying in automatic detection of the model, you should be fine. For `gguf` models, use the `llama` backend. The go backend is deprecated as well but still available as `go-llama`. The go backend supports still features not available in the mainline: speculative sampling and embeddings.
{{% /notice %}}
@@ -65,11 +65,11 @@ parameters:
In the example above we specify `llama` as the backend to restrict loading `gguf` models only.
For instance, to use the `llama-stable` backend for `ggml` models:
For instance, to use the `llama-ggml` backend for `ggml` models:
```yaml
name: llama
backend: llama-stable
backend: llama-ggml
parameters:
# Relative to the models path
model: file.ggml.bin