mirror of
https://github.com/mudler/LocalAI.git
synced 2025-12-30 22:20:20 -06:00
800 lines
28 KiB
Makefile
800 lines
28 KiB
Makefile
GOCMD=go
|
|
GOTEST=$(GOCMD) test
|
|
GOVET=$(GOCMD) vet
|
|
BINARY_NAME=local-ai
|
|
|
|
DETECT_LIBS?=true
|
|
|
|
# whisper.cpp version
|
|
WHISPER_REPO?=https://github.com/ggml-org/whisper.cpp
|
|
WHISPER_CPP_VERSION?=032697b9a850dc2615555e2a93a683cc3dd58559
|
|
|
|
# stablediffusion.cpp (ggml)
|
|
STABLEDIFFUSION_GGML_REPO?=https://github.com/richiejp/stable-diffusion.cpp
|
|
STABLEDIFFUSION_GGML_VERSION?=53e3b17eb3d0b5760ced06a1f98320b68b34aaae
|
|
|
|
# ONEAPI variables for SYCL
|
|
export ONEAPI_VARS?=/opt/intel/oneapi/setvars.sh
|
|
ONEAPI_VERSION=2025.1
|
|
|
|
ONNX_VERSION?=1.20.0
|
|
ONNX_ARCH?=x64
|
|
ONNX_OS?=linux
|
|
|
|
export BUILD_TYPE?=
|
|
export STABLE_BUILD_TYPE?=$(BUILD_TYPE)
|
|
export CMAKE_ARGS?=-DBUILD_SHARED_LIBS=OFF
|
|
export WHISPER_CMAKE_ARGS?=-DBUILD_SHARED_LIBS=OFF
|
|
export BACKEND_LIBS?=
|
|
export WHISPER_DIR=$(abspath ./sources/whisper.cpp)
|
|
export WHISPER_INCLUDE_PATH=$(WHISPER_DIR)/include:$(WHISPER_DIR)/ggml/include
|
|
export WHISPER_LIBRARY_PATH=$(WHISPER_DIR)/build/src/:$(WHISPER_DIR)/build/ggml/src
|
|
|
|
CGO_LDFLAGS?=
|
|
CGO_LDFLAGS_WHISPER?=
|
|
CGO_LDFLAGS_WHISPER+=-lggml
|
|
CUDA_LIBPATH?=/usr/local/cuda/lib64/
|
|
GO_TAGS?=
|
|
BUILD_ID?=
|
|
NATIVE?=false
|
|
|
|
TEST_DIR=/tmp/test
|
|
|
|
TEST_FLAKES?=5
|
|
|
|
RANDOM := $(shell bash -c 'echo $$RANDOM')
|
|
|
|
VERSION?=$(shell git describe --always --tags || echo "dev" )
|
|
# go tool nm ./local-ai | grep Commit
|
|
LD_FLAGS?=-s -w
|
|
override LD_FLAGS += -X "github.com/mudler/LocalAI/internal.Version=$(VERSION)"
|
|
override LD_FLAGS += -X "github.com/mudler/LocalAI/internal.Commit=$(shell git rev-parse HEAD)"
|
|
|
|
OPTIONAL_TARGETS?=
|
|
|
|
export OS := $(shell uname -s)
|
|
ARCH := $(shell uname -m)
|
|
GREEN := $(shell tput -Txterm setaf 2)
|
|
YELLOW := $(shell tput -Txterm setaf 3)
|
|
WHITE := $(shell tput -Txterm setaf 7)
|
|
CYAN := $(shell tput -Txterm setaf 6)
|
|
RESET := $(shell tput -Txterm sgr0)
|
|
|
|
UPX?=
|
|
# check if upx exists
|
|
ifeq (, $(shell which upx))
|
|
UPX=
|
|
else
|
|
UPX=$(shell which upx)
|
|
endif
|
|
|
|
# Default Docker bridge IP
|
|
E2E_BRIDGE_IP?=172.17.0.1
|
|
|
|
ifndef UNAME_S
|
|
UNAME_S := $(shell uname -s)
|
|
endif
|
|
|
|
# IF native is false, we add -DGGML_NATIVE=OFF to CMAKE_ARGS
|
|
ifeq ($(NATIVE),false)
|
|
CMAKE_ARGS+=-DGGML_NATIVE=OFF
|
|
WHISPER_CMAKE_ARGS+=-DGGML_NATIVE=OFF
|
|
endif
|
|
|
|
# Detect if we are running on arm64
|
|
ifneq (,$(findstring aarch64,$(shell uname -m)))
|
|
ONNX_ARCH=aarch64
|
|
endif
|
|
|
|
ifeq ($(OS),Darwin)
|
|
ONNX_OS=osx
|
|
ifneq (,$(findstring aarch64,$(shell uname -m)))
|
|
ONNX_ARCH=arm64
|
|
else ifneq (,$(findstring arm64,$(shell uname -m)))
|
|
ONNX_ARCH=arm64
|
|
else
|
|
ONNX_ARCH=x86_64
|
|
endif
|
|
|
|
ifeq ($(OSX_SIGNING_IDENTITY),)
|
|
OSX_SIGNING_IDENTITY := $(shell security find-identity -v -p codesigning | grep '"' | head -n 1 | sed -E 's/.*"(.*)"/\1/')
|
|
endif
|
|
|
|
# on OSX, if BUILD_TYPE is blank, we should default to use Metal
|
|
ifeq ($(BUILD_TYPE),)
|
|
BUILD_TYPE=metal
|
|
# disable metal if on Darwin and any other value is explicitly passed.
|
|
else ifneq ($(BUILD_TYPE),metal)
|
|
CMAKE_ARGS+=-DGGML_METAL=OFF
|
|
WHISPER_CMAKE_ARGS+=-DGGML_METAL=OFF
|
|
export GGML_NO_ACCELERATE=1
|
|
export GGML_NO_METAL=1
|
|
GO_LDFLAGS_WHISPER+=-lggml-blas
|
|
export WHISPER_LIBRARY_PATH:=$(WHISPER_LIBRARY_PATH):$(WHISPER_DIR)/build/ggml/src/ggml-blas
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),metal)
|
|
CGO_LDFLAGS += -framework Accelerate
|
|
CGO_LDFLAGS_WHISPER+=-lggml-metal -lggml-blas
|
|
CMAKE_ARGS+=-DGGML_METAL=ON
|
|
CMAKE_ARGS+=-DGGML_METAL_USE_BF16=ON
|
|
CMAKE_ARGS+=-DGGML_METAL_EMBED_LIBRARY=ON
|
|
CMAKE_ARGS+=-DGGML_OPENMP=OFF
|
|
WHISPER_CMAKE_ARGS+=-DGGML_METAL=ON
|
|
WHISPER_CMAKE_ARGS+=-DGGML_METAL_USE_BF16=ON
|
|
WHISPER_CMAKE_ARGS+=-DGGML_METAL_EMBED_LIBRARY=ON
|
|
WHISPER_CMAKE_ARGS+=-DWHISPER_BUILD_EXAMPLES=OFF
|
|
WHISPER_CMAKE_ARGS+=-DWHISPER_BUILD_TESTS=OFF
|
|
WHISPER_CMAKE_ARGS+=-DWHISPER_BUILD_SERVER=OFF
|
|
WHISPER_CMAKE_ARGS+=-DGGML_OPENMP=OFF
|
|
export WHISPER_LIBRARY_PATH:=$(WHISPER_LIBRARY_PATH):$(WHISPER_DIR)/build/ggml/src/ggml-metal/:$(WHISPER_DIR)/build/ggml/src/ggml-blas
|
|
else
|
|
CGO_LDFLAGS_WHISPER+=-lggml-blas
|
|
export WHISPER_LIBRARY_PATH:=$(WHISPER_LIBRARY_PATH):$(WHISPER_DIR)/build/ggml/src/ggml-blas
|
|
endif
|
|
else
|
|
CGO_LDFLAGS_WHISPER+=-lgomp
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),openblas)
|
|
CGO_LDFLAGS+=-lopenblas
|
|
export GGML_OPENBLAS=1
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),cublas)
|
|
CGO_LDFLAGS+=-lcublas -lcudart -L$(CUDA_LIBPATH) -L$(CUDA_LIBPATH)/stubs/ -lcuda
|
|
export GGML_CUDA=1
|
|
CMAKE_ARGS+=-DGGML_CUDA=ON
|
|
WHISPER_CMAKE_ARGS+=-DGGML_CUDA=ON
|
|
CGO_LDFLAGS_WHISPER+=-lcufft -lggml-cuda
|
|
export WHISPER_LIBRARY_PATH:=$(WHISPER_LIBRARY_PATH):$(WHISPER_DIR)/build/ggml/src/ggml-cuda/
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),vulkan)
|
|
CMAKE_ARGS+=-DGGML_VULKAN=1
|
|
WHISPER_CMAKE_ARGS+=-DGGML_VULKAN=1
|
|
CGO_LDFLAGS_WHISPER+=-lggml-vulkan -lvulkan
|
|
export WHISPER_LIBRARY_PATH:=$(WHISPER_LIBRARY_PATH):$(WHISPER_DIR)/build/ggml/src/ggml-vulkan/
|
|
endif
|
|
|
|
ifneq (,$(findstring sycl,$(BUILD_TYPE)))
|
|
export GGML_SYCL=1
|
|
CMAKE_ARGS+=-DGGML_SYCL=ON
|
|
WHISPER_CMAKE_ARGS+=-DGGML_SYCL=ON -DCMAKE_C_COMPILER=icx -DCMAKE_CXX_COMPILER=icpx
|
|
export CC=icx
|
|
export CXX=icpx
|
|
CGO_LDFLAGS_WHISPER += -fsycl -L${DNNLROOT}/lib -rpath ${ONEAPI_ROOT}/${ONEAPI_VERSION}/lib -ldnnl ${MKLROOT}/lib/intel64/libmkl_sycl.a -fiopenmp -fopenmp-targets=spir64 -lOpenCL -lggml-sycl
|
|
CGO_LDFLAGS_WHISPER += $(shell pkg-config --libs mkl-static-lp64-gomp)
|
|
CGO_CXXFLAGS_WHISPER += -fiopenmp -fopenmp-targets=spir64
|
|
CGO_CXXFLAGS_WHISPER += $(shell pkg-config --cflags mkl-static-lp64-gomp )
|
|
export WHISPER_LIBRARY_PATH:=$(WHISPER_LIBRARY_PATH):$(WHISPER_DIR)/build/ggml/src/ggml-sycl/
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),sycl_f16)
|
|
export GGML_SYCL_F16=1
|
|
CMAKE_ARGS+=-DGGML_SYCL_F16=ON
|
|
WHISPER_CMAKE_ARGS+=-DGGML_SYCL_F16=ON
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),hipblas)
|
|
ROCM_HOME ?= /opt/rocm
|
|
ROCM_PATH ?= /opt/rocm
|
|
LD_LIBRARY_PATH ?= /opt/rocm/lib:/opt/rocm/llvm/lib
|
|
export CXX=$(ROCM_HOME)/llvm/bin/clang++
|
|
export CC=$(ROCM_HOME)/llvm/bin/clang
|
|
export STABLE_BUILD_TYPE=
|
|
export GGML_HIP=1
|
|
GPU_TARGETS ?= gfx803,gfx900,gfx906,gfx908,gfx90a,gfx942,gfx1010,gfx1030,gfx1032,gfx1100,gfx1101,gfx1102
|
|
AMDGPU_TARGETS ?= "$(GPU_TARGETS)"
|
|
CMAKE_ARGS+=-DGGML_HIP=ON -DAMDGPU_TARGETS="$(AMDGPU_TARGETS)" -DGPU_TARGETS="$(GPU_TARGETS)"
|
|
CGO_LDFLAGS += -O3 --rtlib=compiler-rt -unwindlib=libgcc -lhipblas -lrocblas --hip-link -L${ROCM_HOME}/lib/llvm/lib
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),metal)
|
|
CGO_LDFLAGS+=-framework Foundation -framework Metal -framework MetalKit -framework MetalPerformanceShaders
|
|
export GGML_METAL=1
|
|
endif
|
|
|
|
ifeq ($(BUILD_TYPE),clblas)
|
|
CGO_LDFLAGS+=-lOpenCL -lclblast
|
|
export GGML_OPENBLAS=1
|
|
endif
|
|
|
|
# glibc-static or glibc-devel-static required
|
|
ifeq ($(STATIC),true)
|
|
LD_FLAGS+=-linkmode external -extldflags -static
|
|
endif
|
|
|
|
ALL_GRPC_BACKENDS=backend-assets/grpc/huggingface
|
|
ALL_GRPC_BACKENDS+=backend-assets/grpc/whisper
|
|
|
|
ifeq ($(ONNX_OS),linux)
|
|
ifeq ($(ONNX_ARCH),x64)
|
|
ALL_GRPC_BACKENDS+=backend-assets/grpc/stablediffusion-ggml
|
|
endif
|
|
endif
|
|
|
|
ALL_GRPC_BACKENDS+=backend-assets/grpc/local-store
|
|
ALL_GRPC_BACKENDS+=backend-assets/grpc/silero-vad
|
|
ALL_GRPC_BACKENDS+=$(OPTIONAL_GRPC)
|
|
# Use filter-out to remove the specified backends
|
|
ALL_GRPC_BACKENDS := $(filter-out $(SKIP_GRPC_BACKEND),$(ALL_GRPC_BACKENDS))
|
|
|
|
GRPC_BACKENDS?=$(ALL_GRPC_BACKENDS) $(OPTIONAL_GRPC)
|
|
TEST_PATHS?=./api/... ./pkg/... ./core/...
|
|
|
|
# If empty, then we build all
|
|
ifeq ($(GRPC_BACKENDS),)
|
|
GRPC_BACKENDS=$(ALL_GRPC_BACKENDS)
|
|
endif
|
|
|
|
ifeq ($(BUILD_API_ONLY),true)
|
|
GRPC_BACKENDS=
|
|
endif
|
|
|
|
.PHONY: all test build vendor get-sources prepare-sources prepare
|
|
|
|
all: help
|
|
|
|
## stablediffusion (ggml)
|
|
sources/stablediffusion-ggml.cpp:
|
|
git clone --recursive $(STABLEDIFFUSION_GGML_REPO) sources/stablediffusion-ggml.cpp && \
|
|
cd sources/stablediffusion-ggml.cpp && \
|
|
git checkout $(STABLEDIFFUSION_GGML_VERSION) && \
|
|
git submodule update --init --recursive --depth 1 --single-branch
|
|
|
|
backend/go/image/stablediffusion-ggml/libsd.a: sources/stablediffusion-ggml.cpp
|
|
$(MAKE) -C backend/go/image/stablediffusion-ggml build/libstable-diffusion.a
|
|
$(MAKE) -C backend/go/image/stablediffusion-ggml libsd.a
|
|
|
|
backend-assets/grpc/stablediffusion-ggml: backend/go/image/stablediffusion-ggml/libsd.a backend-assets/grpc
|
|
$(MAKE) -C backend/go/image/stablediffusion-ggml CGO_LDFLAGS="$(CGO_LDFLAGS)" stablediffusion-ggml
|
|
|
|
sources/onnxruntime:
|
|
mkdir -p sources/onnxruntime
|
|
curl -L https://github.com/microsoft/onnxruntime/releases/download/v$(ONNX_VERSION)/onnxruntime-$(ONNX_OS)-$(ONNX_ARCH)-$(ONNX_VERSION).tgz -o sources/onnxruntime/onnxruntime-$(ONNX_OS)-$(ONNX_ARCH)-$(ONNX_VERSION).tgz
|
|
cd sources/onnxruntime && tar -xvf onnxruntime-$(ONNX_OS)-$(ONNX_ARCH)-$(ONNX_VERSION).tgz && rm onnxruntime-$(ONNX_OS)-$(ONNX_ARCH)-$(ONNX_VERSION).tgz
|
|
cd sources/onnxruntime && mv onnxruntime-$(ONNX_OS)-$(ONNX_ARCH)-$(ONNX_VERSION)/* ./
|
|
|
|
backend-assets/lib/libonnxruntime.so.1: backend-assets/lib sources/onnxruntime
|
|
cp -rfv sources/onnxruntime/lib/* backend-assets/lib/
|
|
ifeq ($(OS),Darwin)
|
|
mv backend-assets/lib/libonnxruntime.$(ONNX_VERSION).dylib backend-assets/lib/libonnxruntime.dylib
|
|
else
|
|
mv backend-assets/lib/libonnxruntime.so.$(ONNX_VERSION) backend-assets/lib/libonnxruntime.so.1
|
|
endif
|
|
|
|
## whisper
|
|
sources/whisper.cpp:
|
|
mkdir -p sources/whisper.cpp
|
|
cd sources/whisper.cpp && \
|
|
git init && \
|
|
git remote add origin $(WHISPER_REPO) && \
|
|
git fetch origin && \
|
|
git checkout $(WHISPER_CPP_VERSION) && \
|
|
git submodule update --init --recursive --depth 1 --single-branch
|
|
|
|
sources/whisper.cpp/build/src/libwhisper.a: sources/whisper.cpp
|
|
cd sources/whisper.cpp && cmake $(WHISPER_CMAKE_ARGS) . -B ./build
|
|
cd sources/whisper.cpp/build && cmake --build . --config Release
|
|
|
|
get-sources: sources/stablediffusion-ggml.cpp sources/whisper.cpp
|
|
|
|
replace:
|
|
$(GOCMD) mod edit -replace github.com/ggerganov/whisper.cpp=$(CURDIR)/sources/whisper.cpp
|
|
$(GOCMD) mod edit -replace github.com/ggerganov/whisper.cpp/bindings/go=$(CURDIR)/sources/whisper.cpp/bindings/go
|
|
|
|
dropreplace:
|
|
$(GOCMD) mod edit -dropreplace github.com/ggerganov/whisper.cpp
|
|
$(GOCMD) mod edit -dropreplace github.com/ggerganov/whisper.cpp/bindings/go
|
|
|
|
prepare-sources: get-sources replace
|
|
$(GOCMD) mod download
|
|
|
|
## GENERIC
|
|
rebuild: ## Rebuilds the project
|
|
$(GOCMD) clean -cache
|
|
$(MAKE) -C sources/whisper.cpp clean
|
|
$(MAKE) build
|
|
|
|
prepare: prepare-sources $(OPTIONAL_TARGETS)
|
|
|
|
clean: ## Remove build related file
|
|
$(GOCMD) clean -cache
|
|
rm -f prepare
|
|
rm -rf ./sources
|
|
rm -rf $(BINARY_NAME)
|
|
rm -rf release/
|
|
rm -rf backend-assets/*
|
|
$(MAKE) -C backend/cpp/grpc clean
|
|
$(MAKE) -C backend/go/image/stablediffusion-ggml clean
|
|
$(MAKE) dropreplace
|
|
$(MAKE) protogen-clean
|
|
rmdir pkg/grpc/proto || true
|
|
|
|
clean-tests:
|
|
rm -rf test-models
|
|
rm -rf test-dir
|
|
rm -rf core/http/backend-assets
|
|
|
|
clean-dc: clean
|
|
cp -r /build/backend-assets /workspace/backend-assets
|
|
|
|
## Install Go tools
|
|
install-go-tools:
|
|
go install google.golang.org/grpc/cmd/protoc-gen-go-grpc@1958fcbe2ca8bd93af633f11e97d44e567e945af
|
|
go install google.golang.org/protobuf/cmd/protoc-gen-go@v1.34.2
|
|
go install github.com/GeertJohan/go.rice/rice@latest
|
|
|
|
## Build:
|
|
build: prepare backend-assets grpcs install-go-tools ## Build the project
|
|
$(info ${GREEN}I local-ai build info:${RESET})
|
|
$(info ${GREEN}I BUILD_TYPE: ${YELLOW}$(BUILD_TYPE)${RESET})
|
|
$(info ${GREEN}I GO_TAGS: ${YELLOW}$(GO_TAGS)${RESET})
|
|
$(info ${GREEN}I LD_FLAGS: ${YELLOW}$(LD_FLAGS)${RESET})
|
|
$(info ${GREEN}I UPX: ${YELLOW}$(UPX)${RESET})
|
|
ifneq ($(BACKEND_LIBS),)
|
|
$(MAKE) backend-assets/lib
|
|
cp -f $(BACKEND_LIBS) backend-assets/lib/
|
|
endif
|
|
rm -rf $(BINARY_NAME) || true
|
|
CGO_LDFLAGS="$(CGO_LDFLAGS)" $(GOCMD) build -ldflags "$(LD_FLAGS)" -tags "$(GO_TAGS)" -o $(BINARY_NAME) ./
|
|
rice append --exec $(BINARY_NAME)
|
|
|
|
build-api:
|
|
BUILD_GRPC_FOR_BACKEND_LLAMA=true BUILD_API_ONLY=true GO_TAGS=p2p $(MAKE) build
|
|
|
|
backend-assets/lib:
|
|
mkdir -p backend-assets/lib
|
|
|
|
dist:
|
|
GO_TAGS="p2p" $(MAKE) build
|
|
GO_TAGS="p2p" STATIC=true $(MAKE) build
|
|
mkdir -p release
|
|
# if BUILD_ID is empty, then we don't append it to the binary name
|
|
ifeq ($(BUILD_ID),)
|
|
cp $(BINARY_NAME) release/$(BINARY_NAME)-$(OS)-$(ARCH)
|
|
shasum -a 256 release/$(BINARY_NAME)-$(OS)-$(ARCH) > release/$(BINARY_NAME)-$(OS)-$(ARCH).sha256
|
|
else
|
|
cp $(BINARY_NAME) release/$(BINARY_NAME)-$(BUILD_ID)-$(OS)-$(ARCH)
|
|
shasum -a 256 release/$(BINARY_NAME)-$(BUILD_ID)-$(OS)-$(ARCH) > release/$(BINARY_NAME)-$(BUILD_ID)-$(OS)-$(ARCH).sha256
|
|
endif
|
|
|
|
osx-signed: build
|
|
codesign --deep --force --sign "$(OSX_SIGNING_IDENTITY)" --entitlements "./Entitlements.plist" "./$(BINARY_NAME)"
|
|
|
|
## Run
|
|
run: prepare ## run local-ai
|
|
CGO_LDFLAGS="$(CGO_LDFLAGS)" $(GOCMD) run ./
|
|
|
|
test-models/testmodel.ggml:
|
|
mkdir test-models
|
|
mkdir test-dir
|
|
wget -q https://huggingface.co/mradermacher/gpt2-alpaca-gpt4-GGUF/resolve/main/gpt2-alpaca-gpt4.Q4_K_M.gguf -O test-models/testmodel.ggml
|
|
wget -q https://huggingface.co/ggerganov/whisper.cpp/resolve/main/ggml-base.en.bin -O test-models/whisper-en
|
|
wget -q https://huggingface.co/mudler/all-MiniLM-L6-v2/resolve/main/ggml-model-q4_0.bin -O test-models/bert
|
|
wget -q https://cdn.openai.com/whisper/draft-20220913a/micro-machines.wav -O test-dir/audio.wav
|
|
cp tests/models_fixtures/* test-models
|
|
|
|
prepare-test: grpcs
|
|
cp -rf backend-assets core/http
|
|
cp tests/models_fixtures/* test-models
|
|
|
|
########################################################
|
|
## Tests
|
|
########################################################
|
|
|
|
## Test targets
|
|
test: prepare test-models/testmodel.ggml grpcs
|
|
@echo 'Running tests'
|
|
export GO_TAGS="debug"
|
|
$(MAKE) prepare-test
|
|
HUGGINGFACE_GRPC=$(abspath ./)/backend/python/transformers/run.sh TEST_DIR=$(abspath ./)/test-dir/ FIXTURES=$(abspath ./)/tests/fixtures CONFIG_FILE=$(abspath ./)/test-models/config.yaml MODELS_PATH=$(abspath ./)/test-models BACKENDS_PATH=$(abspath ./)/backends \
|
|
$(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --label-filter="!llama-gguf" --flake-attempts $(TEST_FLAKES) --fail-fast -v -r $(TEST_PATHS)
|
|
$(MAKE) test-llama-gguf
|
|
$(MAKE) test-tts
|
|
$(MAKE) test-stablediffusion
|
|
|
|
backends/llama-cpp: docker-build-llama-cpp docker-save-llama-cpp build-api
|
|
./local-ai backends install "ocifile://$(abspath ./backend-images/llama-cpp.tar)"
|
|
|
|
backends/piper: docker-build-piper docker-save-piper build-api
|
|
./local-ai backends install "ocifile://$(abspath ./backend-images/piper.tar)"
|
|
|
|
########################################################
|
|
## AIO tests
|
|
########################################################
|
|
|
|
docker-build-aio:
|
|
docker build --build-arg MAKEFLAGS="--jobs=5 --output-sync=target" -t local-ai:tests -f Dockerfile .
|
|
BASE_IMAGE=local-ai:tests DOCKER_AIO_IMAGE=local-ai-aio:test $(MAKE) docker-aio
|
|
|
|
e2e-aio:
|
|
LOCALAI_BACKEND_DIR=$(abspath ./backends) \
|
|
LOCALAI_MODELS_DIR=$(abspath ./models) \
|
|
LOCALAI_IMAGE_TAG=test \
|
|
LOCALAI_IMAGE=local-ai-aio \
|
|
$(MAKE) run-e2e-aio
|
|
|
|
run-e2e-aio: protogen-go
|
|
@echo 'Running e2e AIO tests'
|
|
$(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --flake-attempts $(TEST_FLAKES) -v -r ./tests/e2e-aio
|
|
|
|
########################################################
|
|
## E2E tests
|
|
########################################################
|
|
|
|
prepare-e2e:
|
|
mkdir -p $(TEST_DIR)
|
|
cp -rfv $(abspath ./tests/e2e-fixtures)/gpu.yaml $(TEST_DIR)/gpu.yaml
|
|
test -e $(TEST_DIR)/ggllm-test-model.bin || wget -q https://huggingface.co/TheBloke/CodeLlama-7B-Instruct-GGUF/resolve/main/codellama-7b-instruct.Q2_K.gguf -O $(TEST_DIR)/ggllm-test-model.bin
|
|
docker build --build-arg GRPC_BACKENDS="$(GRPC_BACKENDS)" --build-arg IMAGE_TYPE=core --build-arg BUILD_TYPE=$(BUILD_TYPE) --build-arg CUDA_MAJOR_VERSION=12 --build-arg CUDA_MINOR_VERSION=0 --build-arg FFMPEG=true -t localai-tests .
|
|
|
|
run-e2e-image:
|
|
ls -liah $(abspath ./tests/e2e-fixtures)
|
|
docker run -p 5390:8080 -e MODELS_PATH=/models -e THREADS=1 -e DEBUG=true -d --rm -v $(TEST_DIR):/models --gpus all --name e2e-tests-$(RANDOM) localai-tests
|
|
|
|
test-e2e:
|
|
@echo 'Running e2e tests'
|
|
BUILD_TYPE=$(BUILD_TYPE) \
|
|
LOCALAI_API=http://$(E2E_BRIDGE_IP):5390/v1 \
|
|
$(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --flake-attempts $(TEST_FLAKES) -v -r ./tests/e2e
|
|
|
|
teardown-e2e:
|
|
rm -rf $(TEST_DIR) || true
|
|
docker stop $$(docker ps -q --filter ancestor=localai-tests)
|
|
|
|
########################################################
|
|
## Integration and unit tests
|
|
########################################################
|
|
|
|
test-llama-gguf: prepare-test
|
|
TEST_DIR=$(abspath ./)/test-dir/ FIXTURES=$(abspath ./)/tests/fixtures CONFIG_FILE=$(abspath ./)/test-models/config.yaml MODELS_PATH=$(abspath ./)/test-models BACKENDS_PATH=$(abspath ./)/backends \
|
|
$(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --label-filter="llama-gguf" --flake-attempts $(TEST_FLAKES) -v -r $(TEST_PATHS)
|
|
|
|
test-tts: prepare-test
|
|
TEST_DIR=$(abspath ./)/test-dir/ FIXTURES=$(abspath ./)/tests/fixtures CONFIG_FILE=$(abspath ./)/test-models/config.yaml MODELS_PATH=$(abspath ./)/test-models BACKENDS_PATH=$(abspath ./)/backends \
|
|
$(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --label-filter="tts" --flake-attempts $(TEST_FLAKES) -v -r $(TEST_PATHS)
|
|
|
|
test-stablediffusion: prepare-test
|
|
TEST_DIR=$(abspath ./)/test-dir/ FIXTURES=$(abspath ./)/tests/fixtures CONFIG_FILE=$(abspath ./)/test-models/config.yaml MODELS_PATH=$(abspath ./)/test-models BACKENDS_PATH=$(abspath ./)/backends \
|
|
$(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --label-filter="stablediffusion" --flake-attempts $(TEST_FLAKES) -v -r $(TEST_PATHS)
|
|
|
|
test-stores: backend-assets/grpc/local-store
|
|
mkdir -p tests/integration/backend-assets/grpc
|
|
cp -f backend-assets/grpc/local-store tests/integration/backend-assets/grpc/
|
|
$(GOCMD) run github.com/onsi/ginkgo/v2/ginkgo --label-filter="stores" --flake-attempts $(TEST_FLAKES) -v -r tests/integration
|
|
|
|
test-container:
|
|
docker build --target requirements -t local-ai-test-container .
|
|
docker run -ti --rm --entrypoint /bin/bash -ti -v $(abspath ./):/build local-ai-test-container
|
|
|
|
########################################################
|
|
## Help
|
|
########################################################
|
|
|
|
## Help:
|
|
help: ## Show this help.
|
|
@echo ''
|
|
@echo 'Usage:'
|
|
@echo ' ${YELLOW}make${RESET} ${GREEN}<target>${RESET}'
|
|
@echo ''
|
|
@echo 'Targets:'
|
|
@awk 'BEGIN {FS = ":.*?## "} { \
|
|
if (/^[a-zA-Z_-]+:.*?##.*$$/) {printf " ${YELLOW}%-20s${GREEN}%s${RESET}\n", $$1, $$2} \
|
|
else if (/^## .*$$/) {printf " ${CYAN}%s${RESET}\n", substr($$1,4)} \
|
|
}' $(MAKEFILE_LIST)
|
|
|
|
########################################################
|
|
## Backends
|
|
########################################################
|
|
|
|
.PHONY: protogen
|
|
protogen: protogen-go protogen-python
|
|
|
|
.PHONY: protogen-clean
|
|
protogen-clean: protogen-go-clean protogen-python-clean
|
|
|
|
.PHONY: protogen-go
|
|
protogen-go: install-go-tools
|
|
mkdir -p pkg/grpc/proto
|
|
protoc --experimental_allow_proto3_optional -Ibackend/ --go_out=pkg/grpc/proto/ --go_opt=paths=source_relative --go-grpc_out=pkg/grpc/proto/ --go-grpc_opt=paths=source_relative \
|
|
backend/backend.proto
|
|
|
|
.PHONY: protogen-go-clean
|
|
protogen-go-clean:
|
|
$(RM) pkg/grpc/proto/backend.pb.go pkg/grpc/proto/backend_grpc.pb.go
|
|
$(RM) bin/*
|
|
|
|
.PHONY: protogen-python
|
|
protogen-python: bark-protogen coqui-protogen chatterbox-protogen diffusers-protogen exllama2-protogen rerankers-protogen transformers-protogen kokoro-protogen vllm-protogen faster-whisper-protogen
|
|
|
|
.PHONY: protogen-python-clean
|
|
protogen-python-clean: bark-protogen-clean coqui-protogen-clean chatterbox-protogen-clean diffusers-protogen-clean exllama2-protogen-clean rerankers-protogen-clean transformers-protogen-clean kokoro-protogen-clean vllm-protogen-clean faster-whisper-protogen-clean
|
|
|
|
.PHONY: bark-protogen
|
|
bark-protogen:
|
|
$(MAKE) -C backend/python/bark protogen
|
|
|
|
.PHONY: bark-protogen-clean
|
|
bark-protogen-clean:
|
|
$(MAKE) -C backend/python/bark protogen-clean
|
|
|
|
.PHONY: coqui-protogen
|
|
coqui-protogen:
|
|
$(MAKE) -C backend/python/coqui protogen
|
|
|
|
.PHONY: coqui-protogen-clean
|
|
coqui-protogen-clean:
|
|
$(MAKE) -C backend/python/coqui protogen-clean
|
|
|
|
.PHONY: diffusers-protogen
|
|
diffusers-protogen:
|
|
$(MAKE) -C backend/python/diffusers protogen
|
|
|
|
.PHONY: chatterbox-protogen
|
|
chatterbox-protogen:
|
|
$(MAKE) -C backend/python/chatterbox protogen
|
|
|
|
.PHONY: diffusers-protogen-clean
|
|
diffusers-protogen-clean:
|
|
$(MAKE) -C backend/python/diffusers protogen-clean
|
|
|
|
.PHONY: chatterbox-protogen-clean
|
|
chatterbox-protogen-clean:
|
|
$(MAKE) -C backend/python/chatterbox protogen-clean
|
|
|
|
.PHONY: faster-whisper-protogen
|
|
faster-whisper-protogen:
|
|
$(MAKE) -C backend/python/faster-whisper protogen
|
|
|
|
.PHONY: faster-whisper-protogen-clean
|
|
faster-whisper-protogen-clean:
|
|
$(MAKE) -C backend/python/faster-whisper protogen-clean
|
|
|
|
.PHONY: exllama2-protogen
|
|
exllama2-protogen:
|
|
$(MAKE) -C backend/python/exllama2 protogen
|
|
|
|
.PHONY: exllama2-protogen-clean
|
|
exllama2-protogen-clean:
|
|
$(MAKE) -C backend/python/exllama2 protogen-clean
|
|
|
|
.PHONY: rerankers-protogen
|
|
rerankers-protogen:
|
|
$(MAKE) -C backend/python/rerankers protogen
|
|
|
|
.PHONY: rerankers-protogen-clean
|
|
rerankers-protogen-clean:
|
|
$(MAKE) -C backend/python/rerankers protogen-clean
|
|
|
|
.PHONY: transformers-protogen
|
|
transformers-protogen:
|
|
$(MAKE) -C backend/python/transformers protogen
|
|
|
|
.PHONY: transformers-protogen-clean
|
|
transformers-protogen-clean:
|
|
$(MAKE) -C backend/python/transformers protogen-clean
|
|
|
|
.PHONY: kokoro-protogen
|
|
kokoro-protogen:
|
|
$(MAKE) -C backend/python/kokoro protogen
|
|
|
|
.PHONY: kokoro-protogen-clean
|
|
kokoro-protogen-clean:
|
|
$(MAKE) -C backend/python/kokoro protogen-clean
|
|
|
|
.PHONY: vllm-protogen
|
|
vllm-protogen:
|
|
$(MAKE) -C backend/python/vllm protogen
|
|
|
|
.PHONY: vllm-protogen-clean
|
|
vllm-protogen-clean:
|
|
$(MAKE) -C backend/python/vllm protogen-clean
|
|
|
|
## GRPC
|
|
# Note: it is duplicated in the Dockerfile
|
|
prepare-extra-conda-environments: protogen-python
|
|
$(MAKE) -C backend/python/bark
|
|
$(MAKE) -C backend/python/coqui
|
|
$(MAKE) -C backend/python/diffusers
|
|
$(MAKE) -C backend/python/chatterbox
|
|
$(MAKE) -C backend/python/faster-whisper
|
|
$(MAKE) -C backend/python/vllm
|
|
$(MAKE) -C backend/python/rerankers
|
|
$(MAKE) -C backend/python/transformers
|
|
$(MAKE) -C backend/python/kokoro
|
|
$(MAKE) -C backend/python/exllama2
|
|
|
|
prepare-test-extra: protogen-python
|
|
$(MAKE) -C backend/python/transformers
|
|
$(MAKE) -C backend/python/diffusers
|
|
$(MAKE) -C backend/python/chatterbox
|
|
$(MAKE) -C backend/python/vllm
|
|
|
|
test-extra: prepare-test-extra
|
|
$(MAKE) -C backend/python/transformers test
|
|
$(MAKE) -C backend/python/diffusers test
|
|
$(MAKE) -C backend/python/chatterbox test
|
|
$(MAKE) -C backend/python/vllm test
|
|
|
|
backend-assets:
|
|
mkdir -p backend-assets
|
|
ifeq ($(BUILD_API_ONLY),true)
|
|
touch backend-assets/keep
|
|
endif
|
|
|
|
|
|
backend-assets/grpc:
|
|
mkdir -p backend-assets/grpc
|
|
|
|
backend-assets/grpc/huggingface: protogen-go backend-assets/grpc
|
|
$(GOCMD) build -ldflags "$(LD_FLAGS)" -tags "$(GO_TAGS)" -o backend-assets/grpc/huggingface ./backend/go/llm/langchain/
|
|
ifneq ($(UPX),)
|
|
$(UPX) backend-assets/grpc/huggingface
|
|
endif
|
|
|
|
backend-assets/grpc/silero-vad: protogen-go replace backend-assets/grpc backend-assets/lib/libonnxruntime.so.1
|
|
CGO_LDFLAGS="$(CGO_LDFLAGS)" CPATH="$(CPATH):$(CURDIR)/sources/onnxruntime/include/" LIBRARY_PATH=$(CURDIR)/backend-assets/lib \
|
|
$(GOCMD) build -ldflags "$(LD_FLAGS)" -tags "$(GO_TAGS)" -o backend-assets/grpc/silero-vad ./backend/go/vad/silero
|
|
ifneq ($(UPX),)
|
|
$(UPX) backend-assets/grpc/silero-vad
|
|
endif
|
|
|
|
backend-assets/grpc/whisper: protogen-go replace sources/whisper.cpp sources/whisper.cpp/build/src/libwhisper.a backend-assets/grpc
|
|
CGO_LDFLAGS="$(CGO_LDFLAGS) $(CGO_LDFLAGS_WHISPER)" C_INCLUDE_PATH="${WHISPER_INCLUDE_PATH}" LIBRARY_PATH="${WHISPER_LIBRARY_PATH}" LD_LIBRARY_PATH="${WHISPER_LIBRARY_PATH}" \
|
|
CGO_CXXFLAGS="$(CGO_CXXFLAGS_WHISPER)" \
|
|
$(GOCMD) build -ldflags "$(LD_FLAGS)" -tags "$(GO_TAGS)" -o backend-assets/grpc/whisper ./backend/go/transcribe/whisper
|
|
ifneq ($(UPX),)
|
|
$(UPX) backend-assets/grpc/whisper
|
|
endif
|
|
|
|
backend-assets/grpc/local-store: backend-assets/grpc protogen-go replace
|
|
$(GOCMD) build -ldflags "$(LD_FLAGS)" -tags "$(GO_TAGS)" -o backend-assets/grpc/local-store ./backend/go/stores/
|
|
ifneq ($(UPX),)
|
|
$(UPX) backend-assets/grpc/local-store
|
|
endif
|
|
|
|
grpcs: prepare protogen-go $(GRPC_BACKENDS)
|
|
|
|
DOCKER_IMAGE?=local-ai
|
|
DOCKER_AIO_IMAGE?=local-ai-aio
|
|
IMAGE_TYPE?=core
|
|
BASE_IMAGE?=ubuntu:22.04
|
|
|
|
docker:
|
|
docker build \
|
|
--build-arg BASE_IMAGE=$(BASE_IMAGE) \
|
|
--build-arg IMAGE_TYPE=$(IMAGE_TYPE) \
|
|
--build-arg GO_TAGS="$(GO_TAGS)" \
|
|
--build-arg MAKEFLAGS="$(DOCKER_MAKEFLAGS)" \
|
|
--build-arg BUILD_TYPE=$(BUILD_TYPE) \
|
|
-t $(DOCKER_IMAGE) .
|
|
|
|
docker-cuda11:
|
|
docker build \
|
|
--build-arg CUDA_MAJOR_VERSION=11 \
|
|
--build-arg CUDA_MINOR_VERSION=8 \
|
|
--build-arg BASE_IMAGE=$(BASE_IMAGE) \
|
|
--build-arg IMAGE_TYPE=$(IMAGE_TYPE) \
|
|
--build-arg GO_TAGS="$(GO_TAGS)" \
|
|
--build-arg MAKEFLAGS="$(DOCKER_MAKEFLAGS)" \
|
|
--build-arg BUILD_TYPE=$(BUILD_TYPE) \
|
|
-t $(DOCKER_IMAGE)-cuda11 .
|
|
|
|
docker-aio:
|
|
@echo "Building AIO image with base $(BASE_IMAGE) as $(DOCKER_AIO_IMAGE)"
|
|
docker build \
|
|
--build-arg BASE_IMAGE=$(BASE_IMAGE) \
|
|
--build-arg MAKEFLAGS="$(DOCKER_MAKEFLAGS)" \
|
|
-t $(DOCKER_AIO_IMAGE) -f Dockerfile.aio .
|
|
|
|
docker-aio-all:
|
|
$(MAKE) docker-aio DOCKER_AIO_SIZE=cpu
|
|
$(MAKE) docker-aio DOCKER_AIO_SIZE=cpu
|
|
|
|
docker-image-intel:
|
|
docker build \
|
|
--build-arg BASE_IMAGE=intel/oneapi-basekit:${ONEAPI_VERSION}.0-0-devel-ubuntu24.04 \
|
|
--build-arg IMAGE_TYPE=$(IMAGE_TYPE) \
|
|
--build-arg GO_TAGS="$(GO_TAGS)" \
|
|
--build-arg MAKEFLAGS="$(DOCKER_MAKEFLAGS)" \
|
|
--build-arg GRPC_BACKENDS="$(GRPC_BACKENDS)" \
|
|
--build-arg BUILD_TYPE=sycl_f32 -t $(DOCKER_IMAGE) .
|
|
|
|
docker-image-intel-xpu:
|
|
docker build \
|
|
--build-arg BASE_IMAGE=intel/oneapi-basekit:${ONEAPI_VERSION}.0-0-devel-ubuntu22.04 \
|
|
--build-arg IMAGE_TYPE=$(IMAGE_TYPE) \
|
|
--build-arg GO_TAGS="$(GO_TAGS)" \
|
|
--build-arg MAKEFLAGS="$(DOCKER_MAKEFLAGS)" \
|
|
--build-arg GRPC_BACKENDS="$(GRPC_BACKENDS)" \
|
|
--build-arg BUILD_TYPE=sycl_f32 -t $(DOCKER_IMAGE) .
|
|
|
|
########################################################
|
|
## Backends
|
|
########################################################
|
|
|
|
backend-images:
|
|
mkdir -p backend-images
|
|
|
|
docker-build-llama-cpp:
|
|
docker build -t local-ai-backend:llama-cpp -f backend/Dockerfile.llama-cpp .
|
|
|
|
docker-build-bark-cpp:
|
|
docker build -t local-ai-backend:bark-cpp -f backend/Dockerfile.go --build-arg BACKEND=bark-cpp .
|
|
|
|
docker-build-piper:
|
|
docker build -t local-ai-backend:piper -f backend/Dockerfile.go --build-arg BACKEND=piper .
|
|
|
|
docker-save-piper: backend-images
|
|
docker save local-ai-backend:piper -o backend-images/piper.tar
|
|
|
|
docker-save-llama-cpp: backend-images
|
|
docker save local-ai-backend:llama-cpp -o backend-images/llama-cpp.tar
|
|
|
|
docker-save-bark-cpp: backend-images
|
|
docker save local-ai-backend:bark-cpp -o backend-images/bark-cpp.tar
|
|
|
|
docker-build-rerankers:
|
|
docker build -t local-ai-backend:rerankers -f backend/Dockerfile.python --build-arg BACKEND=rerankers .
|
|
|
|
docker-build-vllm:
|
|
docker build -t local-ai-backend:vllm -f backend/Dockerfile.python --build-arg BACKEND=vllm .
|
|
|
|
docker-build-transformers:
|
|
docker build -t local-ai-backend:transformers -f backend/Dockerfile.python --build-arg BACKEND=transformers .
|
|
|
|
docker-build-diffusers:
|
|
docker build -t local-ai-backend:diffusers -f backend/Dockerfile.python --build-arg BACKEND=diffusers .
|
|
|
|
docker-build-kokoro:
|
|
docker build -t local-ai-backend:kokoro -f backend/Dockerfile.python --build-arg BACKEND=kokoro .
|
|
|
|
docker-build-faster-whisper:
|
|
docker build -t local-ai-backend:faster-whisper -f backend/Dockerfile.python --build-arg BACKEND=faster-whisper .
|
|
|
|
docker-build-coqui:
|
|
docker build -t local-ai-backend:coqui -f backend/Dockerfile.python --build-arg BACKEND=coqui .
|
|
|
|
docker-build-bark:
|
|
docker build -t local-ai-backend:bark -f backend/Dockerfile.python --build-arg BACKEND=bark .
|
|
|
|
docker-build-chatterbox:
|
|
docker build -t local-ai-backend:chatterbox -f backend/Dockerfile.python --build-arg BACKEND=chatterbox .
|
|
|
|
docker-build-exllama2:
|
|
docker build -t local-ai-backend:exllama2 -f backend/Dockerfile.python --build-arg BACKEND=exllama2 .
|
|
|
|
docker-build-backends: docker-build-llama-cpp docker-build-rerankers docker-build-vllm docker-build-transformers docker-build-diffusers docker-build-kokoro docker-build-faster-whisper docker-build-coqui docker-build-bark docker-build-chatterbox docker-build-exllama2
|
|
|
|
########################################################
|
|
### END Backends
|
|
########################################################
|
|
|
|
.PHONY: swagger
|
|
swagger:
|
|
swag init -g core/http/app.go --output swagger
|
|
|
|
.PHONY: gen-assets
|
|
gen-assets:
|
|
$(GOCMD) run core/dependencies_manager/manager.go webui_static.yaml core/http/static/assets
|
|
|
|
## Documentation
|
|
docs/layouts/_default:
|
|
mkdir -p docs/layouts/_default
|
|
|
|
docs/static/gallery.html: docs/layouts/_default
|
|
$(GOCMD) run ./.github/ci/modelslist.go ./gallery/index.yaml > docs/static/gallery.html
|
|
|
|
docs/public: docs/layouts/_default docs/static/gallery.html
|
|
cd docs && hugo --minify
|
|
|
|
docs-clean:
|
|
rm -rf docs/public
|
|
rm -rf docs/static/gallery.html
|
|
|
|
.PHONY: docs
|
|
docs: docs/static/gallery.html
|
|
cd docs && hugo serve
|