mirror of
https://github.com/mudler/LocalAI.git
synced 2025-12-20 17:00:19 -06:00
* feat(ui): add watchdog settings Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Do not re-read env Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Some refactor, move other settings to runtime (p2p) Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Add API Keys handling Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Allow to disable runtime settings Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Documentation Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Small fixups Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * show MCP toggle in index Signed-off-by: Ettore Di Giacinto <mudler@localai.io> * Drop context default Signed-off-by: Ettore Di Giacinto <mudler@localai.io> --------- Signed-off-by: Ettore Di Giacinto <mudler@localai.io>
267 lines
8.2 KiB
Go
267 lines
8.2 KiB
Go
package routes
|
|
|
|
import (
|
|
"github.com/labstack/echo/v4"
|
|
"github.com/mudler/LocalAI/core/config"
|
|
"github.com/mudler/LocalAI/core/gallery"
|
|
"github.com/mudler/LocalAI/core/http/endpoints/localai"
|
|
"github.com/mudler/LocalAI/core/http/middleware"
|
|
"github.com/mudler/LocalAI/core/services"
|
|
"github.com/mudler/LocalAI/internal"
|
|
"github.com/mudler/LocalAI/pkg/model"
|
|
)
|
|
|
|
func RegisterUIRoutes(app *echo.Echo,
|
|
cl *config.ModelConfigLoader,
|
|
ml *model.ModelLoader,
|
|
appConfig *config.ApplicationConfig,
|
|
galleryService *services.GalleryService) {
|
|
|
|
// keeps the state of ops that are started from the UI
|
|
var processingOps = services.NewOpCache(galleryService)
|
|
|
|
app.GET("/", localai.WelcomeEndpoint(appConfig, cl, ml, processingOps))
|
|
app.GET("/manage", localai.WelcomeEndpoint(appConfig, cl, ml, processingOps))
|
|
|
|
if !appConfig.DisableRuntimeSettings {
|
|
// Settings page
|
|
app.GET("/settings", func(c echo.Context) error {
|
|
summary := map[string]interface{}{
|
|
"Title": "LocalAI - Settings",
|
|
"BaseURL": middleware.BaseURL(c),
|
|
}
|
|
return c.Render(200, "views/settings", summary)
|
|
})
|
|
}
|
|
|
|
// P2P
|
|
app.GET("/p2p/", func(c echo.Context) error {
|
|
summary := map[string]interface{}{
|
|
"Title": "LocalAI - P2P dashboard",
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"Version": internal.PrintableVersion(),
|
|
//"Nodes": p2p.GetAvailableNodes(""),
|
|
//"FederatedNodes": p2p.GetAvailableNodes(p2p.FederatedID),
|
|
|
|
"P2PToken": appConfig.P2PToken,
|
|
"NetworkID": appConfig.P2PNetworkID,
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/p2p", summary)
|
|
})
|
|
|
|
// Note: P2P UI fragment routes (/p2p/ui/*) were removed
|
|
// P2P nodes are now fetched via JSON API at /api/p2p/workers and /api/p2p/federation
|
|
|
|
// End P2P
|
|
|
|
if !appConfig.DisableGalleryEndpoint {
|
|
registerGalleryRoutes(app, cl, appConfig, galleryService, processingOps)
|
|
registerBackendGalleryRoutes(app, appConfig, galleryService, processingOps)
|
|
}
|
|
|
|
app.GET("/talk/", func(c echo.Context) error {
|
|
modelConfigs, _ := services.ListModels(cl, ml, config.NoFilterFn, services.SKIP_IF_CONFIGURED)
|
|
|
|
if len(modelConfigs) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(302, middleware.BaseURL(c))
|
|
}
|
|
|
|
summary := map[string]interface{}{
|
|
"Title": "LocalAI - Talk",
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"ModelsConfig": modelConfigs,
|
|
"Model": modelConfigs[0],
|
|
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/talk", summary)
|
|
})
|
|
|
|
app.GET("/chat/", func(c echo.Context) error {
|
|
modelConfigs := cl.GetAllModelsConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(modelConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(302, middleware.BaseURL(c))
|
|
}
|
|
modelThatCanBeUsed := ""
|
|
galleryConfigs := map[string]*gallery.ModelConfig{}
|
|
|
|
for _, m := range modelConfigs {
|
|
cfg, err := gallery.GetLocalModelConfiguration(ml.ModelPath, m.Name)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
galleryConfigs[m.Name] = cfg
|
|
}
|
|
|
|
title := "LocalAI - Chat"
|
|
var modelContextSize *int
|
|
|
|
for _, b := range modelConfigs {
|
|
if b.HasUsecases(config.FLAG_CHAT) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Chat with " + modelThatCanBeUsed
|
|
if b.LLMConfig.ContextSize != nil {
|
|
modelContextSize = b.LLMConfig.ContextSize
|
|
}
|
|
break
|
|
}
|
|
}
|
|
|
|
summary := map[string]interface{}{
|
|
"Title": title,
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"GalleryConfig": galleryConfigs,
|
|
"ModelsConfig": modelConfigs,
|
|
"Model": modelThatCanBeUsed,
|
|
"ContextSize": modelContextSize,
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/chat", summary)
|
|
})
|
|
|
|
// Show the Chat page
|
|
app.GET("/chat/:model", func(c echo.Context) error {
|
|
modelConfigs := cl.GetAllModelsConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
galleryConfigs := map[string]*gallery.ModelConfig{}
|
|
modelName := c.Param("model")
|
|
var modelContextSize *int
|
|
|
|
for _, m := range modelConfigs {
|
|
cfg, err := gallery.GetLocalModelConfiguration(ml.ModelPath, m.Name)
|
|
if err != nil {
|
|
continue
|
|
}
|
|
galleryConfigs[m.Name] = cfg
|
|
if m.Name == modelName && m.LLMConfig.ContextSize != nil {
|
|
modelContextSize = m.LLMConfig.ContextSize
|
|
}
|
|
}
|
|
|
|
summary := map[string]interface{}{
|
|
"Title": "LocalAI - Chat with " + modelName,
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"ModelsConfig": modelConfigs,
|
|
"GalleryConfig": galleryConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": modelName,
|
|
"ContextSize": modelContextSize,
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/chat", summary)
|
|
})
|
|
|
|
app.GET("/text2image/:model", func(c echo.Context) error {
|
|
modelConfigs := cl.GetAllModelsConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
summary := map[string]interface{}{
|
|
"Title": "LocalAI - Generate images with " + c.Param("model"),
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"ModelsConfig": modelConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Param("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/text2image", summary)
|
|
})
|
|
|
|
app.GET("/text2image/", func(c echo.Context) error {
|
|
modelConfigs := cl.GetAllModelsConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(modelConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(302, middleware.BaseURL(c))
|
|
}
|
|
|
|
modelThatCanBeUsed := ""
|
|
title := "LocalAI - Generate images"
|
|
|
|
for _, b := range modelConfigs {
|
|
if b.HasUsecases(config.FLAG_IMAGE) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Generate images with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
|
|
summary := map[string]interface{}{
|
|
"Title": title,
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"ModelsConfig": modelConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": modelThatCanBeUsed,
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/text2image", summary)
|
|
})
|
|
|
|
app.GET("/tts/:model", func(c echo.Context) error {
|
|
modelConfigs := cl.GetAllModelsConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
summary := map[string]interface{}{
|
|
"Title": "LocalAI - Generate images with " + c.Param("model"),
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"ModelsConfig": modelConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": c.Param("model"),
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/tts", summary)
|
|
})
|
|
|
|
app.GET("/tts/", func(c echo.Context) error {
|
|
modelConfigs := cl.GetAllModelsConfigs()
|
|
modelsWithoutConfig, _ := services.ListModels(cl, ml, config.NoFilterFn, services.LOOSE_ONLY)
|
|
|
|
if len(modelConfigs)+len(modelsWithoutConfig) == 0 {
|
|
// If no model is available redirect to the index which suggests how to install models
|
|
return c.Redirect(302, middleware.BaseURL(c))
|
|
}
|
|
|
|
modelThatCanBeUsed := ""
|
|
title := "LocalAI - Generate audio"
|
|
|
|
for _, b := range modelConfigs {
|
|
if b.HasUsecases(config.FLAG_TTS) {
|
|
modelThatCanBeUsed = b.Name
|
|
title = "LocalAI - Generate audio with " + modelThatCanBeUsed
|
|
break
|
|
}
|
|
}
|
|
summary := map[string]interface{}{
|
|
"Title": title,
|
|
"BaseURL": middleware.BaseURL(c),
|
|
"ModelsConfig": modelConfigs,
|
|
"ModelsWithoutConfig": modelsWithoutConfig,
|
|
"Model": modelThatCanBeUsed,
|
|
"Version": internal.PrintableVersion(),
|
|
}
|
|
|
|
// Render index
|
|
return c.Render(200, "views/tts", summary)
|
|
})
|
|
}
|