This consists of a fairly comprehensive stack of AI tools that you can run locally for whatever you want.
Take note that this will use 4 GPUs as is. I think you'll need at least 2 for a decent experience and ability to run LLMs and image models. I haven't tested with one GPU, but it may work. To adjust the number of GPUs, change the device_ids
under localai
and comfyui
. Feel free to remove whatever you don't want to use.
When referencing one service from another (like adding a url for comfyui in open webui), make sure you reference the hostname
and internal port of the service. like: http://comfyui:7860
.
- localai: http://machine.ip.address.here:8080
- llm-proxy: http://machine.ip.address.here:3001
- open webui: http://machine.ip.address.here:3000
- searxng: http://machine.ip.address.here:8081
- comfy-ui: http://machine.ip.address.here:7860
- qdrant: http://machine.ip.address.here:6333
- n8n: http://machine.ip.address.here:5678
Enables running open source LLMs/Transformers with support for distributed inferencing.
Thanks to RoboTF AI who helped me get LocalAI working with distributed inferencing.
Check out docker-compose-localai-worker.yml
to run a p2p worker for distributed inferencing. You may need to run 1 worker per GPU on the remote machine.
A simple proxy to make it easier to interact with local AI models. Aggregate models running on separate machines, adds TLS & api keys with the same method as OpenAI's API.
To set this up, refer to the readme.
A feature-rich UI for chatting with and interacting with your LLMs. Open WebUI can use the following services in this stack. SearXNG, ComfyUI, LocalAI, or LocalAI with LLM-Proxy.
A privacy respecting, open source metasearch engine.
SearXNG uses the following services in this stack: Redis (Valkey)
A feature-rich and extensible Stable Diffusion UI, for generating and modifying images from prompt inputs. Works with Flux also.
You'll need to pull a repo for this - it builds locally. stable-diffusion-webui-docker. Check out Techno Tim's tutorial for details on what to do. Techno Tim's Private AI Stack tutorial
This could definitely use some work. I tend to prefer pulling images instead of building locally, but I haven't made the time to do something about that for ComfyUI yet.
A vector database for AI applications. Supports similarity search over any vector space and can be used to store arbitrary data.
A highly extensible low-code automation and integration platform.
n8n uses the following services in this stack. Postgres, Qdrant, and LocalAI with LLM-Proxy.