version: '3.8' services: # --- UI 1: Open WebUI (HP Mini 1) --- openwebui: image: ghcr.io/open-webui/open-webui:main networks: - ai_internal - proxy ports: - "3000:8080" environment: - OLLAMA_BASE_URL=http://ollama:11434 volumes: # BIND MOUNT: Easy to backup config/users - /home/phunter/docker-data/openwebui:/app/backend/data deploy: placement: constraints: - node.role == manager # --- UI 2: AnythingLLM (HP Mini 1) --- anythingllm: image: mintplexlabs/anythingllm networks: - ai_internal - proxy ports: - "3001:3001" environment: - STORAGE_DIR=/app/server/storage - LLM_PROVIDER=ollama - OLLAMA_BASE_PATH=http://ollama:11434 - OLLAMA_MODEL_PREF=llama3 volumes: # BIND MOUNT: Easy to backup PDFs/Workspaces - /home/phunter/docker-data/anythingllm:/app/server/storage deploy: placement: constraints: - node.role == manager # --- UI 3: Lobe Chat (HP Mini 1) --- lobechat: image: lobehub/lobe-chat networks: - ai_internal - proxy ports: - "3210:3210" environment: - OLLAMA_PROXY_URL=http://ollama:11434/v1 deploy: placement: constraints: - node.role == manager networks: ai_internal: driver: overlay attachable: true proxy: external: true