mirror of
https://github.com/kossakovsky/n8n-install.git
synced 2026-03-07 22:33:11 +00:00
216 lines
5.1 KiB
YAML
216 lines
5.1 KiB
YAML
volumes:
|
|
n8n_storage:
|
|
ollama_storage:
|
|
qdrant_storage:
|
|
open-webui:
|
|
flowise:
|
|
caddy-data:
|
|
caddy-config:
|
|
valkey-data:
|
|
|
|
x-n8n: &service-n8n
|
|
image: n8nio/n8n:latest
|
|
environment:
|
|
- DB_TYPE=postgresdb
|
|
- DB_POSTGRESDB_HOST=db
|
|
- DB_POSTGRESDB_USER=postgres
|
|
- DB_POSTGRESDB_PASSWORD=${POSTGRES_PASSWORD}
|
|
- DB_POSTGRESDB_DATABASE=postgres
|
|
- N8N_DIAGNOSTICS_ENABLED=false
|
|
- N8N_PERSONALIZATION_ENABLED=false
|
|
- N8N_ENCRYPTION_KEY
|
|
- N8N_USER_MANAGEMENT_JWT_SECRET
|
|
|
|
x-ollama: &service-ollama
|
|
image: ollama/ollama:latest
|
|
container_name: ollama
|
|
restart: unless-stopped
|
|
ports:
|
|
- 11434:11434
|
|
volumes:
|
|
- ollama_storage:/root/.ollama
|
|
|
|
x-init-ollama: &init-ollama
|
|
image: ollama/ollama:latest
|
|
container_name: ollama-pull-llama
|
|
volumes:
|
|
- ollama_storage:/root/.ollama
|
|
entrypoint: /bin/sh
|
|
command:
|
|
- "-c"
|
|
- "sleep 3; OLLAMA_HOST=ollama:11434 ollama pull qwen2.5:7b-instruct-q4_K_M; OLLAMA_HOST=ollama:11434 ollama pull nomic-embed-text"
|
|
# For a larger context length verison of the model, run these commands:
|
|
# echo "FROM qwen2.5:7b-instruct-q4_K_M\n\nPARAMETER num_ctx 8096" > Modelfile
|
|
# ollama create qwen2.5:7b-8k -f ./Modelfile
|
|
# Change the name of the LLM and num_ctx as you see fit.
|
|
|
|
services:
|
|
flowise:
|
|
image: flowiseai/flowise
|
|
restart: unless-stopped
|
|
container_name: flowise
|
|
environment:
|
|
- PORT=3001
|
|
ports:
|
|
- 3001:3001
|
|
extra_hosts:
|
|
- "host.docker.internal:host-gateway"
|
|
volumes:
|
|
- ~/.flowise:/root/.flowise
|
|
entrypoint: /bin/sh -c "sleep 3; flowise start"
|
|
|
|
open-webui:
|
|
image: ghcr.io/open-webui/open-webui:main
|
|
restart: unless-stopped
|
|
container_name: open-webui
|
|
ports:
|
|
- "3000:8080"
|
|
extra_hosts:
|
|
- "host.docker.internal:host-gateway"
|
|
volumes:
|
|
- open-webui:/app/backend/data
|
|
|
|
n8n-import:
|
|
<<: *service-n8n
|
|
container_name: n8n-import
|
|
entrypoint: /bin/sh
|
|
command:
|
|
- "-c"
|
|
- "n8n import:credentials --separate --input=/backup/credentials && n8n import:workflow --separate --input=/backup/workflows"
|
|
volumes:
|
|
- ./n8n/backup:/backup
|
|
|
|
n8n:
|
|
<<: *service-n8n
|
|
container_name: n8n
|
|
restart: unless-stopped
|
|
ports:
|
|
- 5678:5678
|
|
volumes:
|
|
- n8n_storage:/home/node/.n8n
|
|
- ./n8n/backup:/backup
|
|
- ./shared:/data/shared
|
|
depends_on:
|
|
n8n-import:
|
|
condition: service_completed_successfully
|
|
|
|
qdrant:
|
|
image: qdrant/qdrant
|
|
container_name: qdrant
|
|
restart: unless-stopped
|
|
ports:
|
|
- 6333:6333
|
|
volumes:
|
|
- qdrant_storage:/qdrant/storage
|
|
|
|
caddy:
|
|
container_name: caddy
|
|
image: docker.io/library/caddy:2-alpine
|
|
network_mode: host
|
|
restart: unless-stopped
|
|
volumes:
|
|
- ./Caddyfile:/etc/caddy/Caddyfile:ro
|
|
- caddy-data:/data:rw
|
|
- caddy-config:/config:rw
|
|
environment:
|
|
- N8N_HOSTNAME=${N8N_HOSTNAME:-":8001"}
|
|
- WEBUI_HOSTNAME=${WEBUI_HOSTNAME:-":8002"}
|
|
- FLOWISE_HOSTNAME=${FLOWISE_HOSTNAME:-":8003"}
|
|
- OLLAMA_HOSTNAME=${OLLAMA_HOSTNAME:-":8004"}
|
|
- SUPABASE_HOSTNAME=${SUPABASE_HOSTNAME:-":8005"}
|
|
- SEARXNG_HOSTNAME=${SEARXNG_HOSTNAME:-":8006"}
|
|
- LETSENCRYPT_EMAIL=${LETSENCRYPT_EMAIL:-internal}
|
|
cap_drop:
|
|
- ALL
|
|
cap_add:
|
|
- NET_BIND_SERVICE
|
|
logging:
|
|
driver: "json-file"
|
|
options:
|
|
max-size: "1m"
|
|
max-file: "1"
|
|
|
|
redis:
|
|
container_name: redis
|
|
image: docker.io/valkey/valkey:8-alpine
|
|
command: valkey-server --save 30 1 --loglevel warning
|
|
restart: unless-stopped
|
|
volumes:
|
|
- valkey-data:/data
|
|
cap_drop:
|
|
- ALL
|
|
cap_add:
|
|
- SETGID
|
|
- SETUID
|
|
- DAC_OVERRIDE
|
|
logging:
|
|
driver: "json-file"
|
|
options:
|
|
max-size: "1m"
|
|
max-file: "1"
|
|
|
|
searxng:
|
|
container_name: searxng
|
|
image: docker.io/searxng/searxng:latest
|
|
restart: unless-stopped
|
|
ports:
|
|
- 8080:8080
|
|
volumes:
|
|
- ./searxng:/etc/searxng:rw
|
|
environment:
|
|
- SEARXNG_BASE_URL=https://${SEARXNG_HOSTNAME:-localhost}/
|
|
- UWSGI_WORKERS=${SEARXNG_UWSGI_WORKERS:-4}
|
|
- UWSGI_THREADS=${SEARXNG_UWSGI_THREADS:-4}
|
|
cap_drop:
|
|
- ALL
|
|
cap_add:
|
|
- CHOWN
|
|
- SETGID
|
|
- SETUID
|
|
logging:
|
|
driver: "json-file"
|
|
options:
|
|
max-size: "1m"
|
|
max-file: "1"
|
|
|
|
ollama-cpu:
|
|
profiles: ["cpu"]
|
|
<<: *service-ollama
|
|
|
|
ollama-gpu:
|
|
profiles: ["gpu-nvidia"]
|
|
<<: *service-ollama
|
|
deploy:
|
|
resources:
|
|
reservations:
|
|
devices:
|
|
- driver: nvidia
|
|
count: 1
|
|
capabilities: [gpu]
|
|
|
|
ollama-gpu-amd:
|
|
profiles: ["gpu-amd"]
|
|
<<: *service-ollama
|
|
image: ollama/ollama:rocm
|
|
devices:
|
|
- "/dev/kfd"
|
|
- "/dev/dri"
|
|
|
|
ollama-pull-llama-cpu:
|
|
profiles: ["cpu"]
|
|
<<: *init-ollama
|
|
depends_on:
|
|
- ollama-cpu
|
|
|
|
ollama-pull-llama-gpu:
|
|
profiles: ["gpu-nvidia"]
|
|
<<: *init-ollama
|
|
depends_on:
|
|
- ollama-gpu
|
|
|
|
ollama-pull-llama-gpu-amd:
|
|
profiles: [gpu-amd]
|
|
<<: *init-ollama
|
|
image: ollama/ollama:rocm
|
|
depends_on:
|
|
- ollama-gpu-amd |