For Mac users running OLLAMA locally (#29)

Co-authored-by: Mark Susol <marksusol@gmail.com>
This commit is contained in:
कारतोफ्फेलस्क्रिप्ट™
2025-02-06 12:55:19 +01:00
committed by GitHub
parent b13ff3e76d
commit 22709d4aa5
2 changed files with 20 additions and 3 deletions

View File

@@ -78,8 +78,24 @@ cd self-hosted-ai-starter-kit
docker compose up
```
After you followed the quick start set-up below, change the Ollama credentials
by using `http://host.docker.internal:11434/` as the host.
##### For Mac users running OLLAMA locally
If you're running OLLAMA locally on your Mac (not in Docker), you need to modify the OLLAMA_HOST environment variable
in the n8n service configuration. Update the x-n8n section in your Docker Compose file as follows:
```yaml
x-n8n: &service-n8n
# ... other configurations ...
environment:
# ... other environment variables ...
- OLLAMA_HOST=host.docker.internal:11434
```
Additionally, after you see "Editor is now accessible via: <http://localhost:5678/>":
1. Head to <http://localhost:5678/home/credentials>
2. Click on "Local Ollama service"
3. Change the base URL to "http://host.docker.internal:11434/"
#### For everyone else

View File

@@ -19,6 +19,7 @@ x-n8n: &service-n8n
- N8N_PERSONALIZATION_ENABLED=false
- N8N_ENCRYPTION_KEY
- N8N_USER_MANAGEMENT_JWT_SECRET
- OLLAMA_HOST=ollama:11434
links:
- postgres
@@ -41,7 +42,7 @@ x-init-ollama: &init-ollama
entrypoint: /bin/sh
command:
- "-c"
- "sleep 3; OLLAMA_HOST=ollama:11434 ollama pull llama3.2"
- "sleep 3; ollama pull llama3.2"
services:
postgres: