mirror of
https://github.com/n8n-io/self-hosted-ai-starter-kit.git
synced 2025-11-29 00:23:13 +00:00
1.5 KiB
1.5 KiB
n8n Demo setup
This repo helps quickly bootstrap an n8n demo environment using docker-compose.
Requirements
- Docker compose
- Optionally an Nvidia GPU for faster inference on Ollama
Setup
- Clone this repo
- Optionally edit the credentials in the
.envfile - Start the containers:
- If you have an Nvidia GPU, run
docker compose --profile gpu-nvidia up - Otherwise to run inference services on your CPU, run
docker compose --profile cpu up
- If you have an Nvidia GPU, run
- Wait a couple of minutes for all the containers to become healthy
- Open http://localhost:5678 in your browser and fill in the details
- Open the included workflow: http://localhost:5678/workflow/srOnR8PAY3u4RSwb
- Wait until Ollama has downloaded the
llama3.1model (you can check the docker console)
Included service endpoints
Local files
When running the demo for the first time, Docker will create a folder shared
next to the docker-compose.yml file. You can add files to that, that will be
visible on the /data/shared folder inside the n8n container, and you can use
that, for example, with the Local File Trigger node.
Updating
- Run
docker compose pullto fetch all the latest images - If you use Ollama, use either
docker compose --profile cpu pullordocker compose --profile gpu pullto pull the correct ollama images - Run
docker compose create && docker compose up -dto update and restart all the containers