Files
self-hosted-ai-starter-kit/README.md
2024-08-14 15:46:32 +02:00

1.5 KiB

n8n Demo setup

This repo helps quickly bootstrap an n8n demo environment using docker-compose.

Requirements

  • Docker compose
  • Optionally an Nvidia GPU for faster inference on Ollama

Setup

  • Clone this repo
  • Optionally edit the credentials in the .env file
  • Start the containers:
    • If you have an Nvidia GPU, run docker compose --profile gpu-nvidia up
    • Otherwise to run inference services on your CPU, run docker compose --profile cpu up
  • Wait a couple of minutes for all the containers to become healthy
  • Open http://localhost:5678 in your browser and fill in the details
  • Open the included workflow: http://localhost:5678/workflow/srOnR8PAY3u4RSwb
  • Wait until Ollama has downloaded the llama3.1 model (you can check the docker console)

Included service endpoints

Local files

When running the demo for the first time, Docker will create a folder shared next to the docker-compose.yml file. You can add files to that, that will be visible on the /data/shared folder inside the n8n container, and you can use that, for example, with the Local File Trigger node.

Updating

  • Run docker compose pull to fetch all the latest images
  • If you use Ollama, use either docker compose --profile cpu pull or docker compose --profile gpu pull to pull the correct ollama images
  • Run docker compose create && docker compose up -d to update and restart all the containers