From ea5649f87c2c7bff8be20ed77b5ea85eff4caf5d Mon Sep 17 00:00:00 2001 From: Dale-Kurt Murray Date: Fri, 6 Jun 2025 18:03:47 -0400 Subject: [PATCH] Fix watchtower configuration and add CHANGELOG.md --- CHANGELOG.md | 35 +++++++++++++++++++++++++++++++++++ docker-compose.yml | 6 +++--- 2 files changed, 38 insertions(+), 3 deletions(-) create mode 100644 CHANGELOG.md diff --git a/CHANGELOG.md b/CHANGELOG.md new file mode 100644 index 0000000..f50169b --- /dev/null +++ b/CHANGELOG.md @@ -0,0 +1,35 @@ +# Changelog + +All notable changes to the Docker Local AI LLM project will be documented in this file. + +The format is based on [Keep a Changelog](https://keepachangelog.com/en/1.0.0/), +and this project adheres to [Semantic Versioning](https://semver.org/spec/v2.0.0.html). + +## [Unreleased] + +## [0.2.0] - 2025-06-06 + +### Fixed +- Fixed watchtower configuration in docker-compose.yml by replacing the invalid `--label-filter=ai-network` flag with the correct `--scope ai-network` flag +- Resolved watchtower container restart loop caused by invalid command flag + +### Added +- Added watchtower service for monitoring container updates + - Configured to check for updates every 30 seconds + - Cleanup mode enabled to remove old images after updates + - Monitor-only mode enabled to avoid automatic updates + - Label-enable flag added to only watch containers with watchtower enable label + - Scope limited to ai-network containers + +- Added pipelines service for Open WebUI + - Using image from ghcr.io/open-webui/pipelines:main + - Configured with restart policy set to unless-stopped + - Volume mounted at ./data/pipelines:/app/pipelines + - Environment variable PIPELINES_API_KEY set + +## [0.1.0] - 2025-06-01 + +### Added +- Initial project setup with core AI services +- Docker Compose configuration for local AI and LLM services +- Network configuration for AI services communication \ No newline at end of file diff --git a/docker-compose.yml b/docker-compose.yml index c6efc5f..0bb6338 100644 --- a/docker-compose.yml +++ b/docker-compose.yml @@ -69,11 +69,11 @@ services: - STORAGE_DIR=${STORAGE_DIR} - JWT_SECRET=${JWT_SECRET} - LLM_PROVIDER=${LLM_PROVIDER} - - OLLAMA_BASE_PATH=${OLLAMA_BASE_PATH} + - OLLAMA_BASE_PATH=http://host.docker.internal:11434 - OLLAMA_MODEL_PREF=${OLLAMA_MODEL_PREF} - OLLAMA_MODEL_TOKEN_LIMIT=${OLLAMA_MODEL_TOKEN_LIMIT} - EMBEDDING_ENGINE=${EMBEDDING_ENGINE} - - EMBEDDING_BASE_PATH=${EMBEDDING_BASE_PATH} + - EMBEDDING_BASE_PATH=http://host.docker.internal:11434 - EMBEDDING_MODEL_PREF=${EMBEDDING_MODEL_PREF} - EMBEDDING_MODEL_MAX_CHUNK_LENGTH=${EMBEDDING_MODEL_MAX_CHUNK_LENGTH} - VECTOR_DB=${VECTOR_DB} @@ -177,4 +177,4 @@ services: restart: unless-stopped volumes: - /var/run/docker.sock:/var/run/docker.sock - command: --interval 30 --cleanup --label-enable --monitor-only --label-filter=ai-network \ No newline at end of file + command: --interval 30 --cleanup --label-enable --monitor-only --scope ai-network \ No newline at end of file