Files
local-llm-stack/CHANGELOG.md

1.5 KiB

Changelog

All notable changes to the Docker Local AI LLM project will be documented in this file.

The format is based on Keep a Changelog, and this project adheres to Semantic Versioning.

[Unreleased]

Fixed

  • Fixed PostgreSQL container permission issues by switching from bind mount to named volume
  • Resolved "chown: /var/lib/postgresql/data: Permission denied" errors in PostgreSQL container

[0.2.0] - 2025-06-06

Fixed

  • Fixed watchtower configuration in docker-compose.yml by replacing the invalid --label-filter=ai-network flag with the correct --scope ai-network flag
  • Resolved watchtower container restart loop caused by invalid command flag

Added

  • Added watchtower service for monitoring container updates

    • Configured to check for updates every 30 seconds
    • Cleanup mode enabled to remove old images after updates
    • Monitor-only mode enabled to avoid automatic updates
    • Label-enable flag added to only watch containers with watchtower enable label
    • Scope limited to ai-network containers
  • Added pipelines service for Open WebUI

    • Using image from ghcr.io/open-webui/pipelines:main
    • Configured with restart policy set to unless-stopped
    • Volume mounted at ./data/pipelines:/app/pipelines
    • Environment variable PIPELINES_API_KEY set

[0.1.0] - 2025-06-01

Added

  • Initial project setup with core AI services
  • Docker Compose configuration for local AI and LLM services
  • Network configuration for AI services communication