Files
n8n-install/scripts
Yury Kossakovsky d2abd6e036 Add Ollama service configuration to docker-compose.yml and update related files
- Introduced Ollama service with CPU and GPU profiles in docker-compose.yml, allowing users to run large language models locally.
- Added Ollama selection option in the wizard script for hardware profile configuration.
- Updated README.md to include Ollama as a new available service.
- Adjusted .env.example to include GRAFANA_HOSTNAME in the correct position.
2025-05-15 20:44:36 -06:00
..