Update the README (#6)

This commit is contained in:
jeanpaul
2024-08-20 17:42:03 +02:00
committed by GitHub
parent 541552d7eb
commit a6240edfd4

205
README.md
View File

@@ -1,38 +1,181 @@
# n8n Demo setup # Self-hosted AI starter kit
This repo helps quickly bootstrap an n8n demo environment using docker-compose. **Self-hosted AI Starter Kit** is an open, docker compose template that
quickly bootstraps a fully featured Local AI and Low Code development
environment.
### Requirements Curated by <https://github.com/n8n-io>, it combines the self-hosted n8n
- [Docker compose](https://docs.docker.com/compose/) platform with a curated list of compatible AI products and components to
- **Optionally** an Nvidia GPU for faster inference on Ollama quickly get started with building self-hosted AI workflows.
### Setup > [!TIP]
- Clone this repo > [Read the announcement]()
- **Optionally** edit the credentials in the `.env` file
- Start the containers:
- If you have an Nvidia GPU, run `docker compose --profile gpu-nvidia up`
- Otherwise to run inference services on your CPU, run `docker compose --profile cpu up`
- Wait a couple of minutes for all the containers to become healthy
- Open http://localhost:5678 in your browser and fill in the details
- Open the included workflow: http://localhost:5678/workflow/srOnR8PAY3u4RSwb
- Wait until Ollama has downloaded the `llama3.1` model (you can check the
docker console)
### Included service endpoints ### Whats included
- [n8n](http://localhost:5678/)
- [Ollama](http://localhost:11434/)
- [Qdrant](http://localhost:6333/dashboard)
### Local files ✅ [**Self-hosted n8n**](https://n8n.io/) - Low-code platform with over 400
integrations and advanced AI components
When running the demo for the first time, Docker will create a folder `shared` ✅ [**Ollama**](https://ollama.com/) - Cross-platform LLM platform to install
next to the `docker-compose.yml` file. You can add files to that, that will be and run the latest local LLMs
visible on the `/data/shared` folder inside the n8n container, and you can use
that, for example, with the Local File Trigger node.
### Updating ✅ [**Qdrant**](https://qdrant.tech/) - Open-source, high performance vector
- Run `docker compose pull` to fetch all the latest images store with an comprehensive API
- If you use Ollama, use either `docker compose --profile cpu pull` or
`docker compose --profile gpu pull` to pull the correct ollama images ✅ [**PostgreSQL**](https://www.postgresql.org/) - Workhorse of the Data
- Run `docker compose create && docker compose up -d` to update and restart Engineering world, handles large amounts of data safely.
all the containers
### What you can build
⭐️ AI Agents which can schedule appointments
⭐️ Summarise company PDFs without leaking data
⭐️ Smarter slack bots for company comms and IT-ops
⭐️ Analyse financial documents privately and for little cost
## Installation
### For Nvidia GPU users
```
git clone https://github.com/n8n-io/self-hosted-ai-starter-kit.git
cd self-hosted-ai-starter-kit
docker compose --profile gpu-nvidia up
```
### For everyone else
```
git clone https://github.com/n8n-io/self-hosted-ai-starter-kit.git
cd self-hosted-ai-starter-kit
docker compose --profile cpu up
```
> [!TIP]
> If youre using a Mac with an M1 or newer processor, you can run Ollama on
> your host machine for faster GPU inference. Unfortunately, you cant expose
> the GPU to Docker instances. Check the
> [Ollama homepage](https://ollama.com/) for installation instructions, and
> use `http://host.docker.internal:11434/` as the Ollama host in your
> credentials.
## ⚡️ Quick start and usage
The main component of the self-hosted AI starter kit is a docker compose file
pre-configured with network and disk so there isnt much else you need to
install. After completing the installation steps above, follow the steps below
to get started.
1. Open <http://localhost:5678/> in your browser to set up n8n. Youll only
have to do this once.
2. Open the included workflow:
<http://localhost:5678/workflow/srOnR8PAY3u4RSwb>
3. Select **Test workflow** to start running the workflow.
4. If this is the first time youre running the workflow, you may need to wait
until Ollama finishes downloading Llama3.1. You can inspect the docker
console logs to check on the progress.
To open n8n at any time, visit <http://localhost:5678/> in your browser.
With your n8n instance, youll have access to over 400 integrations and a
suite of basic and advanced AI nodes such as
[AI Agent](https://docs.n8n.io/integrations/builtin/cluster-nodes/root-nodes/n8n-nodes-langchain.agent/),
[Text classifier](https://docs.n8n.io/integrations/builtin/cluster-nodes/root-nodes/n8n-nodes-langchain.text-classifier/),
and [Information Extractor](https://docs.n8n.io/integrations/builtin/cluster-nodes/root-nodes/n8n-nodes-langchain.information-extractor/)
nodes. To keep everything local, just remember to use the Ollama node for your
language model and Qdrant as your vector store.
> [!NOTE]
> This starter kit is designed to help you get started with self-hosted AI
> workflows. While its not fully optimized for production environments, it
> combines robust components that work well together for proof-of-concept
> projects. You can customize it to meet your specific needs
## Upgrading
### For Nvidia GPU users
```
docker compose --profile gpu-nvidia pull
docker compose create && docker compose up
```
### For everyone else
```
docker compose --profile cpu pull
docker compose create && docker compose up
```
## 👓 Recommended reading
n8n is full of useful content for getting started quickly with its AI concepts
and nodes. If you run into an issue, go to [support](#support).
- [AI agents for developers: from theory to practice with n8n](https://blog.n8n.io/ai-agents/)
- [Tutorial: Build an AI workflow in n8n](https://docs.n8n.io/advanced-ai/intro-tutorial/)
- [Langchain Concepts in n8n](https://docs.n8n.io/advanced-ai/langchain/langchain-n8n/)
- [Demonstration of key differences between agents and chains](https://docs.n8n.io/advanced-ai/examples/agent-chain-comparison/)
- [What are vector databases?](https://docs.n8n.io/advanced-ai/examples/understand-vector-databases/)
## 🎥 Video walkthrough
- [Installing and using Local AI for n8n](https://www.youtube.com/watch?v=xz_X2N-hPg0)
## 🛍️ More AI templates
For more AI workflow ideas, visit the [**official n8n AI template
gallery**](https://n8n.io/workflows/?categories=AI). From each workflow,
select the **Use workflow** button to automatically import the workflow into
your local n8n instance.
### Learn AI key concepts
- [AI Agent Chat](https://n8n.io/workflows/1954-ai-agent-chat/)
- [AI chat with any data source (using the n8n workflow too)](https://n8n.io/workflows/2026-ai-chat-with-any-data-source-using-the-n8n-workflow-tool/)
- [Chat with OpenAI Assistant (by adding a memory)](https://n8n.io/workflows/2098-chat-with-openai-assistant-by-adding-a-memory/)
- [Use an open-source LLM (via HuggingFace)](https://n8n.io/workflows/1980-use-an-open-source-llm-via-huggingface/)
- [Chat with PDF docs using AI (quoting sources)](https://n8n.io/workflows/2165-chat-with-pdf-docs-using-ai-quoting-sources/)
- [AI agent that can scrape webpages](https://n8n.io/workflows/2006-ai-agent-that-can-scrape-webpages/)
### Local AI templates
- [Tax Code Assistant](https://n8n.io/workflows/2341-build-a-tax-code-assistant-with-qdrant-mistralai-and-openai/)
- [Breakdown Documents into Study Notes with MistralAI and Qdrant](https://n8n.io/workflows/2339-breakdown-documents-into-study-notes-using-templating-mistralai-and-qdrant/)
- [Financial Documents Assistant using Qdrant and](https://n8n.io/workflows/2335-build-a-financial-documents-assistant-using-qdrant-and-mistralai/) [Mistral.ai](http://mistral.ai/)
- [Recipe Recommendations with Qdrant and Mistral](https://n8n.io/workflows/2333-recipe-recommendations-with-qdrant-and-mistral/)
## Tips & tricks
### Accessing local files
The self-hosted AI starter kit will create a shared folder (by default,
located in the same directory) which is mounted to the n8n container and
allows n8n to access files on disk. This folder within the n8n container is
located at `/data/shared` -- this is the path youll need to use in nodes that
interact with the local filesystem.
**Nodes that interact with the local filesystem**
- [Read/Write Files from Disk](https://docs.n8n.io/integrations/builtin/core-nodes/n8n-nodes-base.filesreadwrite/)
- [Local File Trigger](https://docs.n8n.io/integrations/builtin/core-nodes/n8n-nodes-base.localfiletrigger/)
- [Execute Command](https://docs.n8n.io/integrations/builtin/core-nodes/n8n-nodes-base.executecommand/)
## 📜 License
This project is licensed under the Apache License 2.0 - see the
[LICENSE](LICENSE) file for details.
## 💬 Support
Join the conversation in the [n8n Forum](https://community.n8n.io/), where you
can:
- **Share Your Work**: Show off what youve built with n8n and inspire others
in the community.
- **Ask Questions**: Whether youre just getting started or youre a seasoned
pro, the community and our team are ready to support with any challenges.
- **Propose Ideas**: Have an idea for a feature or improvement? Let us know!
Were always eager to hear what youd like to see next.