n8n-io / self-hosted-ai-starter-kit
- вторник, 24 сентября 2024 г. в 00:00:01
The Self-hosted AI Starter Kit is an open-source template that quickly sets up a local AI environment. Curated by n8n, it provides essential tools for creating secure, self-hosted AI workflows.
Self-hosted AI Starter Kit is an open, docker compose template that quickly bootstraps a fully featured Local AI and Low Code development environment.
Curated by https://github.com/n8n-io, it combines the self-hosted n8n platform with a curated list of compatible AI products and components to quickly get started with building self-hosted AI workflows.
✅ Self-hosted n8n - Low-code platform with over 400 integrations and advanced AI components
✅ Ollama - Cross-platform LLM platform to install and run the latest local LLMs
✅ Qdrant - Open-source, high performance vector store with an comprehensive API
✅ PostgreSQL - Workhorse of the Data Engineering world, handles large amounts of data safely.
⭐️ AI Agents which can schedule appointments
⭐️ Summarise company PDFs without leaking data
⭐️ Smarter slack bots for company comms and IT-ops
⭐️ Analyse financial documents privately and for little cost
git clone https://github.com/n8n-io/self-hosted-ai-starter-kit.git
cd self-hosted-ai-starter-kit
docker compose --profile gpu-nvidia up
Note
If you have not used your Nvidia GPU with Docker before, please follow the Ollama Docker instructions.
If you’re using a Mac with an M1 or newer processor, you can't expose your GPU to the Docker instance, unfortunately. There are two options in this case:
If you want to run Ollama on your mac, check the Ollama homepage for installation instructions, and run the starter kit as follows:
git clone https://github.com/n8n-io/self-hosted-ai-starter-kit.git
cd self-hosted-ai-starter-kit
docker compose up
After you followed the quick start set-up below, change the Ollama credentials
by using http://host.docker.internal:11434/
as the host.
git clone https://github.com/n8n-io/self-hosted-ai-starter-kit.git
cd self-hosted-ai-starter-kit
docker compose --profile cpu up
The main component of the self-hosted AI starter kit is a docker compose file pre-configured with network and disk so there isn’t much else you need to install. After completing the installation steps above, follow the steps below to get started.
To open n8n at any time, visit http://localhost:5678/ in your browser.
With your n8n instance, you’ll have access to over 400 integrations and a suite of basic and advanced AI nodes such as AI Agent, Text classifier, and Information Extractor nodes. To keep everything local, just remember to use the Ollama node for your language model and Qdrant as your vector store.
Note
This starter kit is designed to help you get started with self-hosted AI workflows. While it’s not fully optimized for production environments, it combines robust components that work well together for proof-of-concept projects. You can customize it to meet your specific needs
docker compose --profile gpu-nvidia pull
docker compose create && docker compose --profile gpu-nvidia up
docker compose pull
docker compose create && docker compose up
docker compose --profile cpu pull
docker compose create && docker compose --profile cpu up
n8n is full of useful content for getting started quickly with its AI concepts and nodes. If you run into an issue, go to support.
For more AI workflow ideas, visit the official n8n AI template gallery. From each workflow, select the Use workflow button to automatically import the workflow into your local n8n instance.
The self-hosted AI starter kit will create a shared folder (by default,
located in the same directory) which is mounted to the n8n container and
allows n8n to access files on disk. This folder within the n8n container is
located at /data/shared
-- this is the path you’ll need to use in nodes that
interact with the local filesystem.
Nodes that interact with the local filesystem
This project is licensed under the Apache License 2.0 - see the LICENSE file for details.
Join the conversation in the n8n Forum, where you can: