by Manuel
Who is this template for? This workflow template is ideal for anyone using Notion for project management and Clockify for time tracking. The workflow automatically adds all new clients from Notion to Clockify. How it works Scans your Notion client table every minute for new clients Adds all new clients to your Clockify workspace Set up Steps Set up the Notion trigger node by adding your Notion API credentials as described in the n8n Notion docs. Go to your Notion clients page/table and give your integration permission to acces the data on this page. Go back to n8n and select your Notion client page in the Notion trigger node. Set up the Clockify node by adding your Clockify API credentials as described in the n8n Clockify docs, select your Clockify workspace and map your client name column from Notion to the Clockify "Client Name" field.
by Mario
Purpose This workflow adds the capability to build a RAG on living data. In this case Notion is used as a Knowledge Base. Whenever a page is updated, the embeddings get upserted in a Supabase Vector Store. It can also be fairly easily adapted to PGVector, Pinecone, or Qdrant by using a custom HTTP request for the latter two. Demo How it works A trigger checks every minute for changes in the Notion Database. The manual polling approach improves accuracy and prevents changes from being lost between cached polling intervals. Afterwards every updated page is processed sequentially The Vector Database is searched using the Notion Page ID stored in the metadata of each embedding. If old entries exist, they are deleted. All blocks of the Notion Database Page are retrieved and combined into a single string The content is embedded and split into chunks if necessary. Metadata, including the Notion Page ID, is added during storage for future reference. A simple Question and Answer Chain enables users to ask questions about the embedded content through the integrated chat function Prerequisites To setup a new Vector Store in Supabase, follow this guide Prepare a simple Database in Notion with each Database Page containing at least a title and some content in the blocks section. You can of course also connect this to an existing Database of your choice. Setup Select your credentials in the nodes which require those If you are on an n8n cloud plan, switch to the native Notion Trigger by activating it and deactivating the Schedule Trigger along with its subsequent Notion Node Choose your Notion Database in the first Node related to Notion Adjust the chunk size and overlap in the Token Splitter to your preference Activate the workflow How to use Populate your Notion Database with useful information and use the chat mode of this workflow to ask questions about it. Updates to a Notion Page should quickly reflect in future conversations.
by Artem Boiko
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. CAD-BIM Multi-Format Validation Pipeline This workflow enables automated validation of CAD and BIM files in multiple formats (Revit, IFC, DWG, DGN) for compliance with project standards and requirements. Key Features Converts Revit, IFC, DWG, and DGN models into open data tables Runs automated validation checks on model naming, structure, attributes, and completeness Generates error reports and QTO (Quantity Take-Off) tables for all processed files How it works Upload one or more project files in Revit (.rvt), IFC (.ifc), DWG, or DGN formats The pipeline automatically processes each file and validates against configurable rules in Excel form Error summaries and QTO tables are generated All outputs are available for download as Excel Converter Path:** Make sure the converter executable (e.g. RvtExporter.exe) is placed in DDC Exporter\datadrivenlibs\. Specify the full path in the workflow settings if required. Troubleshooting:** If conversion fails, double-check the path to the executable. Only supported formats can be processed (see GitHub Readme). Review logs in /output for error details. Docs & Issues:** Full Readme on GitHub
by Trung Tran
Multi-Agent Book Creation Workflow with AI Tool Node and GPT-4, DALL-E Who’s it for This workflow is designed for: Content creators** who want to generate books or structured documents automatically. Educators and trainers** who need quick course materials, eBooks, or study guides. Automation enthusiasts* exploring *multi-agent systems* using the newly released *AI Tool Node** in n8n. Developers* looking for a reference template to understand *orchestration of multiple AI agents** with structured output. How it works / What it does This template demonstrates a multi-agent orchestration system powered by AI Tool Nodes: Trigger: Workflow starts when a chat message is received. Book Brief Agent: Generates the initial book concept (title, subtitle, and outline). Book Writer Agent: Expands the outline into full content by collaborating with two sub-agents: Designer Agent → Provides layout/design suggestions. Content Writer Agent → Drafts and refines chapters. Generate Cover Image: AI generates a custom book cover image. Upload to AWS S3: Stores the cover image securely. Configure Metadata: Adds metadata for title, author, and description. Build Book HTML: Converts markdown-based content into HTML format. Upload to Google Drive: Saves the HTML content for processing. Convert to PDF: Transforms the book into a professional PDF. Archive to Google Drive: Final version is archived for safe storage. This workflow showcases multi-agent coordination, structured parsing, and seamless integration with cloud storage services. How to set up Import the workflow into n8n. Configure the following connections: OpenAI (for Book Brief, Book Writer, Designer, and Content Writer Agents). AWS S3 (for image storage). Google Drive (for document storage & archiving). Add your API keys and credentials in n8n credentials manager. Test the workflow by sending a sample chat message (e.g., “Write a book about AI in education”). Verify outputs in Google Drive (HTML + PDF) and AWS S3 (cover image). Requirements n8n* (latest version with *AI Tool Node** support). OpenAI API key** (to power multi-agent models). AWS account** (with S3 bucket for storing images). Google Drive integration** (for document storage and archiving). Basic familiarity with workflow setup in n8n. How to customize the workflow Switch Models**: Replace gpt-4.1-mini with other models (faster, cheaper, or more powerful). Add More Agents: Introduce agents for **editing, fact-checking, or translation. Change Output Format: Export to **EPUB, DOCX, or Markdown instead of PDF. Branding Options: Modify the **cover generation prompt to include company logos or specific style. Extend Storage: Add **Dropbox, OneDrive, or Notion integration for additional archiving. Trigger Alternatives: Replace chat trigger with **form submission, webhook, or schedule-based runs. ✅ This workflow acts as a free, plug-and-play template to showcase how multi-agents + AI Tool Node can work together to automate complex content creation pipelines.
by Iniyavan JC
How it works This workflow creates a multi-talented AI assistant named Simran that interacts with users via Telegram. It can handle text and voice messages, understand the user's intent, and perform various tasks. Step 1: Receive & Transcribe Input The workflow triggers on any new Telegram message. If it's a voice message, it uses AssemblyAI to transcribe it into text; otherwise, it processes the incoming text directly. Step 2: Understand User Intent Using a Large Language Model (LLM), the workflow analyzes the user's message to determine their goal, categorizing it as a general chat, a request to generate an image, a command to set a reminder, or a request to remember a specific piece of information. Step 3: Fetch Context & Route The assistant retrieves past conversation summaries from a MongoDB database to maintain context. Based on the user's intent, the workflow routes the task to the appropriate path. Step 4: Execute the Task Chat: Generates a response using an AI agent whose personality can be toggled between a standard assistant and a "Girlfriend Mode." It also analyzes the user's mood to tailor the response. Generate Image: Creates a detailed prompt and uses an image generation API to create and send a picture. Set Reminder: Parses the natural language request, creates an event in Google Calendar and a task in Google Tasks, and sends a confirmation. Remember Info: Saves specific user-provided information to a dedicated memory collection in MongoDB. Step 5: Respond and Save Memory The final output (text, voice message, or image) is sent back to the user on Telegram. The workflow then summarizes the interaction and saves it to the database to ensure continuity in future conversations. Set up steps Estimated Set up time: 20 - 30 minutes. Configure Credentials: You will need to add credentials for several services in your n8n instance: Telegram (Bot API Token) AssemblyAI (API Key) MongoDB Google (for Calendar, Tasks, Sheets, and Natural Language API) A Large Language Model (the workflow uses Google Gemini but can be adapted) An image generation service (the workflow uses the Together.xyz API) Set up External Services: Ensure your MongoDB instance has two collections: user_memory and memory_auto. Create a Google Sheet to manage the "Girlfriend Mode" status for different users. Ensure edge-tts is installed on the machine running your n8n instance for the text-to-speech functionality. Customize Nodes: Review the nodes with hardcoded IDs, such as Google Tasks and Google Sheets, and update them with your specific Task List ID and Sheet ID. The sticky notes inside the workflow provide more detailed instructions for specific nodes and segments.
by Joseph LePage
🤖 AI-Powered RAG Chatbot with Google Drive Integration This workflow creates a powerful RAG (Retrieval-Augmented Generation) chatbot that can process, store, and interact with documents from Google Drive using Qdrant vector storage and Google's Gemini AI. How It Works Document Processing & Storage 📚 Retrieves documents from a specified Google Drive folder Processes and splits documents into manageable chunks Extracts metadata using AI for enhanced search capabilities Stores document vectors in Qdrant for efficient retrieval Intelligent Chat Interface 💬 Provides a conversational interface powered by Google Gemini Uses RAG to retrieve relevant context from stored documents Maintains chat history in Google Docs for reference Delivers accurate, context-aware responses Vector Store Management 🗄️ Features secure delete operations with human verification Includes Telegram notifications for important operations Maintains data integrity with proper version control Supports batch processing of documents Setup Steps Configure API Credentials: Set up Google Drive & Docs access Configure Gemini AI API Set up Qdrant vector store connection Add Telegram bot for notifications Add OpenAI Api Key to the 'Delete Qdrant Points by File ID' node Configure Document Sources: Set Google Drive folder ID Define Qdrant collection name Set up document processing parameters Test and Deploy: Verify document processing Test chat functionality Confirm vector store operations Check notification system This workflow is ideal for organizations needing to create intelligent chatbots that can access and understand large document repositories while maintaining context and providing accurate responses through RAG technology.
by Oskar
This workflow with AI agent is designed to navigate through the page to retrieve specific type of information (in this example: social media profile links). The agent is equipped with 2 tools: text tool:** to retrieve all the text from the page, URLs tool:** to extract all possible links from the page. 💡 You can edit prompt and JSON schema connected to the agent in order to return other data then social media profile links. 👉 This workflow uses Supabase as storage (input/output). Feel free to change it to any other database of your choice. 🎬 See this workflow in action in my YouTube video. How it works? The workflow uses the input URL (website) as a starting point to retrieve the data (e.g. example.com). Using the "URLs tool", the agent is able to retrieve all links from the page and navigate to them. For example, if you want to retrieve contact information, agent will try to find a subpage that might contain this information (e.g. example.com/contact) and extract the information using the text tool. Set up steps Connect database with input data (website addresses) or pin sample data to trigger node. Configure the crawling agent to retrieve the desired data (e.g. modify prompt and/or parsing schema). Set credentials for OpenAI. Optionally: split agent tools to separate workflows. If you like this workflow, please subscribe to my YouTube channel and/or my newsletter.
by David Ashby
Complete MCP server exposing 1 Listing API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add Listing API credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the Listing API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to https://api.ebay.com{basePath} • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (1 total) 🔧 Item_Draft (1 endpoints) • POST /item_draft/: Create eBay Listing Draft 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native Listing API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by Vadym Nahornyi
> ⚠️ Multi-language WhatsApp Error Notifier Get instant WhatsApp alerts when any workflow fails — perfect for mobile-first monitoring and fast incident response. ✅ No coding required ✅ Works with any workflow via Error Workflow ✅ Step-by-step setup instructions included in: 🇬🇧 English 🇪🇸 Español 🇩🇪 Deutsch 🇫🇷 Français 🇷🇺 Русский 📦 What This Template Does This template sends real-time WhatsApp notifications when a workflow fails. It uses the WhatsApp Business Cloud API to deliver a preformatted error message directly to your phone. The message includes: Workflow name Error message Last executed node Example message: Error on WorkFlow: {{ $json.workflow.name }} Message: {{ $json.execution.error.message }} lastNodeExecuted: {{ $json.execution.lastNodeExecuted }} ⚙️ Prerequisites Before using this template, make sure you have: A verified Facebook Business account Access to WhatsApp Business Cloud API A sender phone number (registered in Meta) An access token (used as credentials in n8n) A pre-approved message template (or be within the 24h session window) More info from Meta Docs → 🚀 How to Use Open the template and insert your WhatsApp credentials Enter your target phone number (e.g. your own) in international format Customize the message body if needed Save the workflow but do not activate it In any other workflow → open Settings → set this as your Error Workflow 🌐 Multi-language Setup Guide Included This template includes full setup instructions with screenshots and message formatting help in: 🇬🇧 English 🇪🇸 Español 🇩🇪 Deutsch 🇫🇷 Français 🇷🇺 Русский Choose your language inside the embedded sticky note in the workflow.
by Marcel Claus-Ahrens
Instructions This automation enables you to just upload any Image (via Form) of a Logo Sheet, containing multiple Images of Product Logos (most likely) which brings them in some context to one another. After submitting an AI-Agent eats that Logo Sheet, turning it into an List of "Productname" and "Attributes", also checks if Tools are kind of similar to another, given the Context of the Image. We utilize AI Vision capabilities for that. NOTE: It might not be able to extract all informations. For a "upload and forget it" Workflow it works for me. You can even run it multiple times, to be sure. But if you need to make sure it extracts everything you might need to think about an Multi-Agent Setup with Validation-Agent Steps. Once the Agent finishes the extraction, it will traditionally and deterministicly add those Attributes to Airtable (Creates those, if not already existing.) and also Upserts the Tool Informations. It uses MD5 Hashes for turning Product Names into.. something fancy really, you could also use it without that, but I wanted to have something that looks atleast like an ID. Setup Set Up the Airtable like shown below. Update and set Credentials for all Airtable Nodes. Check or Adjust the Prompt of the Agent matching your use-case. Activate the Workflow. Open the Form (default: https://your-n8n.io/form/logo-sheet-feeder) Enjoy growing your Airtable. Enjoy the workflow! ❤️ let the work flow — Workflow Automation & Development
by David Ashby
Complete MCP server exposing all Mandrill Tool operations to AI agents. Zero configuration needed - all 2 operations pre-built. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works • MCP Trigger: Serves as your server endpoint for AI agent requests • Tool Nodes: Pre-configured for every Mandrill Tool operation • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Uses official n8n Mandrill Tool tool with full error handling 📋 Available Operations (2 total) Every possible Mandrill Tool operation is included: 💬 Message (2 operations) • Send a message based on a template • Send a message based on HTML 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Resource IDs and identifiers • Search queries and filters • Content and data payloads • Configuration options Response Format: Native Mandrill Tool API responses with full data structure Error Handling: Built-in n8n error management and retry logic 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • Other n8n Workflows: Call MCP tools from any workflow • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Complete Coverage: Every Mandrill Tool operation available • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n error handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by Max aka Mosheh
How it works • Automates multi-platform social media posting (Instagram, YouTube, TikTok, etc.) using AI-generated content • Integrates Airtable, n8n, and Blotato for full content scheduling and publishing • Supports both image and video uploads with dynamic text and account routing Set up steps • Takes ~15–30 minutes to set up depending on how many platforms you connect • Requires Airtable personal access token and Blotato API key • Uses sticky notes throughout the workflow to explain config, tokens, and troubleshooting clearly