by NovaNode
Who is this for? This template is designed for internal support teams, product specialists, and knowledge managers in technology companies who want to automate ingestion of product documentation and enable AI-driven, retrieval-augmented question answering. What problem is this workflow solving? Support agents often spend too much time manually searching through lengthy documentation, leading to inconsistent or delayed answers. This solution automates importing, chunking, and indexing product manuals, then uses retrieval-augmented generation (RAG) to answer user queries accurately and quickly with AI. What these workflows do Workflow 1: Document Ingestion & Indexing Manually triggered to import product documentation from Google Docs. Automatically splits large documents into chunks for efficient searching. Generates vector embeddings for each chunk using OpenAI embeddings. Inserts the embedded chunks and metadata into a MongoDB Atlas vector store, enabling fast semantic search. Workflow 2: AI-Powered Query & Response Listens for incoming user questions (can be extended to webhook). Converts questions to vector embeddings and performs similarity search on MongoDB vector store. Uses OpenAI’s GPT-4o-mini model with retrieval-augmented generation to produce direct, context-aware answers. Maintains short-term conversation context using a memory buffer node. Setup Setting up vector embeddings Authenticate Google Docs and connect your Google Docs URL containing the product documentation you want to index. Authenticate MongoDB Atlas and connect the collection where you want to store the vector embeddings. Create a search index on this collection to support vector similarity queries. Ensure the index name matches the one configured in n8n (data_index). See the example MongoDB search index template below for reference. Setting up chat Configure the AI system prompt in the “Knowledge Base Agent” node to reflect your company’s tone, answering style, and any business rules. Update the workflow description and instructions to help users understand the chat’s purpose and capabilities. Connect the MongoDB collection used for vector search in the chat workflow and update the vector search index if needed to match your setup. Make sure Both MongoDB nodes (in ingestion and chat workflows) are connected to the same collection, with: An embedding field storing vector data, Relevant metadata fields (e.g., document ID, source), and The same vector index name configured (e.g., data_index). Search Index Example: { "mappings": { "dynamic": false, "fields": { "_id": { "type": "string" }, "text": { "type": "string" }, "embedding": { "type": "knnVector", "dimensions": 1536, "similarity": "cosine" }, "source": { "type": "string" }, "doc_id": { "type": "string" } } } }
by Hostinger
This n8n workflow template is designed to help system administrators and DevOps professionals monitor key resource usage metrics — CPU, RAM, and Disk — on a VPS (Virtual Private Server). The workflow automatically checks these resources every 15 minutes and sends an email alert if any resource usage exceeds the 80% threshold. This proactive monitoring helps maintain optimal server performance and prevents resource-related downtimes. Who This Workflow Is For • System Administrators managing Linux-based servers who need to ensure their systems are running smoothly without manual monitoring. • DevOps Professionals who manage multiple environments and need automated tools to alert them to potential issues before they affect operations. • IT Support Teams who require an easy way to keep tabs on server health across an organization’s infrastructure. How It Works Schedule Trigger: The workflow is triggered every 15 minutes by a Cron node. Resource Checks: Separate SSH Command nodes are configured to execute specific commands that check the current usage of RAM, Disk, and CPU. Data Aggregation: The results from each check are merged using a Merge node, which combines the data into a single payload for analysis. Threshold Analysis: A Function node evaluates whether any resource’s usage exceeds the predefined 80% threshold. Alerts: If any metric exceeds the threshold, an email alert is sent through an Email node, ensuring that administrators can react promptly to potential issues. Setup Steps Configure SSH Nodes: Update each SSH node with the appropriate credentials and target server details where the resource checks will be performed. Set Thresholds: If different sensitivity levels are required, review and adjust the resource usage thresholds within the Function node. Email Configuration: Enter the correct email addresses in the Email node for where alerts should be sent. Ensure that your email-sending credentials and server details are correctly configured.
by n8n Team
This workflow creates/updates/deletes a Notion database page when an issue is created/updated/deleted in Jira. Subsequent updates to the issue's title or status in Jira are updated in the Notion database. If you require more fields to send to Notion, this template is easily extendible which will be described in setup. The Notion database will require setup before the workflow can be used. Prerequisites Notion account and Notion credentials. Jira account and Jira credentials. How it works When a new issue is created in Jira, the workflow creates a new page in the Notion database will all the required fields. When the issue's title or status is updated in Jira, the workflow updates the specific Notion database page identified by the "Issue Key" field in Notion. If the status in Jira is set to "Done", the workflow will mark the Notion database page "Done" field as true. When the issue is deleted in Jira, the workflow archives the Notion database page. Setup This workflow requires that you set up a Notion database. To do so, follow the steps below: In Notion, create a new database. Add the following columns to the database: Done (with type "Checkbox") Title (renamed from "Name") Status (with the following options: "To Do", "In Progress", "Done") Link (with type "URL") Issue ID (with type "Number") Issue Key (with type "Text") Add any other fields you require to the database. Your database should look something like this Share the database to n8n. By default, the workflow will fill all the fields provided above, except for any other additional fields you add.
by Srinivasan KB
This n8n workflow provides a ready-to-use API endpoint for extracting structured data from images. It processes an image URL using an AI-powered OCR model and returns the extracted details in a structured JSON format. Use Cases Document OCR** – Extract details from ID cards, invoices, receipts, etc. Text Extraction from Images** – Process screenshots, scanned documents, and photos. Automated Form Processing** – Digitize and capture information from paper forms. Business Card Data Extraction** – Extract names, emails, and phone numbers from business cards. How It Works Send a GET request with an image URL and define the required extraction parameters. The image is converted to base64 for processing. The AI model (Gemini API - Flash Lite) extracts relevant text. The response returns structured JSON data containing only the requested fields. Features ✔️ No-Code API Setup – Easily integrate into any application. ✔️ Customizable Extraction – Modify the request parameters to fit your needs. ✔️ AI-Powered OCR – Uses advanced models for accurate text recognition. ✔️ Automated Processing – Ideal for document processing and digitization. Integration Works with any frontend/backend system that supports API calls. Can be used for workflow automation in CRM, ERP, and document management solutions. Supports further customization based on specific OCR requirements.
by Angel Menendez
Who is this for? This subworkflow is ideal for developers and automation builders working with UniPile and n8n to automate message enrichment and LinkedIn lead routing. What problem is this workflow solving? UniPile separates personal and organization accounts into two different API endpoints. This flow handles both intelligently so you're not missing sender context due to API quirks or bad assumptions. What this workflow does This subworkflow is used by: LinkedIn Auto Message Router with Request Detection** LinkedIn AI Response Generator with Slack Approval** It receives a message sender ID and tries to enrich it using UniPile's /people and /organizations endpoints. It returns a clean, consistent profile object regardless of which source was used. Setup Generate a UniPile API token and save it in your n8n credentials Make sure this subworkflow is triggered correctly by your parent flows Test both people and organization lookups to verify responses are normalized How to customize this workflow to your needs Add a secondary enrichment layer using tools like Clearbit or FullContact Customize the fallback logic or error handling Expand the returned data for more AI context or user routing (e.g., job title, region)
by Abrar Sami
Auto-generate & post content using AI This workflow helps you create daily content using just a topic prompt. It writes a tweet, generates an image, and publishes across Twitter, Facebook, and LinkedIn — all on autopilot. How it works Triggers daily at 10 PM to start the flow Uses OpenAI to generate a niche topic title Writes a short-form post (tweet style) with hashtags Generates a Japanese anime-style image for visual context Saves everything in Google Sheets Publishes automatically on Twitter, LinkedIn, and Facebook Set up steps You’ll need OpenAI, Google Sheets, and social media credentials (Twitter, Facebook, LinkedIn) Takes about 10–15 minutes to configure if you already have the credentials ready Make sure your Sheet and API keys are properly linked before activating 📝 Keep detailed notes inside the workflow with sticky notes for easier handoff or collaboration.
by Tausif
Guidebook: How the Website ChatBot Template Works Chapter 1: Introduction & Objectives This guidebook provides a comprehensive walkthrough of the Website ChatBot developed using n8n and OpenAI. The chatbot is designed to qualify real estate leads and encourage site visits for the Alcove New Kolkata Sangam project through personalized, intelligent conversations. Chapter 2: Tools Required 1. n8n Workflow Automation Tool An open-source workflow builder to automate data flows between services. 2. OpenAI Account with GPT-4o-mini Access For generating AI-based chatbot responses. 3. Web Chat Widget Frontend integration that sends messages via webhook to the chatbot. Chapter 3: Workflow Breakdown Step 1: Webhook Receives POST requests from the chat widget. Endpoint: /webhook/chatbot-webhook Step 2: Set User Message Extracts message from the JSON body. Stores it as user_message. Step 3: Memory Setup Uses session ID to track conversation across messages. Step 4: OpenAI Chat Model GPT-4o-mini processes queries using the defined agent prompt. Step 5: AI Agent (Khusboo) Persona of a pre-sales agent. Uses AIDA + BANT + SPIN + PAS frameworks. Shares videos, responds in Hinglish, schedules site visits. Step 6: Respond to Webhook Formats the chatbot's reply into a JSON response. Chapter 4: Strategy & Psychology Behind Responses | Framework | Purpose | | --------- | ---------------------------------------------------- | | AIDA | Capture attention, interest, desire, action | | BANT | Qualify Budget, Authority, Need, Timing | | SPIN | Understand user's Situation, Problems, Implications | | PAS | Tackle objections using Problem, Agitation, Solution | The chatbot aims to qualify leads and gently move them toward booking a site visit without pushing or over-informing. Chapter 5: Setup Instructions A. n8n Workflow Setup Import the JSON workflow. Ensure OpenAI credentials are set up. Enable webhook at /webhook/chatbot-webhook. B. Frontend Widget Integration Send message as POST to the webhook with structure: { "message": "Looking for 2 BHK", "session_id": "user123" } Chapter 6: Testing & Troubleshooting Test via Postman Send sample request to verify AI response. Common Issues | Issue | Fix | | ---------------- | ----------------------------------- | | No response | Check webhook URL or credentials | | Repeated replies | Ensure memory node is active | | Wrong language | Check system message language rules | Chapter 7: Sample Conversations User: Hi, I’m looking for a home near the Ganga. Bot: Namaste! Main Khusboo hoon, Alcove New Kolkata Sangam se. Aapka naam kya hai? User: Rajat. Bot: Great Rajat! Kya aap apne family ke saath shift hone ka plan kar rahe ho? ... (continues using frameworks) Chapter 8: FAQs & Maintenance Tips Q: Can I update the AI agent persona? A: Yes, by modifying the system message inside the AI Agent node. Q: How do I share new videos or links? A: Add them in the sharingVideos or UserRequests section in the system message. Q: How to scale this for multiple projects? A: Duplicate the workflow and update the aboutProject and links accordingly. End of Guidebook.
by Alex Kim
Automate Video Creation with Luma AI Dream Machine and Airtable (Part 2) Description This is the second part of the Luma AI Dream Machine automation. It captures the webhook response from Luma AI after video generation is complete, processes the data, and automatically updates Airtable with the video and thumbnail URLs. This completes the end-to-end automation for video creation and tracking. 👉 Airtable Base Template 👉 Tutorial Video Setup 1. Luma AI Setup Ensure you’ve created an account with Luma AI and generated an API key. Confirm that the API key has permission to manage video requests. 2. Airtable Setup Make sure your Airtable base includes the following fields (set up in Part 1): Use the Airtable Base Template linked above to simplify setup. Generation ID** – To match incoming webhook data. Status** – Workflow status (e.g., "Done"). Video URL** – Stores the generated video URL. Thumbnail URL** – Stores the thumbnail URL. 3. n8n Setup Ensure that the n8n workflow from Part 1 is set up and configured. Import this workflow and connect it to the webhook callback from Luma AI. How It Works 1. Webhook Trigger The Webhook node listens for a POST response from Luma AI once video generation is finished. The response includes: Video URL – Direct link to the video. Thumbnail URL – Link to the video thumbnail. Generation ID – Used to match the record in Airtable. 2. Process Webhook Data The Set node extracts the video data from the webhook response. The If node checks if the video URL is valid before proceeding. 3. Store in Airtable The Airtable node updates the record with: Video URL – Direct link to the video. Thumbnail URL – Link to the video thumbnail. Status – Marked as "Done." Uses the Generation ID to match and update the correct record. Why This Workflow is Useful ✅ Automates the completion step for video creation ✅ Ensures accurate record-keeping by matching generation IDs ✅ Simplifies the process of managing and organizing video content ✅ Reduces manual effort by automating the update process Next Steps Future Enhancements** – Adding more complex post-processing, video trimming, and multi-platform publishing.
by Khairul Muhtadin
The Error Notification workflow is designed to instantly notify you whenever any other n8n workflow encounters an error, using popular communication channels like Telegram and Gmail—with optional support for Discord, Slack, and WhatsApp. 💡 Why Use Error Notification workflow? Immediate Awareness:** Get instant alerts when workflows fail, preventing unnoticed errors and downtime. Multi-Channel Flexibility:** Notify your team via Telegram, Gmail, and optionally Slack, Discord, or WhatsApp. Detailed Context:** Receive rich error information including the error message, node name, time, and execution link for quicker fixes. Easy Integration:** Built with native n8n nodes and customizable code, simple to adopt without complex setup. Open Source & Free:** Use and adapt this workflow at no cost, making professional error monitoring accessible. ⚡ Who Is This For? n8n Workflow Developers:** Quickly spot and respond to automation issues in development or production. Operations Teams:** Maintain uptime and swiftly troubleshoot errors across multiple workflows. Small to Medium Businesses:** Gain professional error alerting without expensive monitoring tools. Automation Enthusiasts:** Enhance your automation reliability with real-time failure notifications. ❓ What Problem Does It Solve? This workflow embedd error detection and notification directly within your n8n instance. It automates the process of catching errors as they occur, compiling meaningful context, and delivering it instantly via your preferred messaging platforms. This drastically reduces your response time to issues and streamlines error management, improving your automation reliability and operational confidence. 🔧 What This Workflow Does ⏱ Trigger: Listens for any error generated in your n8n workflows using the n8n Error Trigger node. 📎 Step 2: Executes a Code node that formats a detailed error message capturing workflow name, error node, description, timestamp, and an execution URL. 🔍 Step 3: Sends the formatted error notification to multiple communication channels: Telegram and Gmail by default, plus optionally Discord, Slack, and WhatsApp (disabled by default). 💌 Step 4: Delivers rich, parsed HTML-formatted messages to ensure error readability and immediate actionability. 🔐 Setup Instructions Import the provided .json file into your n8n instance (Cloud or self-hosted). Set up credentials: Gmail OAuth credentials for sending emails via Gmail node Telegram API credentials for Telegram notifications (Optional) Discord Webhook URL credential for Discord notifications (Optional) Slack Webhook credential for Slack notifications (Optional) WhatsApp connection credentials (if enabled) Customize the Code node if needed to adjust the error message format or target chat IDs. Update the chat IDs and recipient details in each notification node according to your channels. Test the workflow by manually triggering an error in another workflow to verify proper notifications. 🧩 Pre-Requirements Active n8n instance (cloud or self-hosted) with version supporting Error Trigger node Telegram bot credentials and chat ID (Optional) Gmail, Discord, Slack, or WhatsApp accounts and webhook credentials if you want to use those channels 🛠️ Customize It Further Enable and configure additional notification nodes like Slack or WhatsApp to fit your team's communication style. Customize the error message template in the Code node to include extra metadata or format it differently (e.g., markdown). Integrate with incident management tools via webhook nodes or create tickets automatically on error. 🧠 Nodes Used Error Trigger Code Telegram Gmail Discord (disabled) Slack (disabled) WhatsApp (disabled) Sticky Note (for description) 📞 Support Made by: khaisa Studio Tag: notification,error,monitoring,workflow,automation,alerts Category: Monitoring & Alerts Need a custom? Need a custom? contact me on LinkedIn or Web
by M Sayed
Stop guessing currency rates! Get a quick and clean exchange rate summary sent right to your phone. 📲 This workflow automatically checks the latest rates and builds a simple report for you. What it does: 🤑 Fetches the very latest exchange rates from an API. 🌍 Shows you what major currencies (like USD & EUR) are worth in your chosen local currency. ✍️ Creates a simple, easy-to-read report. 🚀 Delivers it straight to your Telegram! Setup is easy: All you need to do is set your base currency (e.g., 'EGP') and your Telegram Chat ID. Done! ✅
by Ahmed Alnaqa
Who is this template for? This workflow template is designed for content creators, researchers, educators, and professionals who need quick, accurate summaries of YouTube videos. It’s ideal for those looking to save time, extract key insights, or repurpose video content into concise formats for reports, studies, or social media. What does it do? The workflow automates the process of summarizing YouTube videos by extracting the transcript, analyzing the content, and generating a concise summary. It leverages AI tools to ensure accuracy and relevance, making it easier to digest lengthy videos in seconds. Why is it useful? This template saves hours of manual effort by automating video summarization, enabling users to focus on analyzing or sharing insights rather than watching entire videos. It’s particularly useful for staying updated with trends, conducting research, or creating content efficiently. How does it work? The workflow integrates with YouTube’s Transcript API powered by Apify Actor to fetch video transcripts, process the text using AI-powered summarization tools, and deliver a clear, concise summary. Setup Instructions You need an Apify account and an API key to connect with the Actor. Follow the steps below: Create a Free Account. Choose the appropriate Actor from the Apify search. Under the Integration tab, click on “Use API endpoints.” Select the API that best suits your needs.
by Laura Piraux
Use case This automation is for teams working in Notion. When you have a lot of back and forth in the comment section, it’s easy to lose track of what is going on in the conversation. This automation relies on AI to generate a summary of the comment section. How it works Every hour (the trigger can be adapted to your need and usecase), the automation checks if new comments have been added to the pages of your Notion database. If there are new comments, the comments are sent to an AI model to write a summary. The summary is then added to a predefined page property. The automation also updates a “Last execution” property. This prevents to re-generate the AI summary when no new comments have been received. Setup Define your Notion variables: Notion database, property that will hold the AI summary, property that will hold the last execution date of the automation. Set up your Notion credentials. Set up your AI model credentials (API key). How to adjust it to your needs Use the LLM model of your choice. In this template, I used Gemini but you can easily replace it by ChatGPT, Claude, etc. Adapt the prompt to your use case to get better summaries: specify the maximum number of characters, give an example, etc. Adapt the trigger to your needs. You could use Notion webhooks as trigger in order to run the automation only when a new comment is added (this setup is advised if you’re on n8n cloud version).