by Yusei Miyakoshi
Who's it for This template is for teams that want to stay updated on industry trends, tech news, or competitor mentions without manually browsing news sites. It's ideal for marketing, development, and research teams who use Slack as their central hub for automated, timely information. What it does / How it works This workflow runs on a daily schedule (default 9 AM), fetches the top articles from Hacker News for a specific keyword you define (e.g., 'AI'), and uses an AI agent with OpenRouter to generate a concise, 3-bullet point summary in Japanese for each article. The final formatted summary, including the article title, is then posted to a designated Slack channel. The entire process is guided by descriptive sticky notes on the canvas, explaining each configuration step. How to set up In the Configure Your Settings node, change the default keyword AI to your topic of interest and update the slack_channel to your target channel name. Click the OpenRouter Chat Model node and select your OpenRouter API key from the Credentials dropdown. If you haven't connected it yet, you will need to create a new credential. Click the Send Summary to Slack node and connect your Slack account using OAuth2 credentials. (Optional) Adjust the schedule in the Trigger Daily at 9 AM node to change how often the workflow runs. Activate the workflow. Requirements An n8n instance (Cloud or self-hosted). A Slack account and workspace. An OpenRouter API key stored in your n8n credentials. If self-hosting, ensure the LangChain nodes are enabled. How to customize the workflow Change the News Source:* Replace the *Hacker News* node with an *RSS Feed Read** node or another news integration to pull articles from different sources. Modify the AI Prompt:* In the *Summarize Article with AI** node, you can edit the system message to change the summary language, length, or tone. Use a Different AI Model:* Swap the *OpenRouter* node for an *OpenAI, **Anthropic, or any other supported chat model. Track Multiple Keywords:* Modify the workflow to loop through a list of keywords in the *Configure Your Settings** node to monitor several topics at once.
by Fahmi Fahreza
Automated Multi-Bank Balance Sync to BigQuery This workflow automatically fetches balances from multiple financial institutions (RBC, Amex, Wise, PayPal) using Plaid, maps them to QuickBooks account names, and loads structured records into Google BigQuery for analytics. Who’s it for? Finance teams, accountants, and data engineers managing consolidated bank reporting in Google BigQuery. How it works The Schedule Trigger runs weekly. Four Plaid API calls fetch balances from RBC, Amex, Wise, and PayPal. Each response splits out individual accounts and maps them to QuickBooks names. All accounts are merged into one dataset. The workflow structures the account data, generates UUIDs, and formats SQL inserts. BigQuery node uploads the finalized records. How to set up Add Plaid and Google BigQuery credentials, replace client IDs and secrets with variables, test each connection, and schedule the trigger for your reporting cadence.
by Intuz
This n8n template from Intuz provides a complete solution to automate the syncing of new subscribers from Google Sheets to MailerLite. It intelligently identifies and adds only new contacts, preventing duplicates and ensuring your email lists are clean and accurate. Who's this workflow for? Marketing Teams Email Marketers Small Business Owners Community Managers How it works 1. Read from Google Sheets: The workflow begins by reading all contact rows from your designated Google Sheet. 2. Check for Existing Subscribers: For each contact, it performs a search in MailerLite to check if a subscriber with that email address already exists. 3. Handle Duplicates: If the subscriber is found in MailerLite, the workflow stops processing that specific contact, preventing any duplicates from being created. 4. Create New Subscribers: If the contact is not found, the workflow proceeds to create a new subscriber in MailerLite, using all the details from the Google Sheet (like name, company, and country) and assigns them to the specified group. Setup Instructions 1. Google Sheets Setup: Connect your Google Sheets account to n8n. Create a sheet with the required columns: Email, first_name, last_name, Company, Country, and group_id. In the Get row(s) in sheet node, select your credentials and specify the Document ID and Sheet Name. 2. MailerLite Setup: Connect your MailerLite account to n8n using your API key. In both the Get a subscriber and Create subscriber... nodes, select your MailerLite credentials. Make sure the group_id values in your Google Sheet correspond to valid Group IDs in your MailerLite account. 3. Activate Workflow: Save the workflow and click "Execute workflow" to run the sync whenever you need to update your subscriber list. Connect with us Website: https://www.intuz.com/services Email: getstarted@intuz.com LinkedIn: https://www.linkedin.com/company/intuz Get Started: https://n8n.partnerlinks.io/intuz For Custom Worflow Automation Click here- Get Started
by Tushar Mishra
1. Data Ingestion Workflow (Left Panel – Pink Section) This part collects data from the ServiceNow Knowledge Article table, processes it into embeddings, and stores it in Qdrant. Steps: Trigger: When clicking ‘Execute workflow’ The workflow starts manually when you click Execute workflow in n8n. Get Many Table Records Fetches multiple records from the ServiceNow Knowledge Article table. Each record typically contains knowledge article content that needs to be indexed. Default Data Loader Takes the fetched data and structures it into a format suitable for text splitting and embedding generation. Recursive Character Text Splitter Splits large text (e.g., long knowledge articles) into smaller, manageable chunks for embeddings. This step ensures that each text chunk can be properly processed by the embedding model. Embeddings OpenAI Uses OpenAI’s Embeddings API to convert each text chunk into a high-dimensional vector representation. These embeddings are essential for semantic search in the vector database. Qdrant Vector Store Stores the generated embeddings along with metadata (e.g., article ID, title) in the Qdrant vector database. This database will later be used for similarity searches during chatbot interactions. 2. RAG Chatbot Workflow (Right Panel – Green Section) This section powers the Retrieval-Augmented Generation (RAG) chatbot that retrieves relevant information from Qdrant and responds intelligently. Steps: Trigger: When chat message received Starts when a user sends a chat message to the system. AI Agent Acts as the orchestrator, combining memory, tools, and LLM reasoning. Connects to the OpenAI Chat Model and Qdrant Vector Store. OpenAI Chat Model Processes user messages and generates responses, enriched with context retrieved from Qdrant. Simple Memory Stores conversational history or context to ensure continuity in multi-turn conversations. Qdrant Vector Store1 Performs a similarity search on stored embeddings using the user’s query. Retrieves the most relevant knowledge article chunks for the chatbot. Embeddings OpenAI Converts user query into embeddings for vector search in Qdrant.
by Harshil Agrawal
This workflow allows you to create, update and get a post using the Discourse node. Discourse node: This node creates a new post under a category. Based on your use-case, you can select a different category. Discourse1 node: This node updates the content of the post. Discourse2 node: This node fetches the node that we created using the Discourse node. Based on your use-case, you can add or remove nodes to connect Discourse to different services.
by Michael Taleb
Workflow Summary This automation keeps your Supabase vector database synchronized with documents stored in Google Drive, while also making the data contextual and vector based for better retrieval. When a file is added or modified, the workflow extracts its text, splits it into smaller chunks, and enriches each chunk with contextual metadata (such as summaries and document details). It then generates embeddings using OpenAI and stores both the vector data and metadata in Supabase. If a file changes, the old records are replaced with updated, contextualized content. The result is a continuously updated and context-aware vector database, enabling highly accurate hybrid search and retrieval. To setup 1. Connect Google Drive • Create a Google Drive folder to watch. • Connect your Google Drive account in n8n and authorize access. • Point the Google Drive Trigger node to this folder (new/modified files trigger the flow). 2. Configure Supabase • Please refer to the Setting Up Supabase Sticky Note. 3. Connect OpenAI (or your embedding model) • Add your OpenAI API key in n8n credentials.
by Davide
This workflow automates the entire process of creating, managing, and publishing AI-generated videos using OpenAI Sora2 Pro, Google Sheets, Google Drive, and YouTube. Advantages ✅ Fully Automated Video Pipeline From idea to YouTube publication with zero manual intervention after setup. ✅ Centralized Control via Google Sheets Simple spreadsheet interface — no need to use APIs or dashboards directly. ✅ AI-Powered Video Creation Uses OpenAI Sora2 Pro for generating professional-quality videos from text prompts. ✅ SEO-Optimized Titles with GPT-5 Automatically creates catchy, keyword-rich titles optimized for YouTube engagement. ✅ Cloud Integration Seamless use of Google Drive for file management and YouTube for publishing. ✅ Scalable and Repeatable Can handle multiple videos in sequence, triggered manually or at regular intervals. ✅ Error-Resilient and Transparent Uses conditional checks (“Completed?” node) and real-time updates in Google Sheets to ensure reliability and visibility. How it Works This workflow automates the entire process of generating AI videos and publishing them to YouTube, using a Google Sheet as the central control panel. Trigger & Data Fetch: The workflow is triggered either manually or on a schedule. It starts by reading a specific Google Sheet to find new video requests. A new request is identified as a row where the "PROMPT" and "DURATION" columns are filled, but the "VIDEO" column is empty. AI Video Generation: For each new request, it takes the prompt and duration, then sends a request to the Fal.ai Sora-2 Pro model via its API to generate the video. The system then enters a polling loop, checking the video generation status every 60 seconds until it is COMPLETED. Post-Processing & Upload: Once the video is ready, the workflow performs three parallel actions: Fetch Video & Upload to Drive: It retrieves the generated video file and uploads it to a specified folder in Google Drive for archiving. Generate YouTube Title: It sends the original prompt to OpenAI's GPT-5 (or another specified model) to generate an optimized, SEO-friendly title for the YouTube video. Publish to YouTube: It takes the generated video file and the AI-created title and uses the Upload-Post.com service to automatically publish the video to a connected YouTube channel. Update & Log: Finally, the workflow updates the original Google Sheet row with the URL of the archived video in Google Drive and the newly created YouTube video URL, providing a complete audit trail. Set up Steps To configure this workflow, follow these steps: Prepare the Google Sheet: Create a Google Sheet with at least these columns: PROMPT, DURATION, VIDEO, and YOUTUBE_URL. In the n8n "Get new video" and update nodes, configure the documentId and sheetName to point to your specific Google Sheet. Configure Fal.ai API Key: Create an account on fal.ai and obtain your API key. In both the "Create Video" and "Get status" HTTP Request nodes, set up the HTTP Header Authentication. Set the Name to Authorization and the Value to Key YOUR_API_KEY. Set up Upload-Post.com for YouTube: Create an account on Upload-Post.com and get your API key. Connect your YouTube channel as a "profile". In the "HTTP Request" node (for uploading), configure the Header Auth with Name: Authorization and Value: Apikey YOUR_UPLOAD_POST_API_KEY. Replace YOUR_USERNAME in the node's body parameters with the profile name you created on Upload-Post.com (e.g., test1). Configure OpenAI (Optional but Recommended): The "Generate title" node uses an OpenAI model. Ensure you have valid OpenAI API credentials set up in n8n for this node to function and create optimized titles. Finalize Paths and Activate: In the "Upload Video" node, specify the correct Google Drive folder ID where you want the videos to be saved. Once all credentials and paths are set, you can activate the workflow and set the "Schedule Trigger" node to run at your desired interval (e.g., every 5 minutes). Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Yaron Been
CTO Agent with Engineering Team Description Complete AI-powered engineering department with a Chief Technology Officer (CTO) agent orchestrating specialized engineering team members for comprehensive software development and technical operations. Overview This n8n workflow creates a comprehensive engineering department using AI agents. The CTO agent analyzes technical requests and delegates tasks to specialized agents for software architecture, DevOps, security, quality assurance, backend development, and frontend development. Features Strategic CTO agent using OpenAI O3 for complex technical decision-making Six specialized engineering agents powered by GPT-4.1-mini for efficient execution Complete software development lifecycle coverage from architecture to deployment Automated DevOps pipelines and infrastructure management Security assessments and compliance frameworks Quality assurance and test automation strategies Full-stack development capabilities Team Structure CTO Agent**: Technical leadership and strategic delegation (O3 model) Software Architect Agent**: System design, patterns, technology stack decisions DevOps Engineer Agent**: CI/CD pipelines, infrastructure automation, containerization Security Engineer Agent**: Application security, vulnerability assessments, compliance QA Test Engineer Agent**: Test automation, quality strategies, performance testing Backend Developer Agent**: Server-side development, APIs, database architecture Frontend Developer Agent**: UI/UX development, responsive design, frontend frameworks How to Use Import the workflow into your n8n instance Configure OpenAI API credentials for all chat models Deploy the webhook for chat interactions Send technical requests via chat (e.g., "Design a scalable microservices architecture for our e-commerce platform") The CTO will analyze and delegate to appropriate specialists Receive comprehensive technical deliverables Use Cases Full Stack Development**: Complete application architecture and implementation System Architecture**: Scalable designs for microservices and distributed systems DevOps Automation**: CI/CD pipelines, containerization, cloud deployment strategies Security Audits**: Vulnerability assessments, secure coding practices, compliance Quality Assurance**: Test automation frameworks, performance testing strategies Technical Documentation**: API documentation, system diagrams, deployment guides Requirements n8n instance with LangChain nodes OpenAI API access (O3 for CTO, GPT-4.1-mini for specialists) Webhook capability for chat interactions Optional: Integration with development tools and platforms Cost Optimization O3 model used only for strategic CTO decisions GPT-4.1-mini provides 90% cost reduction for specialist tasks Parallel processing enables simultaneous agent execution Code template library reduces redundant development work Integration Options Connect to development platforms (GitHub, GitLab, Bitbucket) Integrate with project management tools (Jira, Trello, Asana) Link to monitoring and logging systems Export to documentation platforms Contact & Resources Website**: nofluff.online YouTube**: @YaronBeen LinkedIn**: Yaron Been Tags #SoftwareEngineering #TechStack #DevOps #SecurityFirst #QualityAssurance #FullStackDevelopment #Microservices #CloudNative #TechLeadership #EngineeringAutomation #n8n #OpenAI #MultiAgentSystem #EngineeringExcellence #DevAutomation #TechInnovation
by Frankie Wong
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This n8n workflow template helps you automatically convert unstructured contact information—such as customer details copied from emails, web forms, or chat messages—into clean, structured JSON using an AI agent. What It Does: Accepts unstructured contact data via a Webhook (as form-data under the key prompt) Uses AI to intelligently extract key fields such as: Company Name First Name Last Name Address City Country Phone Fax Email Parses and formats the extracted data into a valid JSON object Prepares the output for seamless integration into systems like: Dolibarr Other ERP/CRM platforms Any service that consumes JSON via API or webhook Use Cases: Automate manual data entry from emails into your ERP system Clean and normalize contact data from various input sources Reduce human error in your customer onboarding workflows This template saves you time and ensures consistency across your business systems. Simply connect your systems and let the automation handle the rest.
by InfyOm Technologies
✅ What problem does this workflow solve? Salon staff often spend hours juggling appointment calls, managing bookings manually, and keeping track of customer preferences. This workflow automates your entire salon appointment system via WhatsApp, delivering a personalized and human-like booking experience using AI, memory, and Google Sheets. 💡 Main Use Cases 💁♀️ Offer personalized stylist recommendations by remembering customer preferences and past visits. 📅 Provide real-time availability and salon opening hour information. 📝 Book and update appointments directly from customer chat. 🔁 Simplify appointment changes by recalling previous booking details. 🧠 Enable context-aware, memory-driven conversations across multiple interactions. 🧠 How It Works – Step-by-Step 1. 📲 Chat Message Trigger The workflow is triggered whenever a customer sends a message to your WhatsApp salon assistant. 2. 🧠 Memory Buffer for Context Management The assistant uses a Memory Buffer to: Recognize returning customers Avoid repeating questions Maintain conversation flow across multiple sessions This enables a seamless and intelligent dialogue with each customer. 3. 💇 Stylist & Service Lookup When the customer asks for stylist suggestions, available time slots, or services: Extracts request details using AI Queries a Google Sheet containing: Stylist availability Service types Salon opening hours Returns personalized recommendations based on preferences and availability 4. ✅ Appointment Booking Collects all necessary info: Date, time, selected service, stylist, contact info Stores the appointment in Google Sheets Sends a confirmation message to the customer in WhatsApp 5. 🔄 Modify or Cancel Bookings Customers can update or cancel appointments Bot matches records by phone number Modifies or deletes the appointment in the sheet accordingly 🧩 Integrations Used WhatsApp Integration** (via Twilio, Meta API, or other connector) OpenAI/GPT Model** for natural conversation flow and extraction Google Sheets** as a simple and effective appointment database Memory Buffer** for ongoing context across chats 👤 Who can use this? Perfect for: 💇♀️ Salons and barbershops 💅 Spas and beauty centers 🧖♀️ Wellness studios 🛠 Developers building vertical AI assistants for SMBs If you're looking to modernize your booking process and impress customers with an AI-powered, memory-enabled WhatsApp bot—this workflow delivers. 🚀 Benefits ⏰ Save time for your staff 🧠 Offer truly personalized experiences 📲 Book appointments 24/7 via WhatsApp 📋 Keep all records organized in Google Sheets 🧘 Reduce human error and double bookings 📦 Ready to Launch? Just configure: ✅ Your WhatsApp number + webhook integration ✅ Google Sheet with stylist and service data ✅ OpenAI key for AI-powered conversation ✅ Memory Buffer to enable smart replies And your salon will be ready to offer automated, intelligent booking—right from a simple WhatsApp chat.
by Gabriel Santos
Who’s it for Teams and project managers who want to turn meeting transcripts into actionable Trello tasks automatically, without worrying about duplicate cards. What it does This workflow receives a transcript file in .txt format and processes it with AI to extract clear, concise tasks. Each task includes a short title, a description, an assignee (if mentioned), and a deadline (if available). The workflow then checks Trello for duplicates across all lists, comparing both card titles (name) and descriptions (desc). If a matching card already exists, the workflow returns the existing Trello card ID. If not, it creates a new card in the predefined default list. Finally, the workflow generates a user-friendly summary: how many tasks were found, how many already existed, how many new cards were created, and how many tasks had no assignee or deadline. Requirements A Trello account with API credentials configured in n8n (no hardcoded keys). An OpenAI (or compatible) LLM account connected in n8n. How to customize Adjust similarity thresholds for title/description matching in the Trello Sub-Agent. Modify the summary text to always return in your preferred language. Extend the Trello card creation step with labels, members, or due dates.
by Shelly-Ann Davy
Build authentic Reddit presence and generate qualified leads through AI-powered community engagement that provides genuine value without spam or promotion. 🎯 What This Workflow Does: This intelligent n8n workflow monitors 9 targeted subreddits every 4 hours, uses AI to analyze posts for relevance and lead potential, generates authentic helpful responses that add value to discussions, posts comments automatically, and captures high-quality leads (70%+ potential score) directly into your CRM—all while maintaining full Reddit compliance and looking completely human. ✨ Key Features: 6 Daily Checks: Monitors subreddits every 4 hours for fresh content 9 Subreddit Coverage: Customizable list of target communities AI Post Analysis: Determines relevance, intent, and lead potential Intelligent Engagement: Only comments when you can add genuine value Authentic Responses: AI-generated comments that sound human, not promotional Lead Scoring: 0-1.0 scale identifies high-potential prospects (0.7+ captured) Automatic CRM Integration: High-quality leads flow directly to Supabase Rate Limit Protection: 60-second delays ensure Reddit API compliance Native Reddit Integration: Official n8n Reddit node with OAuth2 Beginner-Friendly: 14+ detailed sticky notes explaining every component 🎯 Target Subreddits (Customizable): Insurance & Claims: r/Insurance - General insurance questions r/ClaimAdvice - Claim filing help r/AutoInsurance - Auto coverage discussions r/FloodInsurance - Flood damage queries r/PropertyInsurance - Property coverage Property & Home: r/homeowners - Property issues and claims r/RoofingContractors - Roof damage discussions Financial & Legal: r/PersonalFinance - Insurance decisions r/legaladvice - Legal aspects of claims 🤖 AI Analysis Components: Post Evaluation: Relevance score (0-100%) User intent detection Damage type identification (hail, water, fire, wind) Urgency level (low/medium/high) Lead potential score (0-1.0) Recommended services Engagement opportunity assessment Decision Criteria: Should engage? (boolean) Can we add value? (quality check) Is this promotional? (avoid spam) Lead worth capturing? (70%+ threshold) Typical Engagement Rate: 5-10% of analyzed posts (67-135 comments/day) 🔧 Technical Stack: Trigger: Schedule (every 4 hours, 6x daily) Reddit API: Native n8n node with OAuth2 AI Analysis: Supabase Edge Functions Response Generation: AI-powered contextual replies Lead Capture: Supabase CRM integration Rate Limiting: Wait node (60-second delays)