by Audun
Who is this for? This workflow is tailored for content creators, artists, and developers who use Ko-fi to receive financial support through donations, subscriptions, or product sales. Use case This workflow automates the process of receiving and categorizing payment notifications from Ko-fi, ensuring that creators can focus on their work rather than administrative tasks. What this workflow does Webhook Reception**: The workflow listens for incoming payment notifications from Ko-fi via a configured webhook. Token Verification**: It validates incoming requests to ensure they originate from Ko-fi using a verification token for enhanced security. Type Differentiation**: It categorizes payments into types—donations, subscriptions, and shop orders—allowing for tailored handling for each payment type. Custom Response Options**: Depending on the payment type received, the workflow activates specific actions or processes, enabling seamless integration with other applications or services. Setup Webhook Configuration: Access the Webhook node within the workflow and take note of your unique webhook URL. Visit your Ko-fi webhooks management page at Ko-fi Webhooks Management and input this URL. Verification Token Setup: In your Ko-fi account, locate the verification token in the advanced settings. Input this token in the Prepare node of your n8n workflow. Enable the Workflow: Activate the workflow in n8n to start listening for incoming webhook notifications. Testing: Use the test feature in the Ko-fi webhooks settings to send a test webhook to ensure everything is functioning as expected. How to customize this workflow to your needs Add Actions for Each Payment Type**: You can modify the Donation, Subscription, and Shop Order nodes to include actions such as sending emails, logging payments within a database, or triggering notifications. Enhance Security Measures**: You can further refine the Check token node to include additional checks or to log all incoming webhook requests for monitoring. Integration with Other Services**: Consider linking this workflow with messaging platforms (e.g., Slack, Discord) or CRM tools to keep your supporters informed or to manage relationships more effectively. Custom Fields**: If needed, adjust the fields captured in the Subscription and Shop Order nodes to include more data or different parameters based on your specific use case.
by Niklas Hatje
Who this is for This template is for everyone that wants to download their n8n Cloud invoices automatically as a PDF instead of downloading them manually. How it works This workflow checks your Gmail inbox for new n8n invoice emails from n8n's payment provider Paddle. Once it finds something, it converts the URL into a PDF using pdflayer and saves it in Google Drive. Setup Setup your Gmail and Google Drive credentials Create a free account at https://pdflayer.com/ Insert your pdflayer API key into the Setup node Insert the URL to the wanted drive folder into the setup node (make sure to remove everything after the ?) How to adjust it to your need Instead of saving the PDF in Google drive, you could also save it in your local system, any other storage provider or send the PDF automatically to the right person in your company.
by Onur
Description This workflow empowers you to effortlessly get answers to your n8n platform questions through an AI-powered assistant. Simply send your query, and the assistant will search documentation, forum posts, and example workflows to provide comprehensive, accurate responses tailored to your specific needs. > Note: This workflow uses community nodes (n8n-nodes-mcp.mcpClientTool) and will only work on self-hosted n8n instances. You'll need to install the required community nodes before importing this workflow. ! What does this workflow do? This workflow streamlines the information retrieval process by automatically researching n8n platform documentation, community forums, and example workflows, providing you with relevant answers to your questions. Who is this for? New n8n Users**: Quickly get answers to basic platform questions and learn how to use n8n effectively Experienced Developers**: Find solutions to specific technical issues or discover advanced workflows Teams**: Boost productivity by automating the research process for n8n platform questions Anyone** looking to leverage AI for efficient and accurate n8n platform knowledge retrieval Benefits Effortless Research**: Automate the research process across n8n documentation, forum posts, and example workflows AI-Powered Intelligence**: Leverage the power of LLMs to understand context and generate helpful responses Increased Efficiency**: Save time and resources by automating the research process Quick Solutions**: Get immediate answers to your n8n platform questions Enhanced Learning**: Discover new workflows, features, and best practices to improve your n8n experience How It Works Receive Request: The workflow starts when a chat message is received containing your n8n-related question AI Processing: The AI agent powered by OpenAI GPT-4o analyzes your question Research and Information Gathering: The system searches across multiple sources: Official n8n documentation for general knowledge and how-to guides Community forums for bug reports and specific issues Example workflow repository for relevant implementations Response Generation: The AI agent compiles the research and generates a clear, comprehensive answer Output: The workflow provides you with the relevant information and step-by-step guidance when applicable n8n Nodes Used When chat message received (Chat Trigger) OpenAI Chat Model (GPT-4o mini) N8N AI Agent n8n-assistant tools (MCP Client Tool - Community Node) n8n-assistant execute (MCP Client Tool - Community Node) Prerequisites Self-hosted n8n instance OpenAI API credentials MCP client community node installed MCP server configured to search n8n resources Setup Import the workflow JSON into your n8n instance Configure the OpenAI credentials Configure your MCP client API credentials In the n8n-assistant execute node, ensure the parameter is set to "specific" (corrected from "spesific") Test the workflow by sending a message with an n8n-related question MCP Server Connection To connect to the MCP server that powers this assistant's research capabilities, you need to use the following URL: https://smithery.ai/server/@onurpolat05/n8n-assistant This MCP server is specifically designed to search across three types of n8n resources: Official documentation for general platform information and workflow creation guidance Community forums for bug-related issues and troubleshooting Example workflow repositories for reference implementations Configure this URL in your MCP client credentials to enable the assistant to retrieve relevant information based on user queries. This workflow combines the convenience of chat with the power of AI to provide a seamless n8n platform research experience. Start getting instant answers to your n8n questions today!
by Amjid Ali
Automated Weekly Project Cost Reports with MySQL and Outlook HTML Emails 🧠 Use Case Need to keep your finance or operations team updated on missing project costs? This practical automated report workflow does the job without AI — and saves hours weekly. Runs on a weekly schedule Queries your MySQL database for projects missing cost data Filters by budgeted_project_cost IS NULL Generates a clean HTML email report Sends it through Microsoft Outlook to relevant teams 🚀 How It Works Schedule Trigger – Runs every Monday at 8 AM MySQL Node – Connects and runs SQL to fetch project data missing budgeted_project_cost Switch Node – Routes logic based on cost_center (e.g., Retail, Service, Projects) Outlook Nodes – Sends formatted HTML emails; each node handles a specific group Dynamic Content – Inject values using mustache tags like {{ $json.project_name }} 🔧 Setup Instructions MySQL Setup: Ensure the MySQL node is connected using a valid credential set: Hostname/IP Port (default: 3306) Database name Username with SELECT permissions Password Query Example: SELECT project_name, cost_center FROM tabProject WHERE status = 'Open' AND project_type = 'External' AND budgeted_project_cost IS NULL; Outlook Configuration: Connect your Microsoft Outlook node using OAuth2 credentials. Rename each Outlook node clearly (e.g., Send Email - Retail, Send Email - Service). Switch Node: Modify cost center values as needed to match your organization (e.g., 'Retail', 'Service', 'Projects'). HTML Email Formatting: Customize the HTML message body using inline styles and mustache syntax. Sample: Missing Budgeted Cost Report Project: {{ $json.project_name }} Cost Center: {{ $json.cost_center }} Recipients: Replace amjid@amjidali.com with the actual email addresses of the concerned teams. 📘 Read More 👉 Why Simple Workflows Work 📺 Demo & Tutorial 🎥 Watch the video walkthrough: https://youtube.com/@syncbricks 👤 About the Creator Amjid Ali 🌐 amjidali.com 📘 n8n Book 🎓 Learn n8n > “Start simple, scale smart. Even basic workflows like this can save your team hours!” — Amjid Ali
by Yang
Workflow Description This workflow helps content creators automatically repurpose YouTube videos into SEO-friendly blog posts. It extracts the video transcript, uses AI to generate a full blog post with a relevant image, and sends the complete package via email, ready for publication. Prerequisites/Requirements This workflow relies on external AI services. You will need: OpenAI Account: Used for generating the blog post text (specifically mentioned using GPT-4o in the workflow notes). Credentials: Requires an API key from OpenAI. Cost: OpenAI API usage is typically paid based on the amount of text processed (tokens). Check OpenAI's current pricing. Setup: Sign up at OpenAI and obtain your API key. Dumpling AI Account: Used for retrieving YouTube video transcript and generating the blog post image. Credentials: Requires an API key from Dumpling AI. Cost: Dumpling AI offers 250 free credits to start with and different plans for different levels of usage. Check the pricing page for more details. Setup: Sign up at Dumpling AI and obtain your API key/credentials. Email Account: Credentials for the email service (e.g., Gmail) used to send the final result. How it works Input Video Details: You provide the YouTube video URL and your email address. Get Transcript: The workflow fetches the transcript of the specified YouTube video. Generate Content: An AI model crafts a blog post (title, description, body) based on the transcript. Create Image: Another AI model generates a suitable image for the blog post. Format & Package: The blog post is converted to HTML, and the image is prepared for sending. Email Result: The final HTML blog post and image are emailed to you. Set up steps Configure Variables: Enter the specific YouTube video URL and the recipient email address in the "Set Variables" node. Connect Credentials: Add your credentials for the services used (e.g., OpenAI for text generation, Dumpling AI for YouTube Transcript and AI image generation service). Connect Email Credentials: Authenticate your Gmail account (or chosen email provider) to allow the workflow to send the email. Take it to the next level Direct Publishing:** Instead of emailing the result, connect directly to your CMS (like WordPress, Ghost, Webflow) to automatically create a draft or publish the blog post. AI Agent Integration:** Replace the single "Generate Blog Post" step with an AI Agent for more sophisticated content generation, potentially researching topics or structuring the post section by section based on the transcript. Social Media Snippets:** Add steps to generate companion social media posts (e.g., for Twitter, LinkedIn) summarizing the blog post. Batch Processing:** Modify the trigger to read multiple YouTube URLs from a spreadsheet or database to convert videos in bulk. Enhanced SEO:** Refine the AI prompts to specifically target keywords or incorporate SEO best practices more deeply into the generated content. Multiple Image Options:** Generate several image variations and include them in the email or draft post for selection.
by Jimleuk
This n8n template reviews and audits recently active Google Drive files and reports on files with excessively open permissions. This shows how you can automate simple compliance tasks for access control management. File Sharing Permissions are routinely abused when access needs and scopes expand to many colleagues, clients and users. Often, granting excessively open permissions means you can get back to work rather than deal with numerous access request notifications. Whilst sometimes justified, the problem is that the permissions are rarely reverted to a safer setting at a later date when it is no longer needed. This template serves to improve your security posture by giving frequent reminders of these open files so that they can be actioned and not forgotten about. See example Audit Report here: https://docs.google.com/spreadsheets/d/1V2aiLhp3_nH7EBniMn7D0kFHg7-A5NjpDZXMhb4F5UI/edit?gid=503992967 How it works A scheduled trigger runs everyday to generate a new audit report. A new sheet is created in a designated Google Sheets document to store the day's results. The Google Drive node is used with Advanced Search params to fetch recently modified files for the user with each file result containing the current permission settings. The results are filtered for those with publicly accessible "anyone with link" and sharing with external users via domain. The results are then manipulated into rows so that we can append them to the Sheet we created earlier. The audit Google Sheet is updated with the results and an audit report is sent to the user to action. How to use Set the scheduled trigger to a more appropriate interval which works for you or your organisation. Consider using allowlists for organisations you frequently share with to reduce the number of false positives. The results can be forwarded to other security or analytical products as required. Requirements Google Drive for Document Management Google Sheet for Reports and Data Collection Gmail to Email Reports Customising the workflow Not using Google? Apply the same approach using Microsoft Sharepoint or Dropbox. If your security policies require it, you could automate fixing the file permissions as a proactive action instead and notify the user later.
by Ifeoluwa Ajetomobi
This workflow helps you stay updated with daily launches on Product Hunt. It automatically fetches product details (name, tagline, description, and website), checks if the website redirects to another URL, and logs the final information into a Google Sheet. Perfect for indie hackers, product managers, content curators, and anyone tracking daily launches. How It Works Schedule Trigger – Runs the workflow daily. Set Date – Captures today’s date in ISO format for filtering Product Hunt posts. HTTP Request (Product Hunt API) – Retrieves Product Hunt posts for the day using GraphQL. Extract Product Info (Code Node) – Parses the response to pull key details: Name Tagline Description Website URL HTTP Request (URL Check) – Follows each website URL to detect if it redirects. Merge Data – Combines product info with the final destination URL. Google Sheets Node – Appends all processed product info to your sheet. Pre-conditions A valid Product Hunt API token A Google account with access to Google Sheets A Google Sheet already created with the correct columns (see below) Connected Google Sheets and HTTP credentials in n8n Google Sheets Setup Your spreadsheet should include the following columns (in order): Name Tagline Description Original URL Final URL (after redirect) Ensure your Google Sheets node uses the correct Spreadsheet ID and Sheet Name. Setup Instructions Product Hunt API Auth: Replace {{YOUR_PRODUCT_HUNT_API_KEY}} in the HTTP Request headers: { "Authorization": "Bearer {{YOUR_PRODUCT_HUNT_API_KEY}}" } Google Sheets Node: Connect your Google account. Insert your Spreadsheet ID in the settings. Specify the sheet name (e.g., Daily Launches). Use the “Append” operation and map the 5 data fields accordingly. Notes Only fetches the first 10 posts for the day (can be extended). Consider adding Slack, Discord, or Email nodes to notify you of new entries. Useful for building launch databases, research, or content inspiration.
by Yang
Who is this for? This workflow is built for marketers, sales teams, agencies, virtual assistants, and anyone who regularly researches or contacts local businesses. It's ideal for building lead lists, tracking competitors, or creating location-specific outreach campaigns. What problem is this workflow solving? Instead of manually searching Google Maps and copying business info into spreadsheets, this automation pulls structured business data (e.g. restaurants, gyms, service providers) and logs it directly into Google Sheets. It saves hours of work and ensures cleaner, more usable data. What this workflow does The workflow takes a Google Maps search query (like "best restaurants in New York") and sends it to Dumpling AI. It returns a list of places including their name, address, website, phone number, rating, and more. Each result is split into a row and automatically added to a Google Sheet. Setup Dumpling AI Sign up at Dumpling AI Generate your API key In the HTTP Request node, select Header Auth and paste your key in the Authorization field Google Sheets Create a sheet with tab name Leads Add the following column headers to row 1: Name, Address, Phone number, Website, Rating, Price Level, Type, Booking Link, Position Connect your Google Sheets account and link this sheet in the node Customize the Query In the HTTP node, replace the query string (e.g., "best+restaurants+in+New+York") with your own search term Run It Use the manual trigger to test Optionally swap in a Schedule or Webhook node to run it automatically How to customize this workflow to your needs Change the search query to target different cities or business types Use filters to only save leads with a minimum rating or price level Add GPT to summarize listings or qualify leads Swap Google Sheets for Airtable or a CRM system for deeper integration
by Jimleuk
This n8n demonstrates how to build a simple Youtube MCP server to look up videos on Youtube and download their transcripts for research purposes. Youtube videos are a great source of new and updated information on a variety of cutting edge developments but they''re are not always simple to understand and lengthy videos may take too much time. Using this MCP server, you extract and feed their transcripts for your AI agents which then allows it to breakdown the content into manageble learnings and insights. How it works A MCP server trigger is used and connected to 3 custom workflow tools: Youtube Search, Youtube Transcripts and Usage Reports. Both Youtube tools use an external scraping service called APIFY.com. This is my preference as it's a much simpler interface and there are no rate limits. The Youtube Search fetches 10 results based on the user's query. The Youtube Transcripts downloads the subtitles from one or more given URL. The usage reports pulls in your monthly APIFY.com monthly spending and limits as a way to check your account. How to use This Apify Youtube MCP server allows any compatible MCP client to research YouTube videos for any desired topic. An Apify account is required however to connect and use the service. Connect your MCP client by following the n8n guidelines here - https://docs.n8n.io/integrations/builtin/core-nodes/n8n-nodes-langchain.mcptrigger/#integrating-with-claude-desktop Alternatively, connect any n8n AI agent with the MCP client tool. Try the following queries in your MCP client: "what is MCP?" "How can I use MCP in n8n?" "How can I use Apify's official MCP server?" Requirements APIFY.com for Youtube Scraping. This is a paid service but there is a $5 free tier which is ample for this template. MCP Client or Agent for usage such as Claude Desktop - https://claude.ai/download Customising this workflow Add as many APIFY.com actors as required for your use-case or users. Consider using Apify's official MCP server for 4000+ available tools. Remember to set the MCP server to require credentials before going to production and sharing this MCP server with others!
by Amjid Ali
Automate Digital Delivery After PayPal Purchase Using n8n A Complete Step-by-Step Guide to Seamless Template Delivery Built by Amjid Ali – SyncBricks Deliver personalized files instantly after PayPal transactions using n8n – without writing a single backend line. 🚀 What This n8n Workflow Does This automation template helps you automatically deliver a digital product (such as an n8n template or JSON file) to customers who pay via PayPal — within seconds. You can: Automatically extract customer info Identify what was purchased Send a clean, branded email with the product file Promote your other courses, books, and tools 📦 Use Case Example Product: AI-Powered Social Media Content Generator & Publisher When a customer buys this product through PayPal, this automation: Listens for a successful payment event Fetches order details via API Sends an HTML email with the template attached Promotes your other offerings with embedded links 🔧 Prerequisites You’ll need: An n8n instance (self-hosted or n8n Cloud) A PayPal developer account PayPal OAuth2 credentials configured in n8n Your product hosted as a downloadable .json file (Oracle, Dropbox, GitHub, etc.) SMTP email credentials in n8n 🧠 Step-by-Step Setup 1. Webhook Trigger Node: Webhook Listens for a POST request from PayPal’s webhook for PAYMENT.CAPTURE.COMPLETED events. 📌 Add the webhook to your PayPal Developer App > Webhooks. 2. Wait Node: Wait Adds a brief delay to ensure the payment is completely processed before continuing. 3. Filter Event Type Node: Switch Processes only when the event is PAYMENT.CAPTURE.COMPLETED. 4. Fetch Order Details Node: HTTP Request Retrieves the order information from PayPal's Orders API. URL format: https://api.paypal.com/v2/checkout/orders/{{ order_id }} 5. Extract Email & Product Info Node: Set Extracts first name, last name, email address, and the purchased item name. 6. Identify Product Purchased Node: Switch Checks if the product is “AI-Powered Social Media Content Generator & Publisher”. 7. Download Workflow File Node: HTTP Request Fetches the hosted workflow JSON from object storage (Oracle in this case). 8. Convert to Downloadable File Node: Code Converts the JSON content into a binary file and attaches it. 9. Send Custom Email Node: Send Email Sends a rich HTML email to the buyer with: Their name The file attachment Product name Helpful resource links: 📘 Mastering n8n Course on Udemy 📖 Step-by-Step Guide (n8n Book) 🎓 n8n Video Tutorials (Free Course) ☁️ Sign up for n8n Cloud – Use code AMJID10 🎥 YouTube Video Walkthrough 📚 Additional Learning Resources 🚀 My Full Automation Suite Explore more and master n8n with these resources: 🎓 Mastering n8n (Full Udemy Course) 📕 Get Your Step-by-Step Guide (n8n Book) 🎥 Get Step-by-Step Tutorials (Video Course) ☁️ Sign up for n8n Cloud 💡 Templates, Tools, and More 📺 YouTube Channel – SyncBricks 🙋 Need Help or Customization? Reach out! Email: amjid@amjidali.com LinkedIn: linkedin.com/in/amjidali Website: syncbricks.com
by Xavier
This workflow creates nested Google Drive folders from a path string (like Projects/Clients/Reports). It automatically handles the necessary folder lookups and creation steps required by Google Drive, then outputs the final folder's ID for immediate use. How it works This workflow streamlines the creation of nested folders in Google Drive: Input: Provide a root_folder_id and a path (e.g., Projects/Clients/Reports) as input. Path Parsing: The workflow splits the path into individual folder names (based on the / separator) Iterative Check & Create: Loops through each part of your path: Searches within the current parent folder (starting with the root_folder_id) for a subfolder matching the name. If found: Retrieves the existing folder's ID to use as the parent for the next iteration. If not found: Creates a new folder with that name inside the current parent folder and uses the new folder's ID as the parent for the next iteration. Output: Returns the Google Drive Folder ID of the very last folder in the specified path (e.g., the ID for Reports in the example above). This ID can then be directly used in subsequent n8n nodes to upload files, create documents, or perform other actions within that specific folder. Set up steps Setting up this workflow requires configuring the connection to Google Drive and knowing where to start creating folders: Connect Google Drive Account: Ensure you have a Google Drive credential configured in your n8n instance. Then link your credentials in the workflow: there are 2 Google Drive nodes that will need to be updated. Identify Starting Folder ID: Determine the Google Drive Folder ID where your nested structure should begin. You can either use the root of your Google Drive or a specific folder: To use the root of Google Drive, simply set root_folder_id to root (also called "My Drive" in the UI) To use a specific folder, open the folder in a webbrowser and look at the URL. The folder ID will be in the last part of the URL: https://drive.google.com/drive/folders/THIS_IS_THE_FOLDER_ID. Prepare Inputs for Execution: When running the workflow (or triggering it), you will need to provide: google_drive_folder_id -> this is the root folder ID you identified in step 2. desired_path -> This is the path you want to create (e.g., Projects/Clients/Reports). Here's an example of how you can call this workflow in your other workflows:
by Luciano Gutierrez
Instagram Auto-Comment Responder with AI Agent Integration Version: 1.1.0 ‧ n8n Version: 1.88.0+ ‧ License: MIT A fully automated workflow for managing and responding to Instagram comments using AI agents. Designed to improve engagement and save time, this system listens for new Instagram comments, verifies and filters them, fetches relevant post data, processes valid messages with a natural language AI, and posts context-aware replies directly on the original post. Key Features 💬 AI-Driven Engagement: Intelligent responses to comments via a GPT-powered agent. ✅ Webhook Verification: Handles Instagram webhook handshake to ensure secure integration. 📦 Data Extraction: Maps incoming payload fields (user ID, username, message text, media ID) for processing. 🚫 Self-Comment Filtering: Automatically skips comments made by the account owner to prevent loops. 📡 Post Data Retrieval: Fetches the media’s id and caption from the Graph API (v22.0) before generating a reply. 🧠 Natural Language Processing: Uses a custom system prompt to maintain brand tone and context. 🔁 Automated Replies: Posts the AI-generated message back to the comment thread using Instagram’s API. 🧩 Modular Architecture: Clear separation of steps via sticky notes and dedicated HTTP Request and Agent nodes. Use Cases Social Media Automation**: Keep followers engaged 24/7 with instant, relevant replies. Community Building**: Maintain a consistent voice and tone across all interactions. Brand Reputation Management**: Ensure no valid comment goes unanswered. AI Customer Support**: Triage simple questions and direct followers to resources or support. Technical Implementation Webhook Verification Node: Webhook + Respond to Webhook Echoes hub.challenge to confirm subscription and secure incoming events. Data Extraction Node: Set Maps payload fields into structured variables: conta.id, usuario.id, usuario.name, usuario.message.id, usuario.message.text, usuario.media.id, endpoint. User Validation Node: Filter Skips processing if conta.id equals usuario.id (self-comments). Post Data Retrieval Node: HTTP Request (Get post data) GET https://graph.instagram.com/v22.0/{{ $json.usuario.media.id }}?fields=id,caption&access_token={{ credentials }} Captures the media’s caption for richer context in replies. AI Response Generation Nodes: AI Agent + OpenRouter Chat Model Uses a detailed system prompt with: Profile persona (expert in AI & automations, friendly tone). Input data (username, comment text, post caption). Filtering logic (spam, praise, questions, vague comments). Returns either the reply text or [IGNORE] for irrelevant content. Posting the Reply Node: HTTP Request (Post comment) POST {{ $json.endpoint }}/{{ $json.usuario.message.id }}/replies with message={{ $json.output }} Sends the AI answer back under the original comment. Instructions for Setup Import Workflow In n8n > Workflows > Import from File, upload the provided .json template. Configure Credentials Instagram Graph API (Header Auth or FacebookGraphApi) with instagram_basic, instagram_manage_comments scopes. OpenRouter/OpenAI API key for AI agent. Customize System Prompt Edit the AI Agent’s prompt to adjust brand tone, language (Brazilian Portuguese), length, or emoji usage. Test & Activate Publish a test comment on an Instagram post. Verify each node’s execution, ensuring the webhook, filter, data extraction, HTTP requests, and AI Agent respond as expected. Extend & Monitor Add sentiment analysis or lead capture nodes as needed. Monitor execution logs for errors or rate-limit events. Tags Social Media • Instagram Automation • Webhook Verification • AI Agent • HTTP Request • Auto Reply • Community Management