by Shashikanth
Source code, I maintain this worflow here. Usage Guide This workflow backs up all workflows as JSON files named in the [workflow_name].json format. Steps Create GitHub Repository Skip this step if using an existing repository. Add GitHub Credentials In Credentials, add the GitHub credential for the repository owner. Download and Import Workflow Import this workflow into n8n. Set Global Values In the Globals node, set the following: repo.owner: GitHub username of the repository owner. repo.name: Name of the repository for backups. repo.path: Path to the folder within the repository where workflows will be saved. Configure GitHub Nodes Edit each GitHub node in the workflow to use the added credentials. Workflow Logic Each workflow run handles files based on their status: New Workflow If a workflow is new, create a new file in the repository. Unchanged Workflow If the workflow is unchanged, skip to the next item. Changed Workflow If a workflow has changes, update the corresponding file in the repository. Current Limitations / Needs work Name Change of Workflows If a workflow is renamed or deleted in n8n, the old file remains in the repository. Deleted Workflows Deleted workflows in n8n are not removed from the repository.
by Danielle Gomes
This n8n workflow collects and summarizes news from multiple RSS feeds, using OpenAI to generate a concise summary that can be sent to WhatsApp or other destinations. Perfect for automating your daily news digest. 🔁 Workflow Breakdown: Schedule Trigger Start the workflow on your desired schedule (daily, hourly, etc.). 🟨 Note: Set the trigger however you wish. RSS Feeds (My RSS 01–04) Fetches articles from four different RSS sources. 🟨 Note: You can add as many RSS feeds as you want. Edit Fields (Edit Fields1–3) Normalizes RSS fields (title, link, etc.) to ensure consistency across different sources. Merge (append mode) Combines the RSS items into a single unified list. Filter Optionally filter articles by keywords, date, or categories. Limit Limits the analysis to the 10 most recent articles. 🟨 Note: This keeps the result concise and avoids overloading the summary. Aggregate Prepares the selected news for summarization by combining them into a single content block. OpenAI (Message Assistant) Summarizes the aggregated news items in a clean and readable format using AI. Send Summary to WhatsApp Sends the AI-generated summary to a WhatsApp endpoint via webhook (yoururlapi.com). You can replace this with an email service, Google Drive, or any other destination. 🟨 Note: You can send it to your WhatsApp API, email, drive, etc. No Operation (End) Final placeholder to safely close the workflow. You may expand from here if needed.
by Felix
How It Works This workflow automatically detects duplicate invoices from Gmail. Incoming PDF attachments are scanned by the easybits AI Extractor, then checked against the Master Finance File in Google Sheets. Duplicates trigger a Slack alert – new invoices get added to the sheet. Flow overview: Gmail picks up new emails labeled as invoices (polls every minute) The PDF attachment is extracted and converted to base64 easybits Extractor reads the document and returns structured data The invoice number is compared against all existing entries in Google Sheets If duplicate → Slack DM alert to felix.sattler If new → Invoice is appended to the Master Finance File Step-by-Step Setup Guide 1. Set Up Your easybits Extractor Pipeline Before connecting this workflow, you need a configured extraction pipeline on easybits. Go to extractor.easybits.tech and click "Create a Pipeline". Fill in the Pipeline Name and Description – describe the type of document you're processing (e.g. "Invoice / Receipt"). Upload a sample receipt or invoice as your reference document. Click "Map Fields" and define the following fields to extract: invoice_number (String) – Unique identifier of the invoice, e.g. IN-2026-0022514 total_amount (Number) – Total amount due on the invoice, e.g. 149.99 Click "Save & Test Pipeline" in the Test tab to verify the extraction works correctly. Go to Pipeline Details → View Pipeline and copy your Pipeline ID and API Key. 2. Connect the easybits Node in n8n Open the HTTP Request node in the workflow. Replace the Pipeline ID in the URL with your own. Set up a Bearer Auth credential with your API Key. > The node sends the PDF to your pipeline and receives the extracted fields back under json.data. 3. Connect Gmail Open the Gmail Trigger node. Connect your Gmail account via OAuth2. Create a label called invoice in Gmail (or use your preferred label). Update the label filter in the node to match your label. Make sure Download Attachments is enabled under Options. > The trigger polls every minute for new emails matching the label. 4. Connect Google Sheets Open the Check Google Sheets and Add to Master List nodes. Connect your Google Sheets account via OAuth2. Select your target spreadsheet (Master Finance File) and sheet. Make sure your sheet has at least these columns: Invoice Number and Final Amount (EUR). 5. Connect Slack Go to api.slack.com/apps and create a new Slack App. Under OAuth & Permissions, add these Bot Token Scopes: chat:write, chat:write.public, channels:read, groups:read, users:read, users.profile:read. Install the app to your workspace via Settings → Install App. Copy the Bot User OAuth Token and add it as a Slack API credential in n8n. Open the Slack: Alert Finance node and select the user or channel to receive duplicate alerts. 6. Activate the Workflow Click the "Active" toggle in the top-right corner of n8n to enable the workflow. Label an email with an invoice attachment as invoice in Gmail to test it end to end. Check your Google Sheet – a new row with the invoice number and amount should appear. Send the same invoice again – you should receive a Slack DM alerting you to the duplicate.
by Muzaffer AKYIL
Docker Registry Cleanup Template This template is designed to automatically clean up old image tags in the Docker registry and perform garbage collection. Features List all images in the registry Preserve the last 10 tags for each image (latest tag is always preserved) Delete old tags Email notification for Successful/Excused cancellation Registry garbage collection automation Failure notification in error conditions Prerequisites Docker Registry v2 API access Basic Authentication credentials SMTP email settings (for notifications) SSH node installed on n8n (for garbage collection) Installation 1. Identity Information Add the following credentials in n8n: HTTP Basic Auth**: For Registry access SSH Private Key**: For Garbage collection command Email SMTP**: For notifications 2. Set Variables Replace your-registry-url with your actual registry URL on all nodes: ‘url": ‘https://your.registry.com/v2/_catalog’. Customisation Retention Policy: Set the number of tags to be retained by changing the slice(0, 10) value in the Identify Tags to Remove node Schedule: Change the frequency of operation at the Trigger node Notification Content: Customise email templates according to your needs Notes Check DELETE operations before running in a test environment Make sure that the registry is not in read-only mode The registry may need to be put into maintenance mode for garbage collection Step Details: Retrieving image information:** The workflow starts by fetching a list of images and their associated tags from the Docker registry. Filtering and sorting:** The retrieved tags are then filtered and sorted based on specific criteria, such as creation date and tag name. Deleting old tags:** The workflow identifies old or unused tags and attempts to delete them from the registry. Sending notifications:** The workflow sends email notifications to inform the user about the status of the cleanup process, including any errors or successes. Executing additional cleanup tasks:** Finally, the workflow executes an SSH command on the Docker registry server to perform additional cleanup tasks, such as garbage collection. TL;DR In summary, this n8n template provides a robust and automated solution for managing and cleaning up Docker registries. By regularly running this workflow, users can ensure that their registry remains organized and efficient, and avoid running out of storage space.-
by AiAgent
What It Does This intelligent workflow simplifies the complex task of determining whether a website is legitimate or potentially a scam. By simply submitting a URL through a form, the system initiates a multi-agent evaluation process. Four dedicated AI agents—each powered by GPT-4o and connected to SerpAPI—analyze different dimensions of the website: domain and technical details, search engine signals, product and pricing patterns, and on-site content analysis. Their findings are then passed to a fifth AI agent, the Analyzer, powered by GPT-4o mini, which consolidates the data, scores the site on a scale of 1–10 for scam likelihood, and presents the findings in a clear, structured format for the user. Who It's For This workflow is ideal for anyone who needs to quickly and reliably assess the trustworthiness of a website. Whether you're a consumer double-checking a store before making a purchase, a small business owner validating supplier sites, a cybersecurity analyst conducting threat assessments, or a developer building fraud detection into your platform — this tool offers fast, AI-powered insights without the need for manual research or technical expertise. It's designed for both individuals and teams who value accurate, scalable scam detection. How It Works The process begins with a simple form submission where the user enters the URL of the website they want to investigate. Once submitted, the workflow activates four specialized AI agents—each powered by GPT-4o and connected to SerpAPI—to independently analyze the site from different angles: Agent 1 examines domain age, SSL certificates, and TLD trustworthiness. Agent 2 reviews search engine results, forum mentions, and public scam reports. Agent 3 analyzes product pricing patterns and brand authenticity. Agent 4 assesses on-site content quality, grammar, legitimacy of claims, and presence of business info. Each agent returns its findings, which are then aggregated and passed to a fifth AI agent—the Analyzer. This final agent, powered by GPT-4o mini, evaluates all the input, assigns a scam likelihood score from 1 to 10, and compiles a neatly formatted summary with organized insights and a disclaimer for context. Set UP You will need to obtain an Open AI API key from platform.openai.com/api-keys After you obtain this Open AI API key you will need to connect it to the Open AI Chat Model for all of the Tools agents (Analyzer, Domain & Technical Details, Search Engine Signals, Product & Pricing Patterns, and Content Analysis Tools Agents). You will now need to fund your Open AI account. GPT 4o costs ~$0.01 to run the workflow. Next you will need to create a SerpAPI account at https://serpapi.com/users/sign_up After you create an account you will need to obtain a SerpAPI key. You will then need to use this key to connect to the SerpAPI tool for each of the tools agents (Domain & Technical Details, Search Engine Signals, Product & Pricing Patterns, and Content Analysis Tools Agents) Tip: SerpAPI will allow you to run 100 free searches each month. This workflow uses ~5-15 SerpAPI searches per run. If you would like to utilize the workflow more than that each month, create multiple SerpAPI accounts and have an API key for each account. When you utilize all 100 free searches for an account, switch to the API key for another account within the workflow. Disclaimer This tool is designed to assist in evaluating the potential risk of websites using AI-generated insights. The scam likelihood score and analysis provided are based on publicly available information and should not be considered a definitive or authoritative assessment. This tool does not guarantee the accuracy, safety, or legitimacy of any website. Users should perform their own due diligence and use independent judgment before engaging with any site. N8N, OpenAI, its affiliates, and the creators of this workflow are not responsible for any loss, damages, or consequences arising from the use of this tool or the actions taken based on its results.
by David Olusola
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 📁 Google Drive MCP Workflow – AI-Powered File Management Automation 🚀 🧠 Overview A secure and intelligent n8n workflow that connects with Google Drive via MCP (Model Control Protocol). Ideal for AI agent tasks, compliance-driven storage, and document automation. 🌟 Key Features 🔒 Built-In Safety Backs up files before edits (timestamped) Supports rollback using file history Validates file size, type, and permissions 📁 Smart Organization Automatically converts file types (PDF, DOCX, etc.) Moves files to structured folders Auto-archives old files based on age or rules 🔄 MCP Integration Accepts standardized JSON via webhook Real-time execution for AI agents Fully customizable input (action, fileId, format, etc.) ✅ AI Callable MCP Actions These are the commands AI agents can perform via MCP: Download a file (with optional format conversion) Upload a new file to Google Drive Copy a file for backup Move a file to a specific folder Archive old or inactive files Organize documents into folders Convert files to a new format (PDF, DOCX, etc.) Retrieve and review file history for rollback 📝 Example Input { "action": "download", "fileId": "abc123", "folderPath": "/projects/clientA", "convertFormat": "pdf" } 🔐 Security & Performance OAuth2 secured access to Google Drive API No sensitive data stored in transit Real-time audit logs and alerts Batch-friendly with built-in rate limiting 📌 Ideal For Businesses automating file management AI Agents retrieving, sorting, converting, or archiving files Compliance teams needing file versioning and backups ⚙️ Requirements n8n + Google Drive API v3 MCP server + Webhook integration Google OAuth2 Credentials
by Raz Hadas
Description Transform your investment strategy with a fully automated, AI-driven trading bot. This workflow bridges the gap between AI-powered market insights and real-world trading by executing buy and sell orders directly through the Alpaca paper trading API. Designed to work in tandem with the Automated Stock Sentiment Analysis workflow, this solution takes the top-performing stocks based on daily news sentiment and automatically rebalances your portfolio. It's perfect for algorithmic traders, data-driven investors, and n8n enthusiasts who want to see their AI analysis translate into tangible actions, all while maintaining a comprehensive log of every transaction in Google Sheets. Key Features & Benefits Automated Trading Execution:** Automatically places buy and sell orders on the Alpaca paper trading platform without manual intervention. Sentiment-Driven Decisions:** Leverages the output from the sentiment analysis workflow to make informed decisions, selling positions with waning sentiment and buying into those with high positive sentiment. Dynamic Portfolio Rebalancing:** Intelligently calculates which positions to close and how to allocate the resulting funds into new, high-potential stocks. Paper Trading Ready:** Safely test and refine your trading strategies in a risk-free environment using Alpaca's paper trading API. Daily Performance Tracking:** Automatically logs your account equity and daily percentage change to a Google Sheet, giving you a clear view of your portfolio's performance. Detailed Trade Logging:** Every buy and sell order is meticulously recorded in a Google Sheet for easy review and historical analysis. Scheduled and Autonomous:** The entire process runs on a daily schedule, making it a "set and forget" solution for systematic trading. How It Works This workflow executes a sophisticated, automated trading strategy in a few key stages: Daily Kick-off & Snapshot: The workflow triggers on a daily schedule, first fetching your current Alpaca account balance and logging it to a Google Sheet to track daily performance. Strategy Formulation: It then reads the daily sentiment scores produced by the accompanying "Stock Sentiment Analysis" workflow. A Code node filters these results to identify the top four stocks with the highest positive sentiment. The Decision Engine: The core of the workflow is a custom Code node that acts as the trading brain. It: Retrieves your currently open positions from Alpaca. Compares your holdings against the day's top four sentiment stocks. Generates a "sell list" of positions you hold that are no longer in the top four. Generates a "buy list" of top-sentiment stocks that you don't yet own. Calculates the total cash value from the "sell list" and determines the exact notional value to invest in each stock on the "buy list." Trade Execution: The workflow first iterates through the "sell list" and executes a DELETE request to Alpaca for each, closing the positions. A Wait node pauses the workflow for two minutes to ensure the sell orders are filled and the account balance is updated. It then iterates through the "buy list," executing POST requests to Alpaca to purchase the new assets with the calculated funds. Record Keeping: All executed orders (both buys and sells) are merged and logged in a dedicated Google Sheet, giving you a permanent and detailed transaction history. Nodes Used Schedule Trigger HttpRequest (Alpaca API) Google Sheets Code (JavaScript) SplitOut Wait Merge This workflow is the perfect next step for anyone looking to take their AI analysis to the next level. Take the emotion out of your trading and let this bot systematically execute your data-driven strategy.
by Milan Vasarhelyi - SmoothWork
Video Introduction Want to automate your inbox or need a custom workflow? 📞 Book a Call | 💬 DM me on Linkedin Get Alerted When Your Website Is Down – n8n as an Uptime Robot Alternative If you manage a website (or client sites), getting notified when your site goes down is critical. But you’ve probably experienced alert fatigue if you’ve ever received a flood of downtime emails for every tiny outage. These short hiccups are common, and most monitoring tools don’t let you filter them out. Here’s how you can set up website downtime notifications with n8n - no extra subscriptions, and no more false positives spamming your inbox. For your most important sites, you can even get direct text alerts to your phone. We’ll use n8n, a powerful automation tool with thousands of templates and integrations. You can run it in the cloud or even self-host it for free. Why Use n8n for Website Monitoring? Uptime Robot and similar services have limitations or get expensive fast if you monitor several sites. n8n gives you full control - you choose when and how you get notified, set your own timing and thresholds, and you aren’t stuck with default alert logic. Plus, n8n can automate much more than just uptime monitoring: use it to handle other business workflows too. Quick Start: Free n8n Website Monitoring Template Use the template to get started fast. Log in or sign up for n8n Cloud, or self-host for free if you prefer. Configure your schedule (default: hourly) and list the sites you want to monitor. Key setting: wait seconds. Recommended: 300 seconds (5 minutes). If your site goes down, the workflow waits before sending an alert—no notification for a 1-2 minute outage, only real, persistent problems. How to Test & Use Activate the workflow. Toggle it on in n8n and it’ll check your sites automatically. Test instantly: Add a non-existent URL and run the workflow. After the wait time, you’ll get an email alert for that URL. Notifications stay organized: Alerts go straight to your inbox (see my AI email labelling template if you want color-coded organization). Get Critical Alerts on Your Phone (Telegram) Email is great, but for your most important sites you want instant notifications on your phone. Best option: Use Telegram. n8n can send you messages via a Telegram bot—easy to set up, free, and works in seconds. How: Create a Telegram bot via BotFather (instructions in the template). Add your bot token and chat ID to n8n. Now, you’ll get downtime alerts right on Telegram—no missed notifications, no extra costs. FAQ Can I monitor unlimited sites?** Yes, just list them all in the config. What if downtime is just a few seconds?** The default 5-minute wait filters these out. Do I need a paid n8n plan?** No - self-hosting is free, and the workflow works on free plans. Why not SMS or WhatsApp?** Telegram bots are fastest, easiest, and don’t require extra API setup or subscriptions.
by Jonathan | NEX
Are you drowning in a sea of security notifications? Do your analysts spend more time sifting through low-level logs than investigating real threats? This workflow transforms n8n into an autonomous SOC (Security Operations Center) Analyst, tackling alert fatigue head-on. Leveraging the NixGuard Security RAG connector, this workflow automates the entire alert triage process. It ingests raw security events (from sources like Wazuh, your SIEM, or EDR), uses AI to analyze and assign a priority, and then intelligently routes the alert to the correct Slack channel. How It Works: Ingest & Filter: The workflow runs on a schedule, fetching all recent security alerts. It first performs a basic filtering to isolate events that meet a minimum severity threshold (e.g., level 7+). AI Analysis & Prioritization: The aggregated high-severity alerts are then sent to the AI with a specific prompt, asking it to analyze the situation and return a structured JSON object containing a single, overall priority (Critical, High, Info) and a concise summary. Intelligent Routing: A Switch node reads the AI-assigned priority and routes the notification to the appropriate destination. Critical alerts go to your #security-incident-response channel, high-priority alerts to #security-investigations, and informational ones to #security-logs. Key Features & Benefits: Eliminate Alert Fatigue:** Drastically reduce the noise by having AI pre-process and categorize alerts before they hit your team. Automate SOC Tier 1 Triage:** Free up your human analysts from repetitive triage tasks so they can focus on high-value investigation and threat hunting. Faster Incident Response:** Route critical alerts to the right people in real-time, cutting down on crucial response time. Consistent Prioritization:** Use AI to ensure a consistent, unbiased approach to alert prioritization, 24/7. Smart Routing Logic:** Go beyond simple keyword matching. The Switch node ensures alerts are delivered to the team best equipped to handle them based on AI-assessed severity. Who is this for? SOC Analysts & Security Engineers** looking to automate alert triage and incident response workflows. SecOps and DevOps Teams** who want to build a more efficient, automated security operations pipeline. IT Managers and Directors** aiming to improve their team's efficiency and reduce the risk of missing critical alerts. Anyone using Wazuh, a SIEM, or other security tools that generate a high volume of alerts. Stop manually triaging alerts. Install this workflow to build your own AI-powered security automation platform and let your team focus on what matters most. Don't have the main workflow yet? Get it HERE! 🔗 Learn more about NixGuard: thenex.world 🔗 Get started with a free security subscription: thenex.world/security/subscribe Tags / Keywords: AI, Security, SOC, Automation, Triage, Alerting, Cybersecurity, Wazuh, SIEM, Slack, Incident Response, Alert Fatigue, SecOps, Generative AI, LLM, NixGuard, Routing
by Mario
Template to get your public IP address and push it to Namecheaps Dynamic DNS per subdomain. Open "yourdomain.com" Insert your domain and your Namecheap DDNS password Open "subdomains" Replaces and insert your subdomains Execute Workflow Have fun!
by Sarfaraz Muhammad Sajib
This n8n workflow sends SMS messages through the Textbelt API by accepting phone numbers, messages, and API keys as inputs. It uses a manual trigger to start the process, sets the necessary data, and executes an HTTP POST request to deliver the SMS. Step-by-Step Explanation: Manual Trigger: Starts the workflow manually by clicking ‘Execute workflow’. Set Data Node: Defines the required input parameters (phone, message, and key) that will be sent to the SMS API. You can populate these fields with your target phone number, the text message, and your Textbelt API key. HTTP Request Node: Sends a POST request to https://textbelt.com/tex with the phone number, message, and API key in the request body to send the SMS. The response from the API confirms whether the message was successfully sent.
by Evoort Solutions
✨ AI Text Summarizer with Google Sheets Logging Summarize large blocks of text into concise outputs using the Text Summarizer AI API and automatically log results in Google Sheets. This automation is ideal for content creators, marketers, researchers, and teams who need efficient summarization and record-keeping—without writing a single line of code. 🚀 Features | Feature | Description | |--------------------------|-----------------------------------------------------------------------------| | Web Form Trigger | Collects title, content, mode (Paragraph/Bullet), and length preferences. | | API Integration | Sends content to Text Summarizer AI API via RapidAPI. | | Conditional Logic | Routes success and error cases appropriately. | | Google Sheets Logging| Stores summaries or error messages into Google Sheets. | | Error Handling | Captures failed summaries and ensures no data is lost. | 🧠 Problem Solved Manually summarizing long-form content is time-consuming and repetitive. Storing this output in structured logs (like Google Sheets) adds an extra layer of manual effort. This workflow solves that by: 🔁 Automating AI-driven summarization using Text Summarizer AI API. 📊 Storing results (or fallback errors) directly into a Google Sheet. ✅ Ensuring no request goes unlogged—even on API failure. 🎯 Use Cases 📝 Blog Writers: Quickly convert articles into summaries for social captions. 📚 Students: Break down textbook or lecture content. 🧠 Knowledge Management: Turn raw meeting notes into concise summaries. 📢 SEO Teams: Use bullet-point output for schema markup or meta descriptions. 🧩 Nodes in the Flow | Node Name | Purpose | |----------------------|-------------------------------------------------------------------------| | On form submission | Captures user input via form (title, content, mode, length). | | Mapping | Formats input to match Text Summarizer AI API specs. | | HTTP Request | Sends POST request to the summarization API on RapidAPI. | | If | Validates whether a summary was returned. | | Wait | Adds short delay before writing to spreadsheet (success). | | Google Sheets | Appends summary data to Google Sheet. | | Wait1 | Adds delay for error handling path. | | Google Sheets1 | Logs failure with an “Error occurred” message. | ✅ Benefits 🧠 AI-Powered: Uses Text Summarizer AI API for fast and contextual summaries. 📁 Organized Logs: Google Sheets integration ensures easy tracking and auditing. 🔐 Reliable: Captures all submissions, including failed ones. 🧩 Customizable: Easily adapt inputs or connect to other tools like Notion, Slack, or Airtable. 🛠️ Requirements ✅ n8n account (Cloud or Self-Hosted) ✅ Access to Text Summarizer AI API ✅ Google account (for Sheets integration) 🔑 API key from RapidAPI (used in the HTTP Request node) 🔄 Workflow Overview User submits text and preferences (mode, length) via a form. Workflow triggers and transforms the input. Formatted data is sent to the Text Summarizer AI API. If a valid summary is returned: Log it into Google Sheets. If the API fails: Log the error message instead. 📥 Import Instructions Open n8n and import the workflow JSON. Replace the x-rapidapi-key in the HTTP node with your personal RapidAPI key. Configure your Google Sheets credentials. Deploy and test. ✅ 🔗 Suggested Extensions 🟢 Notion database logging 🟣 Slack/Discord notification for each summary 🔵 CSV download or Airtable sync 🏷 Tags ai summarization text-processing rapidapi google-sheets automation markdown n8n Text Summarizer AI API Create your free n8n account and set up the workflow in just a few minutes using the link below: 👉 Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!