by David Olusola
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 📁 Google Drive MCP Workflow – AI-Powered File Management Automation 🚀 🧠 Overview A secure and intelligent n8n workflow that connects with Google Drive via MCP (Model Control Protocol). Ideal for AI agent tasks, compliance-driven storage, and document automation. 🌟 Key Features 🔒 Built-In Safety Backs up files before edits (timestamped) Supports rollback using file history Validates file size, type, and permissions 📁 Smart Organization Automatically converts file types (PDF, DOCX, etc.) Moves files to structured folders Auto-archives old files based on age or rules 🔄 MCP Integration Accepts standardized JSON via webhook Real-time execution for AI agents Fully customizable input (action, fileId, format, etc.) ✅ AI Callable MCP Actions These are the commands AI agents can perform via MCP: Download a file (with optional format conversion) Upload a new file to Google Drive Copy a file for backup Move a file to a specific folder Archive old or inactive files Organize documents into folders Convert files to a new format (PDF, DOCX, etc.) Retrieve and review file history for rollback 📝 Example Input { "action": "download", "fileId": "abc123", "folderPath": "/projects/clientA", "convertFormat": "pdf" } 🔐 Security & Performance OAuth2 secured access to Google Drive API No sensitive data stored in transit Real-time audit logs and alerts Batch-friendly with built-in rate limiting 📌 Ideal For Businesses automating file management AI Agents retrieving, sorting, converting, or archiving files Compliance teams needing file versioning and backups ⚙️ Requirements n8n + Google Drive API v3 MCP server + Webhook integration Google OAuth2 Credentials
by Raz Hadas
Description Transform your investment strategy with a fully automated, AI-driven trading bot. This workflow bridges the gap between AI-powered market insights and real-world trading by executing buy and sell orders directly through the Alpaca paper trading API. Designed to work in tandem with the Automated Stock Sentiment Analysis workflow, this solution takes the top-performing stocks based on daily news sentiment and automatically rebalances your portfolio. It's perfect for algorithmic traders, data-driven investors, and n8n enthusiasts who want to see their AI analysis translate into tangible actions, all while maintaining a comprehensive log of every transaction in Google Sheets. Key Features & Benefits Automated Trading Execution:** Automatically places buy and sell orders on the Alpaca paper trading platform without manual intervention. Sentiment-Driven Decisions:** Leverages the output from the sentiment analysis workflow to make informed decisions, selling positions with waning sentiment and buying into those with high positive sentiment. Dynamic Portfolio Rebalancing:** Intelligently calculates which positions to close and how to allocate the resulting funds into new, high-potential stocks. Paper Trading Ready:** Safely test and refine your trading strategies in a risk-free environment using Alpaca's paper trading API. Daily Performance Tracking:** Automatically logs your account equity and daily percentage change to a Google Sheet, giving you a clear view of your portfolio's performance. Detailed Trade Logging:** Every buy and sell order is meticulously recorded in a Google Sheet for easy review and historical analysis. Scheduled and Autonomous:** The entire process runs on a daily schedule, making it a "set and forget" solution for systematic trading. How It Works This workflow executes a sophisticated, automated trading strategy in a few key stages: Daily Kick-off & Snapshot: The workflow triggers on a daily schedule, first fetching your current Alpaca account balance and logging it to a Google Sheet to track daily performance. Strategy Formulation: It then reads the daily sentiment scores produced by the accompanying "Stock Sentiment Analysis" workflow. A Code node filters these results to identify the top four stocks with the highest positive sentiment. The Decision Engine: The core of the workflow is a custom Code node that acts as the trading brain. It: Retrieves your currently open positions from Alpaca. Compares your holdings against the day's top four sentiment stocks. Generates a "sell list" of positions you hold that are no longer in the top four. Generates a "buy list" of top-sentiment stocks that you don't yet own. Calculates the total cash value from the "sell list" and determines the exact notional value to invest in each stock on the "buy list." Trade Execution: The workflow first iterates through the "sell list" and executes a DELETE request to Alpaca for each, closing the positions. A Wait node pauses the workflow for two minutes to ensure the sell orders are filled and the account balance is updated. It then iterates through the "buy list," executing POST requests to Alpaca to purchase the new assets with the calculated funds. Record Keeping: All executed orders (both buys and sells) are merged and logged in a dedicated Google Sheet, giving you a permanent and detailed transaction history. Nodes Used Schedule Trigger HttpRequest (Alpaca API) Google Sheets Code (JavaScript) SplitOut Wait Merge This workflow is the perfect next step for anyone looking to take their AI analysis to the next level. Take the emotion out of your trading and let this bot systematically execute your data-driven strategy.
by Milan Vasarhelyi - SmoothWork
Video Introduction Want to automate your inbox or need a custom workflow? 📞 Book a Call | 💬 DM me on Linkedin Get Alerted When Your Website Is Down – n8n as an Uptime Robot Alternative If you manage a website (or client sites), getting notified when your site goes down is critical. But you’ve probably experienced alert fatigue if you’ve ever received a flood of downtime emails for every tiny outage. These short hiccups are common, and most monitoring tools don’t let you filter them out. Here’s how you can set up website downtime notifications with n8n - no extra subscriptions, and no more false positives spamming your inbox. For your most important sites, you can even get direct text alerts to your phone. We’ll use n8n, a powerful automation tool with thousands of templates and integrations. You can run it in the cloud or even self-host it for free. Why Use n8n for Website Monitoring? Uptime Robot and similar services have limitations or get expensive fast if you monitor several sites. n8n gives you full control - you choose when and how you get notified, set your own timing and thresholds, and you aren’t stuck with default alert logic. Plus, n8n can automate much more than just uptime monitoring: use it to handle other business workflows too. Quick Start: Free n8n Website Monitoring Template Use the template to get started fast. Log in or sign up for n8n Cloud, or self-host for free if you prefer. Configure your schedule (default: hourly) and list the sites you want to monitor. Key setting: wait seconds. Recommended: 300 seconds (5 minutes). If your site goes down, the workflow waits before sending an alert—no notification for a 1-2 minute outage, only real, persistent problems. How to Test & Use Activate the workflow. Toggle it on in n8n and it’ll check your sites automatically. Test instantly: Add a non-existent URL and run the workflow. After the wait time, you’ll get an email alert for that URL. Notifications stay organized: Alerts go straight to your inbox (see my AI email labelling template if you want color-coded organization). Get Critical Alerts on Your Phone (Telegram) Email is great, but for your most important sites you want instant notifications on your phone. Best option: Use Telegram. n8n can send you messages via a Telegram bot—easy to set up, free, and works in seconds. How: Create a Telegram bot via BotFather (instructions in the template). Add your bot token and chat ID to n8n. Now, you’ll get downtime alerts right on Telegram—no missed notifications, no extra costs. FAQ Can I monitor unlimited sites?** Yes, just list them all in the config. What if downtime is just a few seconds?** The default 5-minute wait filters these out. Do I need a paid n8n plan?** No - self-hosting is free, and the workflow works on free plans. Why not SMS or WhatsApp?** Telegram bots are fastest, easiest, and don’t require extra API setup or subscriptions.
by Jonathan | NEX
Are you drowning in a sea of security notifications? Do your analysts spend more time sifting through low-level logs than investigating real threats? This workflow transforms n8n into an autonomous SOC (Security Operations Center) Analyst, tackling alert fatigue head-on. Leveraging the NixGuard Security RAG connector, this workflow automates the entire alert triage process. It ingests raw security events (from sources like Wazuh, your SIEM, or EDR), uses AI to analyze and assign a priority, and then intelligently routes the alert to the correct Slack channel. How It Works: Ingest & Filter: The workflow runs on a schedule, fetching all recent security alerts. It first performs a basic filtering to isolate events that meet a minimum severity threshold (e.g., level 7+). AI Analysis & Prioritization: The aggregated high-severity alerts are then sent to the AI with a specific prompt, asking it to analyze the situation and return a structured JSON object containing a single, overall priority (Critical, High, Info) and a concise summary. Intelligent Routing: A Switch node reads the AI-assigned priority and routes the notification to the appropriate destination. Critical alerts go to your #security-incident-response channel, high-priority alerts to #security-investigations, and informational ones to #security-logs. Key Features & Benefits: Eliminate Alert Fatigue:** Drastically reduce the noise by having AI pre-process and categorize alerts before they hit your team. Automate SOC Tier 1 Triage:** Free up your human analysts from repetitive triage tasks so they can focus on high-value investigation and threat hunting. Faster Incident Response:** Route critical alerts to the right people in real-time, cutting down on crucial response time. Consistent Prioritization:** Use AI to ensure a consistent, unbiased approach to alert prioritization, 24/7. Smart Routing Logic:** Go beyond simple keyword matching. The Switch node ensures alerts are delivered to the team best equipped to handle them based on AI-assessed severity. Who is this for? SOC Analysts & Security Engineers** looking to automate alert triage and incident response workflows. SecOps and DevOps Teams** who want to build a more efficient, automated security operations pipeline. IT Managers and Directors** aiming to improve their team's efficiency and reduce the risk of missing critical alerts. Anyone using Wazuh, a SIEM, or other security tools that generate a high volume of alerts. Stop manually triaging alerts. Install this workflow to build your own AI-powered security automation platform and let your team focus on what matters most. Don't have the main workflow yet? Get it HERE! 🔗 Learn more about NixGuard: thenex.world 🔗 Get started with a free security subscription: thenex.world/security/subscribe Tags / Keywords: AI, Security, SOC, Automation, Triage, Alerting, Cybersecurity, Wazuh, SIEM, Slack, Incident Response, Alert Fatigue, SecOps, Generative AI, LLM, NixGuard, Routing
by Mario
Template to get your public IP address and push it to Namecheaps Dynamic DNS per subdomain. Open "yourdomain.com" Insert your domain and your Namecheap DDNS password Open "subdomains" Replaces and insert your subdomains Execute Workflow Have fun!
by Sarfaraz Muhammad Sajib
This n8n workflow sends SMS messages through the Textbelt API by accepting phone numbers, messages, and API keys as inputs. It uses a manual trigger to start the process, sets the necessary data, and executes an HTTP POST request to deliver the SMS. Step-by-Step Explanation: Manual Trigger: Starts the workflow manually by clicking ‘Execute workflow’. Set Data Node: Defines the required input parameters (phone, message, and key) that will be sent to the SMS API. You can populate these fields with your target phone number, the text message, and your Textbelt API key. HTTP Request Node: Sends a POST request to https://textbelt.com/tex with the phone number, message, and API key in the request body to send the SMS. The response from the API confirms whether the message was successfully sent.
by Evoort Solutions
✨ AI Text Summarizer with Google Sheets Logging Summarize large blocks of text into concise outputs using the Text Summarizer AI API and automatically log results in Google Sheets. This automation is ideal for content creators, marketers, researchers, and teams who need efficient summarization and record-keeping—without writing a single line of code. 🚀 Features | Feature | Description | |--------------------------|-----------------------------------------------------------------------------| | Web Form Trigger | Collects title, content, mode (Paragraph/Bullet), and length preferences. | | API Integration | Sends content to Text Summarizer AI API via RapidAPI. | | Conditional Logic | Routes success and error cases appropriately. | | Google Sheets Logging| Stores summaries or error messages into Google Sheets. | | Error Handling | Captures failed summaries and ensures no data is lost. | 🧠 Problem Solved Manually summarizing long-form content is time-consuming and repetitive. Storing this output in structured logs (like Google Sheets) adds an extra layer of manual effort. This workflow solves that by: 🔁 Automating AI-driven summarization using Text Summarizer AI API. 📊 Storing results (or fallback errors) directly into a Google Sheet. ✅ Ensuring no request goes unlogged—even on API failure. 🎯 Use Cases 📝 Blog Writers: Quickly convert articles into summaries for social captions. 📚 Students: Break down textbook or lecture content. 🧠 Knowledge Management: Turn raw meeting notes into concise summaries. 📢 SEO Teams: Use bullet-point output for schema markup or meta descriptions. 🧩 Nodes in the Flow | Node Name | Purpose | |----------------------|-------------------------------------------------------------------------| | On form submission | Captures user input via form (title, content, mode, length). | | Mapping | Formats input to match Text Summarizer AI API specs. | | HTTP Request | Sends POST request to the summarization API on RapidAPI. | | If | Validates whether a summary was returned. | | Wait | Adds short delay before writing to spreadsheet (success). | | Google Sheets | Appends summary data to Google Sheet. | | Wait1 | Adds delay for error handling path. | | Google Sheets1 | Logs failure with an “Error occurred” message. | ✅ Benefits 🧠 AI-Powered: Uses Text Summarizer AI API for fast and contextual summaries. 📁 Organized Logs: Google Sheets integration ensures easy tracking and auditing. 🔐 Reliable: Captures all submissions, including failed ones. 🧩 Customizable: Easily adapt inputs or connect to other tools like Notion, Slack, or Airtable. 🛠️ Requirements ✅ n8n account (Cloud or Self-Hosted) ✅ Access to Text Summarizer AI API ✅ Google account (for Sheets integration) 🔑 API key from RapidAPI (used in the HTTP Request node) 🔄 Workflow Overview User submits text and preferences (mode, length) via a form. Workflow triggers and transforms the input. Formatted data is sent to the Text Summarizer AI API. If a valid summary is returned: Log it into Google Sheets. If the API fails: Log the error message instead. 📥 Import Instructions Open n8n and import the workflow JSON. Replace the x-rapidapi-key in the HTTP node with your personal RapidAPI key. Configure your Google Sheets credentials. Deploy and test. ✅ 🔗 Suggested Extensions 🟢 Notion database logging 🟣 Slack/Discord notification for each summary 🔵 CSV download or Airtable sync 🏷 Tags ai summarization text-processing rapidapi google-sheets automation markdown n8n Text Summarizer AI API Create your free n8n account and set up the workflow in just a few minutes using the link below: 👉 Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!
by vinci-king-01
Daily Report Generator with Mattermost and HubSpot This workflow automatically compiles key metrics from HubSpot (and optional internal data sources) into a concise daily summary and posts it to a designated Mattermost channel. It helps sales and marketing teams stay informed without manually pulling reports or navigating multiple dashboards. Pre-conditions/Requirements Prerequisites An n8n instance (self-hosted or n8n Cloud) HubSpot account with a Private App token Mattermost workspace with an incoming webhook enabled (Optional) Internal REST API or database endpoint for additional data Required Credentials HubSpot Private App Token** – Grants API access to Deals, Contacts, Activities, etc. Mattermost Personal Access Token** (or Incoming Webhook URL) – Permits message posting to channels n8n User Management** – Ensure the workflow has network access to HubSpot and Mattermost Specific Setup Requirements | Component | Requirement | Example/Notes | |-------------|---------------------------------------------------------|-------------------------------------------------------| | Mattermost | Create an Incoming Webhook or generate a PAT | System Console → Integrations | | HubSpot | Create a Private App → Scopes: crm.objects.deals.read | Settings → Integrations → Private Apps | | Scheduler | External cron job or n8n’s internal trigger* | curl https://n8n.yourdomain.com/webhook/daily_report | *The provided template uses a Webhook node so you can trigger it via any scheduler (e.g., crontab, Zapier, GitHub Actions). Replace with the Cron node if preferred. How it works This workflow automatically compiles key metrics from HubSpot (and optional internal data sources) into a concise daily summary and posts it to a designated Mattermost channel. It helps sales and marketing teams stay informed without manually pulling reports or navigating multiple dashboards. Key Steps: Webhook Trigger**: Waits for a daily call from an external scheduler. HubSpot Node**: Retrieves deal statistics, new contacts, and other CRM metrics. HTTP Request Node**: (Optional) Pulls supplementary data from an internal API. Merge Node**: Consolidates HubSpot and optional data sources. Code / Set Nodes**: Formats numbers, calculates KPIs, and builds a markdown message. If Node**: Guards against empty datasets or API failures. Mattermost Node**: Posts the formatted report to the chosen channel. Respond to Webhook**: Returns a JSON confirmation to the scheduler. Set up steps Setup Time: 15-20 minutes Import Template: In n8n, go to “Workflows → Import from File” and select the JSON template. Add Credentials: a. HubSpot → New credential → Paste Private App token b. Mattermost → New credential → Paste PAT or Webhook URL Configure Webhook URL: Copy the production URL of the Webhook node and add it to your external scheduler (e.g., crontab, Zapier). Adjust Query Parameters (HubSpot node): Modify filters (e.g., deal stage, create date) as needed. Edit Message Template (Code node): Update markdown formatting, include/exclude sections. Test Run: Manually execute the workflow. Verify the JSON response and Mattermost post. Activate: Toggle workflow to “Active”. Confirm your scheduler triggers it at the desired time. Node Descriptions Core Workflow Nodes: stickyNote** – Contains inline documentation and instructions. Webhook** – Primary trigger, receives the daily HTTP call. HubSpot** – Pulls deals, contacts, and engagement data. HTTP Request** – Fetches optional internal statistics (e.g., support tickets). Merge** – Combines HubSpot and HTTP results into one object. Set** – Selects and renames fields for clarity. Code** – Calculates KPIs (e.g., conversion rate) and assembles a markdown summary. If** – Checks for empty data arrays or API errors. Mattermost** – Sends the final message to a channel. Respond to Webhook** – Returns a success/failure payload to the caller. Data Flow: Webhook → HubSpot Webhook → HTTP Request HubSpot + HTTP Request → Merge → Set → Code → If → Mattermost → Respond to Webhook Customization Examples Change Report Time Range // HubSpot Node → Additional Fields { "filterGroups": [{ "filters": [{ "propertyName": "createdate", "operator": "BETWEEN", "highValue": Date.now(), "value": Date.now() - 24 * 60 * 60 * 1000 // last 24h }] }] } Format Mattermost Message with Emojis // Code Node (return statement) return [{ json: { text: :bar_chart: Daily CRM Report\n\n• New Deals: ${newDeals}\n• New Contacts: ${newContacts}\n• Win Rate: ${winRate}% } }]; Data Output Format The workflow outputs structured JSON data: { "status": "success", "date": "2024-05-23", "hubspot": { "new_deals": 12, "new_contacts": 34, "win_rate": 27.1 }, "internal": { "tickets_opened": 8, "tickets_closed": 6 }, "mattermostPostId": "abc123xyz" } Troubleshooting Common Issues 401 Unauthorized (HubSpot) – Verify Private App token and scopes. Regenerate if necessary. Message Not Posting – Ensure the Mattermost token has post:write or the webhook URL is valid. Performance Tips Cache HubSpot responses during testing to avoid hitting API limits. Reduce payload size by selecting only the fields you need in the Set node. Pro Tips: Replace the Webhook node with the Cron node for an all-in-n8n schedule. Use environment variables for tokens ({{$env.HUBSPOT_TOKEN}}) to avoid hard-coding secrets. Add a second Mattermost node to DM managers for critical alerts (e.g., low win rate). This is a community workflow template provided “as-is.” It is not officially supported by n8n GmbH. Always review and test in a development environment before deploying to production.
by Dr. Firas
💥 Generate AI Music & Publish to YouTube Automatically with Blotato This workflow automatically generates AI music, creates a thumbnail, converts everything into a video, and publishes it directly to YouTube — fully automated. Perfect for building faceless YouTube music channels, AI content automation, or passive content generation pipelines. Who is this for? This template is ideal for: YouTube creators building AI music channels Content creators creating faceless videos Automation enthusiasts using n8n for content pipelines Agencies generating bulk YouTube content Anyone wanting to automate music creation and publishing Whether you're creating lofi music, ambient tracks, meditation sounds, or background music, this workflow removes manual work entirely. What problem is this workflow solving? Creating AI music videos typically involves multiple manual steps: Generating music Creating thumbnails Converting audio into video Uploading to YouTube Writing titles and metadata This process is time-consuming and repetitive. This workflow automates everything into a single end-to-end pipeline, allowing you to: Generate unlimited AI music Automatically create thumbnails Convert audio into video Publish directly to YouTube What this workflow does This workflow performs the following steps: User submits a music request via form AI generates an optimized music prompt ElevenLabs generates AI music AI generates thumbnail prompt Atlas Cloud generates thumbnail image Cloudinary uploads generated assets Shotstack creates video (image + audio) Blotato uploads video to YouTube automatically SEO-optimized YouTube title is generated automatically Final result: AI-generated music video automatically published to YouTube The entire process runs fully automated. Setup To use this workflow, configure the following credentials: Required APIs OpenAI API Key — Prompt generation Elevenlabs** API Key — Music generation AtlasCloud** API Key — Thumbnail generation Cloudinary Account — Media hosting Shotstack API Key — Video creation Blotato** API Key — YouTube publishing Community Node Requirement This workflow uses the Blotato community node. 🎥 Watch This Tutorial 👋 Need help or want to customize this? 📩 Contact: LinkedIn 📺 YouTube: @DRFIRASS 🚀 Workshops: Mes Ateliers n8n Need help customizing? Contact me for consulting and support : Linkedin / Youtube / 🚀 Mes Ateliers n8n
by Dataki
This workflow helps you generate an llms.txt file (if you're unfamiliar with it, check out this article) using a Screaming Frog export. Screaming Frog is a well-known website crawler. You can easily crawl a website. Then, export the "internal_html" section in CSV format. How It Works: A form allows you to enter: The name of the website A short description The internal_html.csv file from your Screaming Frog export Once the form is submitted, the workflow is triggered automatically, and you can download the llms.txt file directly from n8n. Downloading the File Since the last node in this workflow is "Convert to File", you will need to download the file directly from the n8n UI. However, you can easily add a node (e.g., Google Drive, OneDrive) to automatically upload the file wherever you want. AI-Powered Filtering (Optional): This workflow includes a text classifier node, which is deactivated by default. You can activate it to apply a more intelligent filter to select URLs for the llms.txt file. Consider modifying the description in the classifier node to specify the type of URLs you want to include. How to Use This Workflow Crawl the website you want to generate an llms.txt file for using Screaming Frog. Export the "internal_html" section in CSV format. In n8n, click "Test Workflow", fill in the form, and upload the internal_html.csv file. Once the workflow is complete, go to the "Export to File" node and download the output. That's it! You now have your llms.txt file! Recommended Usage: Use this workflow directly in the n8n UI by clicking 'Test Workflow' and uploading the file in the form.
by GiovanniSegar
How to use the workflow This workflow takes a Loom link, extracts the video ID, uses the Loom API to download the video, then sends it to Gemini along with your question. Finally, it sends the output to Slack. To use it, you just need to add your own API key for Gemini and Slack connection. Click the link above to get your Gemini API key, then add a generic "Query auth" type credential in n8n. The name will be "key" and the value will be your API key. One way to customize this workflow would be to make the trigger any received email, extract the Loom link, and run an auto-prompt like "Describe this video in detail".
by Hans Blaauw
This flow is supported by a Chrome plugin created with Cursor AI. The idea was to create a Chrome plugin and a backend service in N8N to do chart analytics with OpenAI. It's a good sample on how to submit a screenshot from the browser to N8N. Who is it for? N8N developers who want to learn about using a Chrome plugin, an N8N webhook and OpenAI. What opportunity does it present? This sample opens up a whole range of N8N connected Chrome extensions that can analyze screenshots by using OpenAI. What this workflow does? The workflow contains: a webhook trigger an OpenAI node with GPT-4O-MINI and Analyze Image selected a response node to send back the Text that was created after analysing the screenshot. All this is needed to talk to the Chrome extension which is created with Cursor AI. The idea is to visit the tradingview.com crypto charts, click the Chrome plugin and get back analytics about the shown chart in understandable language. This is driven by the N8N flow. With the new image analytics capabilities of OpenAI this opens up a world of opportunities. Requirements/setup OpenAI API key Cursor AI installed The Chrome extension. Download The N8N JSON code. Download How to customize it to your needs? Both the Chrome extension and N8N flow can be adapted to use on other websites. You can consider: analyzing a financial screen and ask questions about the data shown analyzing other charts extending the N8N workflow with other AI nodes With AI and image analytics the sky is the limit and in some cases it saves you from creating complex API integrations. Download Chrome extension