by Agent Circle
This N8N template demonstrates how to use our tool to collect key information from any YouTube channel - including title, description, custom URL, stats (views, subscribers, videos), creation date, country, keywords, thumbnails, and channel ID. Everything is fetched and saved directly in a linked Google Sheet. Use cases are many: Whether you're part of an MCN or YouTube channel network, a digital marketing agency, an influencer analyst, or a growth-focused YouTube consultant, this tool helps you instantly access detailed channel data for scouting, benchmarking, and strategic planning. It's perfect for competitive research, brand vetting, or building your next performance dashboard. How It Works The workflow starts when you manually click Test Workflow or Execute Workflow in N8N. It reads through the list of full channel URLs or custom channel URLs from the Channel URLs tab in a connected Google Sheet. Only the URLs marked with the Ready status will be processed. The tool loops through each row and prepares the necessary data for the YouTube API call later. A Switch node detects the URL type of each input (whether it's a full or custom channel URL) and routes it accordingly. Based on the input type, the appropriate YouTube API endpoint is triggered to fetch channel metrics. It checks whether the API call is successful before continuing: If successful, the fetched data is written back to the connected Google Sheet in the same row and the row's status is updated to Finished. If it fails, the row's status is updated to Error. How To Use Download the workflow package. Import the workflow package into your N8N interface. Duplicate the YouTube - Get Channel Information Google Sheet template into your Google Sheets account. Set up Google Cloud Console credentials in the following nodes in N8N, ensuring enabled access and suitable rights to Google Sheets and YouTube services: For Google Sheets access, ensure each node is properly connected to the correct tab in your connected Google Sheet template: Node Google Sheets - Get Channel URLs → connected to Tab Channel URLs; Node Google Sheets - Update Data → connected to Tab Channel URLs; Node Google Sheets - Update Data - Error → connected to Tab Channel URLs. For YouTube access, set up a GET method to connect to YouTube API in the following nodes: Node HTTP Request - Get Comments For Video URLs 1; Node HTTP Request - Get Comments For Video URLs 2. In your connected Google Sheet, enter the full or custom channel URLs that you want to crawl and set the rows' status to Ready. Run the workflow by clicking Execute Workflow or Test Workflow in N8N. View the results in your Google Sheet: Successful fetches will update channel data in the same row in Channel URLs tab and the row's status is updated to Finished; otherwise, the row's status will be changed to Error. Requirements Basic setup in Google Cloud Console (OAuth or API Key method enabled) with enabled access to YouTube and Google Sheets. How To Customize By default, the workflow is manually triggered in N8N. However, you can automate the process by adding a Google Sheets trigger that monitors new entries in your YouTube – Get Channel Information template and starts the workflow automatically. Need Help? Join our community on different platforms for support, inspiration and tips from others. Website: https://www.agentcircle.ai/ Etsy: https://www.etsy.com/shop/AgentCircle Gumroad: http://agentcircle.gumroad.com/ Discord Global: https://discord.gg/d8SkCzKwnP FB Page Global: https://www.facebook.com/agentcircle/ FB Group Global: https://www.facebook.com/groups/aiagentcircle/ X: https://x.com/agent_circle YouTube: https://www.youtube.com/@agentcircle LinkedIn: https://www.linkedin.com/company/agentcircle
by Hostinger
This workflow automates the routine check for upgradable packages on your Ubuntu server, ensuring you stay updated with the latest software patches and security improvements. By running a daily script, it efficiently monitors any available package upgrades and promptly notifies you via email, saving you time and enhancing your server’s security. How It Works: Daily Monitoring**: The workflow is configured to execute a script daily that connects to your Ubuntu server and checks for any upgradable packages. Email Notification**: If any upgradable packages are detected during the check, the workflow triggers an alert mechanism that automatically sends you a notification email detailing the available updates. Set Up Steps: SSH Credentials**: Provide the SSH login credentials for your Ubuntu server. This will allow the workflow to securely connect and perform checks for software updates. SMTP Credentials**: Provide SMTP login details for your email account. These credentials are used to configure the email notifications system, enabling it to send alerts about the upgradable packages. Benefits: Timeliness**: Receive prompt updates on critical software upgrades to maintain the optimal performance and security of your server. Automation**: Reduces the need for manual checks, allowing you to focus on other critical tasks with peace of mind. Customizable**: Easily adjust the checking frequency or update the notification settings according to your preferences.
by Airtop
LinkedIn Post Engagement Data Extractor Use Case This automation is designed to extract key engagement metrics and audience data from a LinkedIn post. It's useful for analyzing the impact of content and identifying engaged users for lead generation, marketing, or research purposes. What It Does Given a LinkedIn post URL and an Airtop profile, this automation extracts: Total number of reactions Total number of comments Total number of reposts A list of users who reacted or commented, including: Their full name Their job title A link to their LinkedIn profile Input Parameters | Name | Description | Required | |------------------|------------------------------------------------------------|----------| | airtop_profile | The name of an Airtop Profile that's logged into LinkedIn | Yes | | linkedin_post_url| The full URL of the LinkedIn post you want to analyze | Yes | How It Works The workflow starts when triggered manually or from another workflow/form. It maps input fields for Airtop profile and post URL. Airtop opens a browser session and loads the LinkedIn post. An AI agent is instructed to extract engagement data via prompt-based analysis. The response is parsed and output in a structured format. Output Format The output will be a structured JSON object with the following fields: { "interactors": [ { "name": "Jane Doe", "job_title": "Marketing Director at ExampleCorp", "profile_url": "https://linkedin.com/in/janedoe" } // ... more interactors ], "reactions_count": 153, "comments_count": 21, "reposts_count": 8 } Read more about how to extract Linkedin post comments and reactions.
by LuisBetancourt.co
Description Whenever a Zoom “Meeting assets” email arrives in your Gmail inbox, this workflow will: 1) Trigger on new Gmail messages filtered by the subject “Meeting assets”. 2) Extract from the email (HTML or plain text): 3) Type of session (e.g. “1 hour”, “2 hours”, or “exploratory call”). Client’s full name. Session date & time (from the GMT… timestamp). Duration (HH:MM:SS). Recording link. Quick summary. Detailed summary. List of next steps. 4) Lookup the client in your Master Airtable base, table People, by full name. 5) Send a personalized Gmail to the client with all extracted details. 6) Create a new record in your Sessions table in Airtable, linking back to that client. Quick Start Import this JSON into n8n as a new workflow. Connect your Gmail credentials (OAuth2). Connect your Airtable credentials (Personal Access Token). In the Search Records node: Base → your Master base ID. Table → “Your people table”. Filter By Formula → ={Full Name} = '{{ $json.clientName }}'. In the Create Record node: Table → “Sessions”. Map each field (dateTime, duration, summaries, next steps, client link). Activate the workflow. Prerequisites n8n v1.50 or higher A Gmail account with OAuth2 credentials configured An Airtable base containing: Table People with a Full Name field (and email). Table Sessions with fields: DateTime, Duration, Quick Summary, Detailed Summary, Next Steps, and a Linked Record to People. An Airtable Personal Access Token with read/write access to that base. Tips & Extensions Timezone conversion: Use a Function node with moment-timezone to convert UTC if needed. Error handling: Add a catch node to log or notify if any field fails to parse. Alternate notifications: Swap the Gmail node for Slack, Microsoft Teams, or SMS integrations. With this documentation, your team can import and deploy the workflow in minutes. Enjoy!
by Łukasz
Who is it for? If you are having a lot of meetings as a project manager, CFO, CTO, CEO or any other role that requires handling many meetings, AND you are working with people in different timezones, you may have noticed that it is not uncommon that daylight savings time change day may differ from timezone to timezone. This may be very troublesome at times. If DST change day differs between timezones, then you might need to adjust your meetings time accordingly. And this happens twice a year. So it's good to get notification beforehand (at least a day before). This automation will notify you if tomorrow you can expect DST in any zone you provide. How It Works? Script runs daily and loops through provided timezones Checks if there is DST change to or from the tomorrow (if you want to be notified sooner, just adjust number of days) If there is DST change, script provides you with Slack notification (replace with email if needed) How to set up? Add and/or edit timezones you want to monitor in "Timezones List" node Adjust "Calculate Tomorrow's Date" if you want to be notified sooner than 1 day before DST change Adjust "Send Notification on Upcoming Change" to set where on Slack you want to be notified And that's it. Hope that you won't miss any other meeting because of DST!
by Niko
Capture URL Screenshots Automatically from Google Sheets & Drive with ScreenshotOne & Gmail Alerts Summary This automation template streamlines the process of capturing screenshots for multiple URLs. Instead of manually visiting each URL, taking a screenshot, and organizing the results, this workflow automates everything. When a spreadsheet is added to a designated Google Drive folder, the template extracts URLs from the column named "Url." These URLs are then processed through ScreenshotOne to capture screenshots, which are saved back to the same folder. Finally, an email notification is sent via Gmail with a link to the folder containing the screenshots. Problem Solved This template addresses the challenge of manual screenshot capture for multiple URLs. Without this automation, a user would need to: Open each URL from a spreadsheet. Take a screenshot manually. Save each screenshot with an appropriate name. Organize the screenshots in a folder. Notify stakeholders when the process is complete. These steps are not only time-consuming but also repetitive, especially when handling a large number of URLs. Who Can Benefit: Digital Marketers:** Monitor website appearances for competitive analysis or to track campaign landing pages. Web Developers/Designers:** Capture screenshots of multiple websites for inspiration or reference. QA Teams:** Document the visual state of web pages during various stages of development. SEO Specialists:** Track visual changes to websites they are optimizing. Content Managers:** Monitor how content appears across various web properties. Prerequisites Google Drive Node:** Must have appropriate permissions to create and access folders. Connected Google Sheets Node:** To extract URLs from the spreadsheet. Authenticated Gmail Node:** For sending notifications. ScreenshotOne Account:* Either a free or paid plan depending on volume needs, along with an *Access key**. Ensure you replace the placeholder --YOUR ACCESS KEY-- with your generated access key in the "Get Screenshots" node. Workflow Details Step 1: Google Drive Integration Trigger Node:** Monitors a specific folder in Google Drive. When a spreadsheet is added, the workflow is initiated. Step 2: Google Sheets Processing Google Sheets Node:** Extracts URLs from the column named "Url." Step 3: Screenshot Capture Get Screenshots Node:** Sends each extracted URL to ScreenshotOne to capture screenshots. Step 4: Saving Screenshots and Notifications Google Drive Node:** Saves the captured screenshots back into the same folder. Gmail Node:** Sends an email notification with a link to the folder, alerting stakeholders that the screenshots are ready. Customization Guidance Folder Monitoring: The workflow is set to monitor a specific Google Drive folder. It can be customized by selecting a different folder in the node settings. Spreadsheet Structure: While the template expects a spreadsheet with a column named "Url." for extracting URLs, users can add additional columns (e.g., titles, categories, or tags) and modify the workflow to utilize them as needed. Email Settings: Customize the recipient, subject, and body of the notification email to suit your needs. If required, enable optional notifications for different stakeholders. ScreenshotOne Access Key & Configurations: A valid ScreenshotOne Access key is required to capture screenshots. Users can further refine screenshot settings (e.g., viewport size, device emulation, or delay timing) by exploring the available options in the ScreenshotOne API documentation.
by PiAPI
Who is the template for? This workflow is specifically designed for content creators and social media professionals, enabling Instagram and X (Twitter) influencers to produce highly artistic visual posts, empowering marketing teams to quickly generate event promotional graphics, assisting blog authors in creating featured images and illustrations, and helping knowledge-based creators transform key insights into easily shareable card visuals. Set up Instructions Fill in your API key from PiAPI. Fill in Basic Params Node following the sticky note guidelines. Set up a design template in Canvas Switchboard. Make a simple template in Switchboard. Click Crul and get the API code to fill in JSON of Design in Canvas. Click Test Workflow and get a url result. Use Case Here we will provide some setting examples to help users find a proper way to use this workflow. User could change these settings based on specific purposes. Basic Params Setting: theme: Hope scenario: Don't know about the future, confused and feel lost with tech-development. style: Cinematic Grandeur, Sci-Tech Aesthetic, 3D style example: 1. March. Because of your faith, it will happen. 2. Something in me will save me. 3. To everyone carrying a heavy heart in silence. You are going to be okay. 4. Tomorrow will be better. image prompt: A cinematic sci-fi metropolis where Deep Neural Nets control a hyper-connected society. Holographic interfaces glow in the air as robotic agents move among humans, symbolizing Industry 4.0. The scene contrasts organic human emotion with cold machine precision, rendered in a hyper-realistic 3D style with futuristic lighting. Epic wide shots showcase the grandeur of this civilization’s industrial evolution. Output Image: More Example Results for Reference
by Davide
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This n8n workflow integrates the powerful Pipedream MCP server with AI capabilities to create a smart, extensible assistant that can interact with over 2,700 APIs and 10,000+ tools — all within a secure and modular structure. This setup seamlessly integrates Pipedream's MCP server with n8n, enabling your AI assistant to leverage thousands of APIs and tools securely. Benefits Massive Tool Access**: Instantly connect 2,700+ APIs using Pipedream MCP tools — from productivity apps to custom APIs — with zero-code integration. Dynamic AI Agent**: The use of a LangChain agent allows for flexible tool execution and contextual conversations, powered by GPT. Easy Customization**: Simply copy your MCP tool URL into the respective sseEndpoint field to extend the agent’s capabilities. Scalable and Modular**: Add or remove tools (like Slack, Notion, Stripe, etc.) without altering the core logic. Secure and Revocable**: Credentials and API access can be managed directly via Pipedream’s MCP dashboard. How It Works Chat Trigger: The workflow begins when a chat message is received via the When chat message received node, which acts as the entry point. AI Agent Processing: The message is passed to the AI Agent node, which orchestrates the interaction using the connected tools and memory. Language Model: The OpenAI Chat Model (GPT-4.1-mini) processes the user's input and generates responses or actions. Memory: The Simple Memory node retains context from the conversation to enable coherent multi-turn interactions. Tool Integration: The Calendly and Gmail nodes (connected via Pipedream's MCP server) allow the AI to perform actions like scheduling events or sending emails. These tools use SSE (Server-Sent Events) endpoints provided by Pipedream. Response: The AI Agent combines the model's output and tool responses to deliver a final reply to the user. Set Up Steps Sign Up for Pipedream: Create an account on and set up your MCP server. Configure MCP Tools: Connect your accounts (e.g., Calendly, Gmail) in Pipedream and obtain the SSE endpoints for each tool (e.g., https://mcp.pipedream.net/xxx/calendly_v2). Update n8n Nodes: Replace the placeholder SSE endpoints in the Calendly and Gmail nodes with your Pipedream MCP URLs. OpenAI Credentials: Ensure the OpenAI Chat Model node has valid API credentials (configured under "OpenAi account"). Activate Workflow: Enable the When chat message received node (currently disabled) and deploy the workflow. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Automate With Marc
🤖 Grok-4 Customer Support Agent with Document-Based Intelligence Template [RAG] This workflow creates a smart, AI-powered customer support agent using Grok-4 that can answer questions based on a preloaded Google Doc knowledge base. It listens for incoming customer queries via Telegram, then uses Grok-4’s language reasoning + memory features to generate helpful responses pulled directly from the doc. Watch the Step-by-Step Tutorial of this Workflow: https://www.youtube.com/watch?v=OXzsh-Ba-8Y&t=2s It’s perfect for solopreneurs, startups, or businesses that want to: Automate first-level support Build a Telegram-based knowledge agent Answer FAQs using internal docs (like manuals, policies, product details) 🔍 How It Works: Telegram Trigger – Listens for incoming messages from users Google Docs Tool – Retrieves a specified doc to serve as the knowledge base Grok-4 AI Agent – Uses xAI’s latest LLM with built-in memory and the document as a tool Memory Buffer – Keeps track of ongoing context in the conversation Telegram Reply – Sends the final response back to the customer 🧠 Tools & Integrations Used: xAI Grok-4 Model (via Langchain-compatible node) Google Docs Tool (as a reference knowledge base) Telegram Bot API (chat interface) n8n Agent Framework (for chaining memory, model, and tools) 💡 Use Cases: AI-powered FAQ assistant for your product Internal HR bot answering company policy questions Support assistant trained on onboarding documents or technical manuals Private support bot for VIP groups on Telegram
by Max Mitcham
An intelligent system that monitors social media conversations, identifies high-value engagement opportunities, and generates strategic comments to establish thought leadership while adding genuine value to discussions. Overview This automation workflow leverages Trigify's social listening platform to intelligently identify and respond to social media conversations. It combines AI-powered analysis with strategic comment generation to build authentic thought leadership presence across social platforms. 🔄 Workflow Process 1. Social Listening Webhook Real-time social media monitoring Integrated with Trigify.io social listening platform Monitors conversations across multiple social platforms Captures post content, author details, engagement metrics, and URLs Filters incoming posts by predefined keywords and topics Processes posts in real-time as they're discovered 2. Platform Validation Filter Platform-specific engagement optimization Checks post source (LinkedIn, Twitter, Reddit, etc.) Currently optimized for LinkedIn engagement Filters out non-relevant platforms Maintains platform-specific engagement strategies Routes posts based on platform requirements 3. Post Relevance Analyzer Agent AI-powered opportunity assessment Analyzes post content against expertise domains: Social Media Intelligence Competitive Analysis B2B Marketing Attribution Evaluates value-add potential and audience quality Scores engagement opportunity and confidence levels Identifies natural connection points to demonstrate authority Filters out low-quality or irrelevant conversations Returns structured analysis with TRUE/FALSE relevance decision 4. Engagement Decision Gate Quality control checkpoint Processes AI analysis results Only proceeds with TRUE relevance scores Prevents engagement on inappropriate content Maintains high-quality engagement standards Protects brand reputation through selective filtering 5. Strategic Comment Generator Agent Authentic thought leadership responses Generates comments under 30 words for maximum impact Focuses on tactical advice, strategic insights, or pattern recognition Avoids promotional language or forced statistics Incorporates domain expertise naturally Maintains conversational, helpful tone Uses experience-based insights over generic advice 6. Web Search Integration Enhanced context gathering Optional web search capability for additional context Provides current market insights when needed Supplements comment generation with real-time data Ensures comments are informed and relevant 7. Output Formatting Structured data preparation Compiles post URL, suggested comment, and post summary Formats data for Slack notification system Maintains context across workflow steps Prepares actionable engagement package 8. Slack Notification System Team collaboration and review Sends formatted notifications to #comment-strategy channel Includes post summary, suggested comment, and direct link Provides action buttons (View Post, Copy Comment, Skip) Enables team review before engagement Maintains engagement tracking and decision history 🛠️ Technology Stack n8n**: Workflow orchestration and webhook management Claude Sonnet 4**: Multi-agent AI analysis and content generation Trigify.io**: Social listening and post monitoring platform Slack API**: Team notifications and collaboration OpenAI API**: Optional web search for enhanced context Webhook Integration**: Real-time post processing ✨ Key Features Real-time social media monitoring via Trigify integration AI-powered relevance scoring and quality assessment Strategic comment generation focused on thought leadership Platform-specific engagement optimization (LinkedIn-focused) Team collaboration through Slack notifications Selective engagement to maintain high-quality interactions Expertise-based content analysis across multiple domains Anti-promotional safeguards for authentic engagement 🎯 Ideal Use Cases Perfect for professionals seeking to build authentic thought leadership: B2B Executives** building thought leadership presence Marketing Professionals** demonstrating industry expertise Sales Leaders** engaging prospects through valuable insights Consultants** establishing authority in their domains Business Development Teams** nurturing relationship building Companies** wanting systematic social media engagement Teams** requiring quality control over social interactions Professionals** seeking authentic network growth through value-add 📈 Business Impact Transform passive social listening into active thought leadership: Establishes thought leadership** through strategic engagement Builds authentic professional relationships** naturally Demonstrates expertise** without direct promotion Increases visibility** among target audience Creates networking opportunities** through valuable contributions Maintains consistent social media presence** systematically Scales personal engagement** while preserving authenticity This workflow ensures every engagement adds genuine value while naturally showcasing professional expertise, creating a sustainable approach to social media thought leadership.
by Oneclick AI Squad
This n8n workflow automatically tracks assignment deadlines and sends reminders to students and teachers. It checks for upcoming assignments daily, organizes the data, and sends email notifications to ensure deadlines are met. Good to Know Fully Automated**: Runs daily at 9 AM on weekdays to check assignments. Regular Updates**: Sends reminders for upcoming deadlines. Clear Notifications**: Emails a list of assignments to students and teachers. Error Handling**: Skips execution if no assignments are due. Scalable**: Works for multiple assignments and users. How It Works Reminder and Tracking Flow Set Schedule for Trigger: Starts the workflow daily at 9 AM on weekdays. Get Assignments: Retrieves assignment data from Notion database. IF Assignments Exist: Checks if there are any upcoming assignments. Split Items: Breaks down the assignment list for individual processing. Send Email Reminder: Emails reminders to students and teachers. No Assignments: Stops the workflow if no assignments are found. Example Database Columns Assignment ID**: Unique identifier for each assignment. Title**: Name of the assignment. Due Date**: Deadline for submission. Student ID**: Unique identifier for the student. Teacher ID**: Unique identifier for the teacher. Status**: Current status (e.g., Pending, Completed). How to Use Import Workflow: Add the workflow to n8n using the “Import Workflow” option. Set Up Notion: Configure n8n with Notion API credentials to fetch assignments. Configure Email: Add student and teacher email addresses and set up an email service (e.g., Gmail). Activate Workflow: Save and turn on the workflow in n8n. Check Logs: Verify reminders are sent and tracked. Requirements n8n Instance**: Self-hosted or cloud-based n8n setup. Notion Database**: API access with assignment data. Email Service**: SMTP setup (e.g., Gmail) for sending reminders. Admin Oversight**: Someone to monitor and adjust as needed. Customizing This Workflow Change Schedule**: Adjust the trigger to run at a different time or frequency. Add More Data**: Include additional fields like priority or notes. Custom Email**: Modify the email template for specific details.
by mariskarthick
Reduce human delays between malware detection and remediation in MSSP/SOC environments. This workflow automates full endpoint antivirus scanning immediately after high-severity endpoint infection wazuh alerts, closing the gap between alerting and action. Why Use This Workflow? Malware alerts are only effective if acted upon swiftly. Manual follow-ups are slow or often missed, letting threats persist. Automates detection, triage, scan initiation, and notification—all within one minute of alerting. Ensures consistent, auditable actions across endpoints running Linux or Windows. 🔑 Key Features Listens for high-severity Wazuh AV infection alerts (e.g., rule 52502). Uses GPT-4 for AI-powered alert summaries to speed triage and decision making. Extracts exact infected file paths using AI and regex for targeted scanning. Runs ClamAV/defender scans directly on endpoints via SSH with least-privilege credentials. Sends real-time scan results and remediation updates through Telegram, Slack, or email. Runs locally with limited permissions—no need for elevated Wazuh manager access. 🎯 Impact Eliminates manual lag—scans start automatically and immediately. Standardizes response playbooks for reliable, repeatable remediation. Reduces threat dwell time, minimizing risk exposure. Provides full event-to-remediation visibility via logs and notifications. 🚀 Get Started Configure Wazuh Manager to forward AV alerts to this n8n webhook. Import this workflow JSON into your n8n instance. Set up required credentials: OpenAI API, SSH access for ClamAV scanning, notification channels (Telegram/Slack/email). Activate the workflow and monitor alerts triggering automated scans and reports. 📂 Enjoy customizing Swap ClamAV with your preferred antivirus commands (e.g., Defender) as needed. Integrate with your existing communication or ticketing systems. Extend or adapt for multi-endpoint orchestration or other alert rules. Created by Mariskarthick M Senior Security Analyst | Detection Engineer | Threat Hunter | Open-Source Enthusiast