by Jimleuk
This n8n workflow assists property managers and surveyors by reducing the time and effort it takes to complete property inventory surveys. In such surveys, articles and goods within a property may need to be captured and reported as a matter of record. This can take a sizable amount of time if the property or number of items is big enough. Our solution is to delegate this task to a capable AI Agent who can identify and fill out the details of each item automatically. How it works An AirTable Base is used to capture just the image of an item within the property Our workflow monitoring this AirTable Base sends the photo to an AI image recognition model to describe the item for purpose of identification. Our AI agent uses this description and the help of Google's reverse image search in an attempt to find an online product page for the item. If found, the product page is scraped for the item's specifications which are then used to fill out the rest of the details of the item in our Airtable. Requirements Airtable for capturing photos and product information OpenAI account to for image recognition service and AI for agent SerpAPI account for google reverse image search. Firecrawl.dev account for webspacing. Customising this workflow Try building an internal inventory database to query and integrate into the workflow. This could save on costs by avoiding fetching new each time for common items.
by Angel Menendez
Enhance Security Operations with the Venafi Slack CertBot! Venafi Presentation - Watch Video Our Venafi Slack CertBot is strategically designed to facilitate immediate security operations directly from Slack. This tool allows end users to request Certificate Signing Requests that are automatically approved or passed to the Secops team for manual approval depending on the Virustotal analysis of the requested domain. Not only does this help centralize requests, but it helps an organization maintain the security certifications by allowing automated processes to log and analyze requests in real time. Workflow Highlights: Interactive Modals**: Utilizes Slack modals to gather user inputs for scan configurations and report generation, providing a user-friendly interface for complex operations. Dynamic Workflow Execution**: Integrates seamlessly with Venafi to execute CSR generation and if any issues are found, AI can generate a custom report that is then passed to a slack teams channel for manual approval with the press of a single button. Operational Flow: Parse Webhook Data**: Captures and parses incoming data from Slack to understand user commands accurately. Execute Actions**: Depending on the user's selection, the workflow triggers other actions within the flow like automatic Virustotal Scanning. Respond to Slack**: Ensures that every interaction is acknowledged, maintaining a smooth user experience by managing modal popups and sending appropriate responses. Setup Instructions: Verify that Slack and Qualys API integrations are correctly configured for seamless interaction. Customize the modal interfaces to align with your organization's operational protocols and security policies. Test the workflow to ensure that it responds accurately to Slack commands and that the integration with Qualys is functioning as expected. Need Assistance? Explore Venafi's Documentation or get help from the n8n Community for more detailed guidance on setup and customization. Deploy this bot within your Slack environment to significantly enhance the efficiency and responsiveness of your security operations, enabling proactive management of CSR's.
by Thibaud
How it works: Schedule Trigger** on a daily basis (configured at 7.30 am) Connects Google Contacts** to get personal information from them Field Checker** on birthday date & firstnames. And see if there is any celebration for today Send a Telegram Notification** and display the message on a Google Home Speaker via Home Assistant if any celebration has matched Set up steps: Download* the workflow and *import** it into your n8n instance Configure accounts** for Google Contacts, Telegram and Home Assistant And you will be good to go
by InfraNodus
The Ultimate Gmail Analysis and Visual Summarization Template This workflow showcases various useful Gmail search, filter, and AI categorization operations and generates a knowledge graph for your mail using the InfraNodus GraphRAG API, which you can use to reveal the main topics and blind spots in your correspondence. InfraNodus will then target those blind spots to generate interesting research questions for you and send the topical summary and insights via Telegram. You can also click the generated graph and explore the blind spots inside InfraNodus using the interactive visual interface: What is it useful for? Learn about advanced Gmail search, filtering, and AI categorization functions** that can be useful for your other workflows Analyze all your personal messages for the last week to get an overview of the main topics Analyze all your Sent messages to find recurrent topics and gaps and generate ideas based. on those gaps Generate ideas based on specific message filters (Personal, Promos, from a specific person, AI-defined criteria, e.g. urgency) Get an overview of an interaction with a specific person / company Get an overview of your notes Generate new ideas based on your correspondence on a certain topic (e.g. "business") Learn about various n8n nodes useful for email processing, filtering, and data conversion Never miss important topics, use AI filter to get notified of the urgent and important emails via Telegram How it works This template can be triggered in multiple ways: automatically in regular intervals (daily, weekly), manually in n8n, or via a private password-protected URL form where you can specify your search and filtering criteria When you start the workflow, you specify: your Gmail search filters (can be combined, e.g. after:2025/06/01 label:personal business to search for all emails received after 1 June 2025, filed in the Personal category containing the word "business". (optional, if empty, will retrieve all the emails or limited to the number you set in the Gmail node) Additional Gmail labels (e.g. SENT or CATEGORY_PERSONAL or your custom categories). Use the search filter for faster processing (e.g. prefer label:person to CATEGORY_PERSONAL, but labels can be useful for additional filtering for your search queries) (optional, if empty, will retrieve all the emails) AI filtering criteria** — set an additional classification criteria used to filter out the emails, e.g. "Only the urgent, personal emails" — in that case, AI classification node working with Google's Gemini AI will be activated and will only pass through the email based on the criteria you specify. Whether you want to build a text graph or a social graph — see the workflow for detailed explanation of each Use snippets of emails (default) or full text (for thorough analysis). We prefer snippets as it's faster and your graph context doesn't get biased towards longer emails this way. Once you set up your search parameters in Steps 1 and 2, the template will follow the following steps: Step 3 — retrieve Google emails that satisfy your filter criteria. Filter them by additional labels provided if applicable. Step 4 - if the user chooses to analyze full text, use additional Gmail node that retrieves the full text of the email message Step 5 — if AI filter rule is provided, use the AI Classifier node with Google Gemini Pro 2.5 model to classify the email based on the rule provided. Bypass if empty. Step 6 - format the text or the email snippets to add the sender meta-data and category and to prepare to submit to InfraNodus Step 7 - submit the data to the InfraNodus HTTP graphAndEntries endpoint and generate a knowledge graph Step 8 - access this graph via the graphAndAdvice endpoint) and generate a topical summary based on the GraphRAG representation and insight questions bridging the gaps identified. Send the results via a Telegram bot. We use Telegram, because it takes only 30 seconds to set up a bot with an API, unlike Discord or Slack, which is long and cumbersome to set up. You can also attach a Gmail send node and generate an email instead. How to use You need an InfraNodus GraphRAG API account and key to use this workflow. Create an InfraNodus account Get the API key at https://infranodus.com/api-access and create a Bearer authorization key for the InfraNodus HTTP nodes. Add this Authorization code in Steps 7 and 8 of the workflow. Come up with the name of the graph and change it in the HTTP InfraNodus nodes in the steps 7 and 8 and also in the Telegram nodes that send a link to the graph. For additional settings you can use in the HTTP InfraNodus nodes, see the InfraNodus access points page. Authorize your Gmail account for Steps 2 and 3 Gmail nodes. The easiest way to set it up is to open a free Google Console API account and to create an OAuth access point for n8n. You can then reuse it with other Google services like Google Sheets, Drive, etc. So it's a useful thing to have in general. Set up the Gemini AI API key using the instructions in the Step 5 Gemini AI node. Set up the Telegram node bot for the Step 8. It takes only 30 seconds: just go to @botfather and type in /newbot and you'll have an API key ready. To get the conversation ID, follow the n8n / Telegram instructions in the node itself. Once everything is ready, try to run the default automated workflow to test if everything works well, then use the Form for playing around with specific filters that you may find useful. Requirements An InfraNodus account and API key An Google Cloud API OAuth client and key for Gmail access A Gemini AI API key A Telegram bot API key FAQ 1. What's the best search query to use? I personally like starting with analyzing the messages Gmail tags as "personal" from the last week (using the after:2025/05/28 label:personal search query) using the social graph settings. It helps me see who I interacted with, what it was about, and gives me a good bird's eye view into my last week's interactions, helping me see if I didn't miss anything. I also find it useful to analyze the sent messages (using the after:2025/05/28 label:sent search filter or SENT category filter) as it helps me see what I was writing about recently and understand some recurrent topics and gaps in my interactions. Finally, I also like to analyze notes (label:notes) or specific correspondence (from:your_friend@gmail.com) to get an overview and find gaps in the conversations. 2. Why use InfraNodus and not an AI summarization module? You probably get a lot of spam, so your AI will get overwhelmed with the content that's not really useful. The InfraNodus graph helps you see the important patterns and discover what's missing by focusing on the gaps. You can use the interactive graph to quickly remove the stuff you don't need and to focus on the most relevant topics and conversations. Customizing this workflow You can connect a Gmail node instead of the Telegram one if you prefer to receive notifications directly by email. I don't like using Slack and Discord because their bots are too difficult to set up and take too long. Check out the complete setup guide for this workflow at https://support.noduslabs.com/hc/en-us/articles/20394884531996-Build-a-Knowledge-Graph-and-Extract-Insights-from-Gmail-Emails-with-n8n-and-InfraNodus with a video tutorial coming soon and the links to other n8n workflows. Check our other n8n workflows at https://n8n.io/creators/infranodus/ for useful content gap analysis, expert panel, and marketing, and research workflows that utilize GraphRAG for better AI generation. Finally, check out https://infranodus.com to learn more about our network analysis technology used to build knowledge graphs from text.
by Dr. Firas
Generate AI videos with Seedance & Blotato, upload to TikTok, YouTube & Instagram Who is this for? This template is ideal for creators, content marketers, social media managers, and AI enthusiasts who want to automate the production of short-form, visually captivating videos for platforms like TikTok, YouTube Shorts, and Instagram Reels — all without manual editing or publishing. What problem is this workflow solving? Creating engaging videos requires: Generating creative ideas Writing detailed scene prompts Producing realistic video clips and sound effects Editing and stitching the final video Publishing across multiple platforms This workflow automates the entire process, saving hours of manual work and ensuring consistent, AI-driven content output ready for social distribution. What this workflow does This end-to-end AI video automation workflow: Generates a creative idea using OpenAI and LangChain Creates detailed video prompts with Seedance AI Generates video clips via Wavespeed AI Generates sound effects with Fal AI Stitches the final video using Fal AI’s ffmpeg API Logs metadata and video links to Google Sheets Uploads the video to Blotato Auto-publishes to TikTok, YouTube, Instagram, and other platforms Setup Add your OpenAI API key in the LLM nodes Set up Seedance and Wavespeed AI credentials for video prompt and clip generation Add your Fal AI API key for sound and stitching steps Connect your Google Sheets account for tracking ideas and outputs Set your Blotato API key and fill in the platform account IDs in the Assign Social Media IDs node Adjust the Schedule Trigger to control when the automation runs How to customize this workflow to your needs Change the AI prompts** to target your niche (e.g., ASMR, product videos, humor) Add a Telegram or Slack step** for video preview before publishing Tweak scene structure** or video duration to match your style Disable platforms** you don’t want by turning off specific HTTP Request nodes Edit the sound generation prompts** for different moods or effects 📄 Documentation: Notion Guide Need help customizing? Contact me for consulting and support : Linkedin / Youtube
by Angel Menendez
Who’s it for This template is perfect for OMI pendant users or anyone with AI-generated memory transcripts who want to: Automatically create daily journals in Markdown Extract actionable tasks from conversations Store memories in Google Drive Sync action items to Google Tasks Great for creators, ADHD professionals, techies, or productivity hackers who want to build a second brain workflow with no manual data entry. What it does / How it works This workflow: Accepts POST data from the OMI AI pendant (via webhook) Extracts structured summaries, tasks, events, and raw transcript data Converts the transcript into Markdown using metadata like emoji, category, and overview Uses Google Gemini or an AI Agent to generate a high-quality journal entry Splits out action items and creates tasks in Google Tasks Uploads both the transcription and the final journal file into separate Google Drive folders for archival Deletes processed files if needed (cleanup path is included) How to set up Connect your OMI device to send daily summaries to the webhook endpoint Authenticate your Google Drive and Google Tasks accounts Replace any hardcoded values (like folder IDs or task list IDs) with your own Review the system prompt in the AI Agent node if you'd like to personalize your journal style ## Requirements OMI pendant or device that generates .md summaries via API or webhook Google Drive & Google Tasks credentials set up in n8n Optional: Google Gemini or OpenAI for natural language journal generation ## How to customize Change the output folder IDs for GDrive in the Upload Transcription and Upload Journal nodes. One folder is for long term storage and the other is short term, the contents of which are deleted every night to generate the journal entries. Ensure your workflow timezone is set correctly in the settings. Replace Google Tasks with another todo app (e.g. Notion, Todoist) using HTTP or native nodes Customize the AI prompt in the AI Agent or Gemini Chat node to reflect your tone (e.g., poetic, minimalist, spiritual) Modify the Markdown format in the Build Markdown Transcription node for your preferred structure
by Abrar Sami
Turn Reddit Questions into SEO Articles Automatically This workflow takes real user questions from Reddit and transforms them into fully structured blog posts — title, intro, steps, and conclusion — using AI. How it works Manually triggered when you want to run it Scrapes the latest posts from a specific subreddit (e.g. r/n8n) Filters only posts that are real questions (based on keywords like “how,” “what,” “why”) Logs relevant questions into a Google Sheet as raw input Enhances each question using AI (rephrases, creates a clean title and slug) Generates full-length blog content: ✏️ Intro paragraph ✅ Step-by-step guide 🧠 Clear conclusion Saves the final blog content to a second Google Sheet for publishing Set up steps You’ll need access to: Reddit API (OAuth) OpenAI API Google Sheets Takes around 15–20 minutes to connect all the credentials and tweak prompts Customize the subreddit or topic focus by changing the Reddit node config Perfect for content teams who want to scale content output using real community pain points — without ever starting from a blank page.
by Trung Tran
Multi-Agent Book Creation Workflow with AI Tool Node and GPT-4, DALL-E Who’s it for This workflow is designed for: Content creators** who want to generate books or structured documents automatically. Educators and trainers** who need quick course materials, eBooks, or study guides. Automation enthusiasts* exploring *multi-agent systems* using the newly released *AI Tool Node** in n8n. Developers* looking for a reference template to understand *orchestration of multiple AI agents** with structured output. How it works / What it does This template demonstrates a multi-agent orchestration system powered by AI Tool Nodes: Trigger: Workflow starts when a chat message is received. Book Brief Agent: Generates the initial book concept (title, subtitle, and outline). Book Writer Agent: Expands the outline into full content by collaborating with two sub-agents: Designer Agent → Provides layout/design suggestions. Content Writer Agent → Drafts and refines chapters. Generate Cover Image: AI generates a custom book cover image. Upload to AWS S3: Stores the cover image securely. Configure Metadata: Adds metadata for title, author, and description. Build Book HTML: Converts markdown-based content into HTML format. Upload to Google Drive: Saves the HTML content for processing. Convert to PDF: Transforms the book into a professional PDF. Archive to Google Drive: Final version is archived for safe storage. This workflow showcases multi-agent coordination, structured parsing, and seamless integration with cloud storage services. How to set up Import the workflow into n8n. Configure the following connections: OpenAI (for Book Brief, Book Writer, Designer, and Content Writer Agents). AWS S3 (for image storage). Google Drive (for document storage & archiving). Add your API keys and credentials in n8n credentials manager. Test the workflow by sending a sample chat message (e.g., “Write a book about AI in education”). Verify outputs in Google Drive (HTML + PDF) and AWS S3 (cover image). Requirements n8n* (latest version with *AI Tool Node** support). OpenAI API key** (to power multi-agent models). AWS account** (with S3 bucket for storing images). Google Drive integration** (for document storage and archiving). Basic familiarity with workflow setup in n8n. How to customize the workflow Switch Models**: Replace gpt-4.1-mini with other models (faster, cheaper, or more powerful). Add More Agents: Introduce agents for **editing, fact-checking, or translation. Change Output Format: Export to **EPUB, DOCX, or Markdown instead of PDF. Branding Options: Modify the **cover generation prompt to include company logos or specific style. Extend Storage: Add **Dropbox, OneDrive, or Notion integration for additional archiving. Trigger Alternatives: Replace chat trigger with **form submission, webhook, or schedule-based runs. ✅ This workflow acts as a free, plug-and-play template to showcase how multi-agents + AI Tool Node can work together to automate complex content creation pipelines.
by Jay Emp0
Overview Fetch Multiple Google Analytics GA4 metrics daily, post to Discord, update previous day’s entry as GA data finalizes over seven days. Benefits Automates daily traffic reporting Maintains single message per day, avoids channel clutter Provides near–real-time updates by editing prior messages Use Case Teams tracking website performance via Discord (or any chat tool) without manual copy–paste. Marketing managers, community moderators, growth hackers. If your manager asks you for daily marketing report every morning, you can now automate it Notes google analytics node in n8n does not provide real time data. The node updates previous values for the next 7 days discord node on n8n does not have features to update an exisiting message by message id. So we have used the discord api for this most businesses use multiple google analytics properties across their digital platforms Core Logic Schedule trigger fires once a day. Google Analytics node retrieves metrics for date ranges (past 7 days) Aggregate node collates all records. Discord node fetches the last 10 messages in the broadcast channel Code node maps existing Discord messages by to the google analytics data using the date fields For each GA record: If no message exists → send new POST to the discord channel If message exists and metrics changed, send an update patch to the existing discord message Batch loops + wait nodes prevent rate-limit. Setup Instructions Import workflow JSON into n8n. Follow the n8n guide to Create Google Analytics OAuth2 credential with access to all required GA accounts. Follow the n8n guide to Create Discord OAuth2 credential for “Get Messages” operations. Follow the Discord guide to Create HTTP Header Auth credential named “Discord-Bot” with header Key: Authorization Value: Bot <your-bot-token> In the two Set nodes in the beginning of the flow, assign discord_channel_id and google_analytics_id. Get your discord channel id by sending a text on your discord channel and then copy message link Paste the text below and you will see your message link in the form of https://discord.com/channels/server_id/channel_id/message_id , you will want to get the channel_id which is the number in the middle Find your google analytics id by going to google analytics dashboard, seeing the properties in the top right and copy paste that number to the flow Adjust schedule trigger times to your preferred report hour. Activate workflow. Customization Replace Discord HTTP Request nodes with Slack, ClickUp, WhatsApp, Telegram integrations by swapping POST/PATCH endpoints and authentication.
by Marcel Claus-Ahrens
Instructions This automation enables you to just upload any Image (via Form) of a Logo Sheet, containing multiple Images of Product Logos (most likely) which brings them in some context to one another. After submitting an AI-Agent eats that Logo Sheet, turning it into an List of "Productname" and "Attributes", also checks if Tools are kind of similar to another, given the Context of the Image. We utilize AI Vision capabilities for that. NOTE: It might not be able to extract all informations. For a "upload and forget it" Workflow it works for me. You can even run it multiple times, to be sure. But if you need to make sure it extracts everything you might need to think about an Multi-Agent Setup with Validation-Agent Steps. Once the Agent finishes the extraction, it will traditionally and deterministicly add those Attributes to Airtable (Creates those, if not already existing.) and also Upserts the Tool Informations. It uses MD5 Hashes for turning Product Names into.. something fancy really, you could also use it without that, but I wanted to have something that looks atleast like an ID. Setup Set Up the Airtable like shown below. Update and set Credentials for all Airtable Nodes. Check or Adjust the Prompt of the Agent matching your use-case. Activate the Workflow. Open the Form (default: https://your-n8n.io/form/logo-sheet-feeder) Enjoy growing your Airtable. Enjoy the workflow! ❤️ let the work flow — Workflow Automation & Development
by Madame AI
Synchronize WooCommerce Inventory & Create Products with Gemini AI & BrowserAct This sophisticated n8n template automates WooCommerce inventory management by scraping supplier data, updating existing products, and intelligently creating new ones with AI-formatted descriptions. This workflow is essential for e-commerce operators, dropshippers, and inventory managers who need to ensure their product pricing and stock levels are synchronized with multiple third-party suppliers, minimizing overselling and maximizing profit. Self-Hosted Only This Workflow uses a community contribution and is designed and tested for self-hosted n8n instances only. How it works The workflow is typically run by a Schedule Trigger (though a Manual Trigger is also shown) to check stock automatically. It reads a list of suppliers and their inventory page URLs from a central Google Sheet. The workflow loops through each supplier: A BrowserAct node scrapes the current stock and price data from the supplier's inventory page. A Code node parses this bulk data into individual product items. It then loops through each individual product found. The workflow checks WooCommerce to see if the product already exists based on its name. If the product exists: It proceeds to update the existing product's price and stock quantity. If the product DOES NOT exist: An If node checks if the missing product's category matches a predefined type (optional filtering). If it passes the filter, a second BrowserAct workflow scrapes detailed product attributes from a dedicated product page (e.g., DigiKey). An AI Agent (Gemini) transforms these attributes into a specific, styled HTML table for the product description. Finally, the product is created in WooCommerce with all scraped details and the AI-generated description. Error Handling:* Multiple *Slack** nodes are configured to alert your team immediately if any scraping task fails or if the product update/creation process encounters an issue. Note: This workflow does not support image uploads for new products. To enable this functionality, you must modify both the n8n and BrowserAct workflows. Requirements BrowserAct** API account for web scraping BrowserAct** n8n Community Node -> (n8n Nodes BrowserAct) BrowserAct* templates named *“WooCommerce Inventory & Stock Synchronization”* and *“WooCommerce Product Data Reconciliation”** Google Sheets** credentials for the supplier list WooCommerce** credentials for product management Google Gemini** account for the AI Agent Slack** credentials for error alerts Need Help? How to Find Your BrowseAct API Key & Workflow ID How to Connect n8n to Browseract How to Use & Customize BrowserAct Templates How to Use the BrowserAct N8N Community Node Workflow Guidance and Showcase STOP Overselling! Auto-Sync WooCommerce Inventory from ANY Supplier
by Anna Bui
Automatically monitor LinkedIn posts from your community members and create AI-powered content digests for efficient social media curation. This template is perfect for community managers, content creators, and social media teams who need to track LinkedIn activity from their network without spending hours manually checking profiles. It fetches recent posts, extracts key information, and creates digestible summaries using AI. Good to know API costs apply** - LinkedIn API calls ($0.01-0.05 per profile check) and OpenAI processing ($0.001-0.01 per post) Rate limiting included** - Built-in random delays prevent API throttling issues Flexible scheduling** - Easy to switch from daily schedule to webhook triggers for real-time processing Requires API setup** - Need RapidAPI access for LinkedIn data and OpenAI for content processing How it works Daily profile scanning** - Automatically checks each LinkedIn profile in your Airtable for posts from yesterday Smart data extraction** - Pulls post content, engagement metrics, author information, and timestamps AI-powered summarization** - Creates 30-character previews of posts for quick content scanning Duplicate prevention** - Checks existing records to avoid storing the same post multiple times Structured storage** - Saves all processed data to Airtable with clean formatting and metadata Batch processing** - Handles multiple profiles efficiently with proper error handling and delays How to use Set up Airtable base** - Create tables for LinkedIn profiles and processed posts using the provided structure Configure API credentials** - Add your RapidAPI LinkedIn access and OpenAI API key to n8n credentials Import LinkedIn profiles** - Add community members' LinkedIn URLs and URNs to your profiles table Test the workflow** - Run manually with a few profiles to ensure everything works correctly Activate schedule** - Enable daily automation or switch to webhook triggers for real-time processing Requirements Airtable account** - For storing profile lists and managing processed posts with proper field structure RapidAPI Professional Network Data API** - Access to LinkedIn post data (requires subscription) OpenAI API account** - For intelligent content summarization and preview generation LinkedIn profile URNs** - Properly formatted LinkedIn profile identifiers for API calls Customising this workflow Change monitoring frequency** - Switch from daily to hourly checks or use webhook triggers for real-time updates Expand data extraction** - Add company information, hashtag analysis, or engagement trending Integrate notification systems** - Add Slack, email, or Discord alerts for high-engagement posts Connect content tools** - Link to Buffer, Hootsuite, or other social media management platforms for direct publishing Add filtering logic** - Set up conditions to only process posts with minimum engagement thresholds Scale with multiple communities** - Duplicate workflow for different LinkedIn communities or industry segments