by Simon
This n8n workflow simplifies the process of removing backgrounds from images stored in Google Drive. By leveraging the PhotoRoom API, this template enables automatic background removal, padding adjustments, and output formatting, all while storing the updated images back in a designated Google Drive folder. This workflow is very useful for companies or individuals that are spending a lot of time into removing the background from product images. How it Works The workflow begins with a Google Drive Trigger node that monitors a specific folder for new image uploads. Upon detecting a new image, the workflow downloads the file and extracts essential metadata, such as the file size. Configurations are set for background color, padding, output size, and more, which are all customizable to match specific requirements. The PhotoRoom API is called to process the image by removing its background and adding padding based on the settings. The processed image is saved back to Google Drive in the specified output folder with an updated name indicating the background has been removed. Requirements PhotoRoom API Key Google Drive API Access Customizing the Workflow Easily adjust the background color, padding, and output size using the configuration node. Modify the output folder path in Google Drive or replace Google Drive with another storage service if needed. For advanced use cases, integrate further image processing steps, such as adding captions or analyzing content using AI.
by tanaypant
This workflow is the third of three. You can find the other workflkows here: Incident Response Workflow - Part 1 Incident Response Workflow - Part 2 Incident Response Workflow - Part 3 We have the following nodes in the workflow: Webhook node: This trigger node listens to the event when the Resolve button is clicked. PagerDuty node: This node changes the status of the incident report from Acknowledged to Resolved in PagerDuty. Jira Software node: This node moves the incident issue to Done. Mattermost node: This node publishes a message in the auxiliary channel mentioning that the incident has been marked as resolved in PagerDuty and Jira. Mattermost node: This node publishes a message in the specified Incidents channel that the incident has been resolved by the on-call team.
by Angel Menendez
CallForge - AI Sales Call Processing & Insights Extraction Automate sales call analysis with AI-powered insights for sales, marketing, and product teams. Who is This For? This workflow is designed for: β Sales teams looking to extract structured insights from Gong call transcripts. β Marketing professionals seeking AI-driven customer pain points & content strategy. β Product teams needing feedback from sales calls to prioritize feature development. π What Problem Does This Workflow Solve? Manually analyzing Gong.io sales call transcripts is slow, inconsistent, and lacks structured insights. With CallForge, you can: β Extract AI-powered insights about use cases, objections, competitors, and next steps. β Provide structured marketing & product intelligence to enhance strategy. β Automatically store call insights in Notion and Salesforce for easy access. β Ensure resilience with automated reruns on failed workflows (handling Notion API limits). β Improve decision-making with AI-powered competitor and sentiment analysis. π Key Workflow Features π€ AI-Powered Transcript Analysis Uses AI to identify use cases, objections, competitors, and customer pain points. Categorizes insights for sales, marketing, and product teams. π AI Agent Breakdown πΉ Sales AI Agent β Extracts customer objections, pain points, competitors, and next steps. πΉ Marketing AI Agent β Identifies recurring topics, keyword trends, and content opportunities. πΉ Product AI Agent β Captures feature requests and AI/ML-related references. π Structured Output Processing Sales Data Processor* β Stores insights in *Notion & Salesforce** for sales tracking. Marketing Data Processor* β Extracts *SEO & content strategy insights** for marketing teams. Product AI Data Processor* β Logs *customer feedback* to prioritize *feature development**. π‘ Competitor & Integration Analysis Tracks competing products mentioned in calls**. Identifies integration needs**, flagging workarounds used by prospects. π’ Real-Time Slack Notifications Alerts teams on workflow progress** and completed call analyses. π Failure Resilience & Automated Re-Runs If a Notion API limit is reached, the process resumes automatically. π How This Works π 1. Trigger & Call Data Processing The workflow retrieves Gong call transcripts and metadata. Normalizes data**, correcting common mispronunciations like "n8n." π€ 2. AI Agents Analyze the Call Sales Agent** β Extracts actionable insights for sales follow-ups. Marketing Agent* β Identifies *recurring themes* and *keyword trends**. Product Agent* β Captures *feature requests and AI/ML usage mentions**. π‘ 3. Data is Stored in Notion & Salesforce Logs AI-extracted insights* in *Notion** for structured tracking. Pushes sales-related data* to *Salesforce** for team accessibility. π 4. Slack Alerts for Teams Notifies sales, marketing, and product teams** about extracted insights. CallForge - 01 - Filter Gong Calls Synced to Salesforce by Opportunity Stage CallForge - 02 - Prep Gong Calls with Sheets & Notion for AI Summarization CallForge - 03 - Gong Transcript Processor and Salesforce Enricher CallForge - 04 - AI Workflow for Gong.io Sales Calls CallForge - 05 - Gong.io Call Analysis with Azure AI & CRM Sync CallForge - 06 - Automate Sales Insights with Gong.io, Notion & AI CallForge - 07 - AI Marketing Data Processing with Gong & Notion CallForge - 08 - AI Product Insights from Sales Calls with Notion π Sample Output Data 1οΈβ£ Sales Insights { "UseCases": [ { "Summary": "A manufacturing company wants to automate inventory tracking and reduce manual entry delays.", "DepartmentTags": ["Operations"], "IndustryTags": ["Manufacturing"], "ImplementationStatus": "Evaluating" } ], "Objection": { "ObjectionTags": ["Feature Limitation"], "Nature": "The prospect wanted a deeper integration with their ERP system, which n8n currently lacks." }, "CallSummary": "The call focused on automation for supply chain processes. The prospect expressed interest but wanted confirmation on ERP integration capabilities.", "NextSteps": ["Schedule a follow-up demo for ERP integration."] } 2οΈβ£ Marketing Insights { "MarketingInsights": [ { "Tag": "Workflow Template Request", "Summary": "The prospect requested a template for automating CRM lead tracking." } ], "RecurringTopics": [ { "Topic": "CRM Integration", "Mentions": 3, "Context": "Discussed how n8n could sync CRM data automatically." } ], "ActionableInsights": [ { "RecommendationType": "Tutorial", "Title": "How to Automate CRM Lead Tracking with n8n", "Topic": "CRM Integration", "Rationale": "The prospect expressed a need for CRM automation templates." } ] } 3οΈβ£ Product Feedback { "ProductFeedback": [ { "Sentiment": "Positive", "Feedback": "The external speaker praised the simplicity of n8n's UI, making it easier for non-developers to automate tasks." }, { "Sentiment": "Negative", "Feedback": "The external speaker mentioned frustration over the lack of a dedicated ERP integration node." } ], "AI_ML_References": { "Exist": true, "Context": "The external speaker mentioned using AI for automating customer ticket categorization.", "Details": { "DevelopmentStatus": "Building", "Department": "Support", "RequiresAgents": true, "RequiresRAG": false, "RequiresChat": "Yes: External App (e.g., Slack)" } } } π§ How to Customize This Workflow π‘ π Change Data Storage β Swap Notion for Airtable, HubSpot, or another CRM. π‘ π© Customize Slack Notifications β Send alerts via email, webhook, or another channel. π‘ π Modify AI Processing β Adjust AI models or processing prompts. π‘ π Add More Integrations β Sync insights with Pipedrive, HubSpot, or another CRM. π Why Use This Workflow? β Automates Gong call transcript analysis, eliminating manual work. β Improves collaboration by structuring insights for sales, marketing, and product teams. β Boosts sales conversions by identifying objections and next steps. β Enhances marketing and SEO strategy with AI-driven insights. β Optimizes product roadmap decisions based on customer feedback. This workflow scales AI-powered sales intelligence for better decision-making, content strategy, and sales enablement. π
by Nskha
Overview The [n8n] YouTube Channel Advanced RSS Feeds Generator workflow facilitates the generation of various RSS feed formats for YouTube channels without requiring API access or administrative permissions. It utilizes third-party services to extract data, making it extremely user-friendly and accessible. Key Use Cases and Benefits Content Aggregation**: Easily gather and syndicate content from any public YouTube channel. No API Key Required**: Avoid the complexities and limitations of Google's API. Multiple Formats**: Supports ATOM, JSON, MRSS, Plaintext, Sfeed, and direct YouTube XML feeds. Flexibility**: Input can be a YouTube channel or video URL, ID, or username. Services/APIs Utilized This workflow integrates with: commentpicker.com**: For retrieving YouTube channel IDs. rss-bridge.org**: To generate various RSS formats. Configuration Instructions Start the Workflow: Activate the workflow in your n8n instance. Input Details: Enter the YouTube channel or video URL, ID, or username via the provided form trigger. Run the Workflow: Execute the workflow to receive links to 13 different RSS feeds, including community and video content feeds. Screenshots Additional Notes Customization**: You can modify the RSS feed formats or integrate additional services as needed. Support and Contributions For support, questions, or contributions, please visit the n8n community forum or the GitHub repository. We welcome contributions from the community!
by Eduard
Create, iterate, and share! Transform a single image through multiple scenes while maintaining consistency. β¨ What this workflow does This template showcases FLUX.1 Kontext - Black Forest Labs' in-context image generation model that excels at maintaining character features across multiple transformations. Combined with the Upload Post community node for effortless multi-platform social media posting, you can create and share compelling visual stories instantly. The workflow demonstrates FLUX Kontext's core strength: character consistency across multiple image generations. Starting with a single input image, it: πΌοΈ Loads an initial character image (example: a cute animal mascot) π Defines multiple scene transformation prompts π Iteratively generates new scenes while preserving exact character features π― Maintains visual consistency by reusing binary data from previous generations π± Auto-posts the complete transformation series to multiple social platforms simultaneously π Key Features: The Consistency Advantage Character Preservation**: FLUX Kontext's signature feature - maintains character features and style across transformations (requires specific prompting techniques) Iterative Context Building**: Each generation uses the previous image as context, creating visual continuity Binary Data Reuse**: Smart workflow design that feeds output from one generation as input to the next Multi-Scene Storytelling**: Transform your character across different environments while keeping them recognizable One-Click Multi-Platform Posting*: Upload Post eliminates the tedious process of posting to each platform individually π± Why use Upload Post? Posting the same content to TikTok, Instagram, LinkedIn, YouTube, Facebook, X (Twitter), and Threads individually is time-consuming and error-prone. The Upload Post service* simplifies this process: β Connect once, post everywhere: Link all your social media accounts to Upload Post β Single API call: Post to multiple platforms with one simple node β No more platform juggling: Skip the endless switching between apps and dashboards β Consistent timing: All platforms get your content simultaneously β Trusted by 3,751+ users: Proven solution for content creators and marketers Instead of spending 30+ minutes manually posting to each platform, Upload Post does it all in seconds with a single n8n node! π οΈ Prerequisites Required Accounts: Black Forest Labs API: Create account at dashboard.bfl.ai Get your API key for FLUX Kontext Pro access Upload Post Account: Sign up at upload-post.com* Connect your social media profiles (TikTok, Instagram, LinkedIn, YouTube, Facebook, X/Twitter, Threads) Get API credentials for automated posting Free tier available: 10 uploads/month π‘ Perfect For: Character Designers** maintaining brand character integrity across scenes Social Media Managers** creating engaging visual story series without manual posting Brand Marketers** ensuring character consistency across campaigns Storytellers** building visual narratives with consistent protagonists Agencies** managing multiple client accounts efficiently π§ Customization Options: Modify transformation prompts** to create your own character journey Adjust iteration steps** Change initial character image** Configure social platform targeting** (choose which platforms to post to) Customize post content** and formatting Experiment with different consistency scenarios** \ Affiliate link*
by Harshil Agrawal
Based on your use case, you might want to trigger a workflow if new data gets added to your database. This workflow allows you to send a message to Mattermost when new data gets added in Google Sheets. The Interval node triggers the workflow every 45 minutes. You can modify the timing based on your use case. You can even use the Cron node to trigger the workflow. If you wish to fetch new Tweets from Twitter, replace the Google Sheet node with the respective node. Update the Function node accordingly.
by Harshil Agrawal
This workflow allows you to get analytics of a website and store it Airtable. In this workflow, we get the analytics for the sessions grouped by the country. Based on your use-case, you can select different Dimensions and set different Metrics. You can use the Cron node or the Interval node to trigger the workflow on a particular interval and fetch the analytics data regularly. Based on your use-case, you might want to store the data returned by Google Analytics to a database or a Google Sheet. Replace the Airtable node with the appropriate node.
by Shashikanth
Source code, I maintain this worflow here. Usage Guide This workflow backs up all workflows as JSON files named in the [workflow_name].json format. Steps Create GitHub Repository Skip this step if using an existing repository. Add GitHub Credentials In Credentials, add the GitHub credential for the repository owner. Download and Import Workflow Import this workflow into n8n. Set Global Values In the Globals node, set the following: repo.owner: GitHub username of the repository owner. repo.name: Name of the repository for backups. repo.path: Path to the folder within the repository where workflows will be saved. Configure GitHub Nodes Edit each GitHub node in the workflow to use the added credentials. Workflow Logic Each workflow run handles files based on their status: New Workflow If a workflow is new, create a new file in the repository. Unchanged Workflow If the workflow is unchanged, skip to the next item. Changed Workflow If a workflow has changes, update the corresponding file in the repository. Current Limitations / Needs work Name Change of Workflows If a workflow is renamed or deleted in n8n, the old file remains in the repository. Deleted Workflows Deleted workflows in n8n are not removed from the repository.
by Danielle Gomes
This n8n workflow collects and summarizes news from multiple RSS feeds, using OpenAI to generate a concise summary that can be sent to WhatsApp or other destinations. Perfect for automating your daily news digest. π Workflow Breakdown: Schedule Trigger Start the workflow on your desired schedule (daily, hourly, etc.). π¨ Note: Set the trigger however you wish. RSS Feeds (My RSS 01β04) Fetches articles from four different RSS sources. π¨ Note: You can add as many RSS feeds as you want. Edit Fields (Edit Fields1β3) Normalizes RSS fields (title, link, etc.) to ensure consistency across different sources. Merge (append mode) Combines the RSS items into a single unified list. Filter Optionally filter articles by keywords, date, or categories. Limit Limits the analysis to the 10 most recent articles. π¨ Note: This keeps the result concise and avoids overloading the summary. Aggregate Prepares the selected news for summarization by combining them into a single content block. OpenAI (Message Assistant) Summarizes the aggregated news items in a clean and readable format using AI. Send Summary to WhatsApp Sends the AI-generated summary to a WhatsApp endpoint via webhook (yoururlapi.com). You can replace this with an email service, Google Drive, or any other destination. π¨ Note: You can send it to your WhatsApp API, email, drive, etc. No Operation (End) Final placeholder to safely close the workflow. You may expand from here if needed.
by Felix
How It Works This workflow automatically detects duplicate invoices from Gmail. Incoming PDF attachments are scanned by the easybits AI Extractor, then checked against the Master Finance File in Google Sheets. Duplicates trigger a Slack alert β new invoices get added to the sheet. Flow overview: Gmail picks up new emails labeled as invoices (polls every minute) The PDF attachment is extracted and converted to base64 easybits Extractor reads the document and returns structured data The invoice number is compared against all existing entries in Google Sheets If duplicate β Slack DM alert to felix.sattler If new β Invoice is appended to the Master Finance File Step-by-Step Setup Guide 1. Set Up Your easybits Extractor Pipeline Before connecting this workflow, you need a configured extraction pipeline on easybits. Go to extractor.easybits.tech and click "Create a Pipeline". Fill in the Pipeline Name and Description β describe the type of document you're processing (e.g. "Invoice / Receipt"). Upload a sample receipt or invoice as your reference document. Click "Map Fields" and define the following fields to extract: invoice_number (String) β Unique identifier of the invoice, e.g. IN-2026-0022514 total_amount (Number) β Total amount due on the invoice, e.g. 149.99 Click "Save & Test Pipeline" in the Test tab to verify the extraction works correctly. Go to Pipeline Details β View Pipeline and copy your Pipeline ID and API Key. 2. Connect the easybits Node in n8n Open the HTTP Request node in the workflow. Replace the Pipeline ID in the URL with your own. Set up a Bearer Auth credential with your API Key. > The node sends the PDF to your pipeline and receives the extracted fields back under json.data. 3. Connect Gmail Open the Gmail Trigger node. Connect your Gmail account via OAuth2. Create a label called invoice in Gmail (or use your preferred label). Update the label filter in the node to match your label. Make sure Download Attachments is enabled under Options. > The trigger polls every minute for new emails matching the label. 4. Connect Google Sheets Open the Check Google Sheets and Add to Master List nodes. Connect your Google Sheets account via OAuth2. Select your target spreadsheet (Master Finance File) and sheet. Make sure your sheet has at least these columns: Invoice Number and Final Amount (EUR). 5. Connect Slack Go to api.slack.com/apps and create a new Slack App. Under OAuth & Permissions, add these Bot Token Scopes: chat:write, chat:write.public, channels:read, groups:read, users:read, users.profile:read. Install the app to your workspace via Settings β Install App. Copy the Bot User OAuth Token and add it as a Slack API credential in n8n. Open the Slack: Alert Finance node and select the user or channel to receive duplicate alerts. 6. Activate the Workflow Click the "Active" toggle in the top-right corner of n8n to enable the workflow. Label an email with an invoice attachment as invoice in Gmail to test it end to end. Check your Google Sheet β a new row with the invoice number and amount should appear. Send the same invoice again β you should receive a Slack DM alerting you to the duplicate.
by Muzaffer AKYIL
Docker Registry Cleanup Template This template is designed to automatically clean up old image tags in the Docker registry and perform garbage collection. Features List all images in the registry Preserve the last 10 tags for each image (latest tag is always preserved) Delete old tags Email notification for Successful/Excused cancellation Registry garbage collection automation Failure notification in error conditions Prerequisites Docker Registry v2 API access Basic Authentication credentials SMTP email settings (for notifications) SSH node installed on n8n (for garbage collection) Installation 1. Identity Information Add the following credentials in n8n: HTTP Basic Auth**: For Registry access SSH Private Key**: For Garbage collection command Email SMTP**: For notifications 2. Set Variables Replace your-registry-url with your actual registry URL on all nodes: βurl": βhttps://your.registry.com/v2/_catalogβ. Customisation Retention Policy: Set the number of tags to be retained by changing the slice(0, 10) value in the Identify Tags to Remove node Schedule: Change the frequency of operation at the Trigger node Notification Content: Customise email templates according to your needs Notes Check DELETE operations before running in a test environment Make sure that the registry is not in read-only mode The registry may need to be put into maintenance mode for garbage collection Step Details: Retrieving image information:** The workflow starts by fetching a list of images and their associated tags from the Docker registry. Filtering and sorting:** The retrieved tags are then filtered and sorted based on specific criteria, such as creation date and tag name. Deleting old tags:** The workflow identifies old or unused tags and attempts to delete them from the registry. Sending notifications:** The workflow sends email notifications to inform the user about the status of the cleanup process, including any errors or successes. Executing additional cleanup tasks:** Finally, the workflow executes an SSH command on the Docker registry server to perform additional cleanup tasks, such as garbage collection. TL;DR In summary, this n8n template provides a robust and automated solution for managing and cleaning up Docker registries. By regularly running this workflow, users can ensure that their registry remains organized and efficient, and avoid running out of storage space.-
by AiAgent
What It Does This intelligent workflow simplifies the complex task of determining whether a website is legitimate or potentially a scam. By simply submitting a URL through a form, the system initiates a multi-agent evaluation process. Four dedicated AI agentsβeach powered by GPT-4o and connected to SerpAPIβanalyze different dimensions of the website: domain and technical details, search engine signals, product and pricing patterns, and on-site content analysis. Their findings are then passed to a fifth AI agent, the Analyzer, powered by GPT-4o mini, which consolidates the data, scores the site on a scale of 1β10 for scam likelihood, and presents the findings in a clear, structured format for the user. Who It's For This workflow is ideal for anyone who needs to quickly and reliably assess the trustworthiness of a website. Whether you're a consumer double-checking a store before making a purchase, a small business owner validating supplier sites, a cybersecurity analyst conducting threat assessments, or a developer building fraud detection into your platform β this tool offers fast, AI-powered insights without the need for manual research or technical expertise. It's designed for both individuals and teams who value accurate, scalable scam detection. How It Works The process begins with a simple form submission where the user enters the URL of the website they want to investigate. Once submitted, the workflow activates four specialized AI agentsβeach powered by GPT-4o and connected to SerpAPIβto independently analyze the site from different angles: Agent 1 examines domain age, SSL certificates, and TLD trustworthiness. Agent 2 reviews search engine results, forum mentions, and public scam reports. Agent 3 analyzes product pricing patterns and brand authenticity. Agent 4 assesses on-site content quality, grammar, legitimacy of claims, and presence of business info. Each agent returns its findings, which are then aggregated and passed to a fifth AI agentβthe Analyzer. This final agent, powered by GPT-4o mini, evaluates all the input, assigns a scam likelihood score from 1 to 10, and compiles a neatly formatted summary with organized insights and a disclaimer for context. Set UP You will need to obtain an Open AI API key from platform.openai.com/api-keys After you obtain this Open AI API key you will need to connect it to the Open AI Chat Model for all of the Tools agents (Analyzer, Domain & Technical Details, Search Engine Signals, Product & Pricing Patterns, and Content Analysis Tools Agents). You will now need to fund your Open AI account. GPT 4o costs ~$0.01 to run the workflow. Next you will need to create a SerpAPI account at https://serpapi.com/users/sign_up After you create an account you will need to obtain a SerpAPI key. You will then need to use this key to connect to the SerpAPI tool for each of the tools agents (Domain & Technical Details, Search Engine Signals, Product & Pricing Patterns, and Content Analysis Tools Agents) Tip: SerpAPI will allow you to run 100 free searches each month. This workflow uses ~5-15 SerpAPI searches per run. If you would like to utilize the workflow more than that each month, create multiple SerpAPI accounts and have an API key for each account. When you utilize all 100 free searches for an account, switch to the API key for another account within the workflow. Disclaimer This tool is designed to assist in evaluating the potential risk of websites using AI-generated insights. The scam likelihood score and analysis provided are based on publicly available information and should not be considered a definitive or authoritative assessment. This tool does not guarantee the accuracy, safety, or legitimacy of any website. Users should perform their own due diligence and use independent judgment before engaging with any site. N8N, OpenAI, its affiliates, and the creators of this workflow are not responsible for any loss, damages, or consequences arising from the use of this tool or the actions taken based on its results.