by Sirhexalot
This n8n workflow enables you to export data from Zammad, including Users, Roles, Groups, and Organizations, into individual Excel files. It simplifies data handling and reporting by creating structured outputs for further processing or sharing. Features Export Users with associated details such as email, firstname, lastname, role_ids, and group_ids. Export Roles and Organizations with their respective identifiers and names. Convert all data into separate Excel files for easy access and use. Usage Import this workflow into your n8n instance. Configure the required Zammad API credentials (zammad_base_url and zammad_api_key) in the Basic Variables node. Run the workflow to generate Excel files containing Zammad data. Issues and Suggestions If you encounter any issues or have suggestions for improvement, please report them on the GitHub repository. We appreciate your feedback to help enhance this workflow!
by Zacharia Kimotho
What problem is this workflow solving? This workflow is aimed for email marketing enthusiasts looking for an easy way to either extract the domain from an email ad also check if the syntax is correct without having to use the code node. How this works For this to work, replace the debugger node with your actual data source. Map your data at match the above layout Run your workflow and check for all the emails that are either valid or not Once done, you will have a list of all your emails, domains, and whether they are valid or not.
by Harshil Agrawal
This workflow allows you to receive a Mattermost message when meeting notes get added to the Notion. Prerequisites Create a table in Notion similar to this: Meeting Notes Follow the steps mentioned in the documentation, to create credentials for the Notion Trigger node. Create create credentials for Mattermost. Notion Trigger: The Notion Trigger node will trigger the workflow when new data gets added to Notion. IF node: This node will check if the notes belong to the team Marketing. If the team is Marketing the node will true, otherwise false. Mattermost node: This node will send a message about the new data in the channel 'Marketing' in Mattermost. If you have a different channel, use that instead. You can even replace the Mattermost node with nodes of other messaging platforms, like Slack, Telegram, Discord, etc. NoOp node: Adding this node here is optional, as the absence of this node won't make a difference to the functioning of the workflow.
by Manu
This workflow will take all emails you put into a certain folder, upload any attachements to Nextcloud, and mark the emails as read (configurable). Attachements will be saved with automatically generated filenames: 2021-01-01_From-Sender-Name_Filename-of-attachement.pdf Instructions: Allow lodash to be used in n8n (or rewrite the code...) NODE_FUNCTION_ALLOW_EXTERNAL=lodash (environment variable) Import workflow Set credentials for Email & Nextcloud nodes Configure to use correct folder / custom filters Activate Custom filter examples: Only unread emails: Custom Email Config = ["UNSEEN"] Filter emails by 'to' address: Custom Email Config = [["TO", "example+invoices@posteo.de"]]
by Recrutei Automaรงรตes
What This Workflow Does This workflow automates the candidate nurturing process, solving the common problem of candidates losing interest or "ghosting" after an application. It keeps them engaged and informed by sending a personalized, multi-channel (WhatsApp & Gmail) sequence of follow-up messages over their first week. The automation triggers when a new candidate is added to your ATS (e.g., via a Recrutei webhook). It then uses AI to generate a custom 3-part message (for Day 1, Day 3, and Day 7) tailored to the candidate's age and the specific job they applied for, ensuring a professional and empathetic experience that strengthens your employer brand. How it Works Trigger: A Webhook node captures the new candidate data from your Applicant Tracking System (ATS) or form. Data Preparation: Two Code nodes clean the incoming data. The first (Separating information) extracts key fields and formats the phone number. The second (Extract age) calculates the candidate's age from their birthday to be used by the AI. AI Content Generation: The workflow sends the candidate's details (name, age, job title) to an AI model (AI Recruitment Assistant). The AI has a detailed system prompt to generate three distinct messages for Day 1 (Thank You), Day 3 (Friendly Reminder), and Day 7 (Final Reinforcement), adapting its tone based on the candidate's age. Split Messages: A Code node (Separating messages per days) receives the single text block from the AI and splits it into three separate variables (day1, day3, day7). Day 1 Send: The workflow immediately sends the day1 message via both Gmail and WhatsApp (configured for Evolution API). Day 3 Send: A "Wait" node pauses the workflow for 2 days, after which it sends the day3 message. Day 7 Send: Another "Wait" node pauses for 4 more days, then sends the final day7 message, completing the 7-day nurturing sequence. Setup Instructions This workflow is plug-and-play once you configure the following 5 steps: Webhook Node: Copy the Test URL from the Webhook node and configure it in your ATS (e.g., Recrutei) or form builder to trigger whenever a new candidate is added. Run one test submission to make the data structure visible to n8n. AI Credentials: In the AI Recruitment Assistant node, select or create your OpenAI API credential. MCP Credential (Optional): If you use a Recrutei MCP, paste your endpoint URL into the MCP Recrutei node. Gmail Credentials: In all three Message Gmail nodes (Day 1, 3, 7), select or create your Gmail (OAuth2) credential. Optional: In the same nodes, go to Options and change the Sender Name from your_company to your actual company name. WhatsApp (Evolution API): This template is pre-configured for the Evolution API. In all three Message WhatsApp nodes (Day 1, 3, 7), you must: URL: Replace {server-url} and {instance} with your Evolution API details. Headers: In the "Header Parameters" section, replace your_api_key with your actual Evolution API key.
by Simon
This n8n workflow simplifies the process of removing backgrounds from images stored in Google Drive. By leveraging the PhotoRoom API, this template enables automatic background removal, padding adjustments, and output formatting, all while storing the updated images back in a designated Google Drive folder. This workflow is very useful for companies or individuals that are spending a lot of time into removing the background from product images. How it Works The workflow begins with a Google Drive Trigger node that monitors a specific folder for new image uploads. Upon detecting a new image, the workflow downloads the file and extracts essential metadata, such as the file size. Configurations are set for background color, padding, output size, and more, which are all customizable to match specific requirements. The PhotoRoom API is called to process the image by removing its background and adding padding based on the settings. The processed image is saved back to Google Drive in the specified output folder with an updated name indicating the background has been removed. Requirements PhotoRoom API Key Google Drive API Access Customizing the Workflow Easily adjust the background color, padding, and output size using the configuration node. Modify the output folder path in Google Drive or replace Google Drive with another storage service if needed. For advanced use cases, integrate further image processing steps, such as adding captions or analyzing content using AI.
by tanaypant
This workflow is the third of three. You can find the other workflkows here: Incident Response Workflow - Part 1 Incident Response Workflow - Part 2 Incident Response Workflow - Part 3 We have the following nodes in the workflow: Webhook node: This trigger node listens to the event when the Resolve button is clicked. PagerDuty node: This node changes the status of the incident report from Acknowledged to Resolved in PagerDuty. Jira Software node: This node moves the incident issue to Done. Mattermost node: This node publishes a message in the auxiliary channel mentioning that the incident has been marked as resolved in PagerDuty and Jira. Mattermost node: This node publishes a message in the specified Incidents channel that the incident has been resolved by the on-call team.
by Angel Menendez
CallForge - AI Sales Call Processing & Insights Extraction Automate sales call analysis with AI-powered insights for sales, marketing, and product teams. Who is This For? This workflow is designed for: โ Sales teams looking to extract structured insights from Gong call transcripts. โ Marketing professionals seeking AI-driven customer pain points & content strategy. โ Product teams needing feedback from sales calls to prioritize feature development. ๐ What Problem Does This Workflow Solve? Manually analyzing Gong.io sales call transcripts is slow, inconsistent, and lacks structured insights. With CallForge, you can: โ Extract AI-powered insights about use cases, objections, competitors, and next steps. โ Provide structured marketing & product intelligence to enhance strategy. โ Automatically store call insights in Notion and Salesforce for easy access. โ Ensure resilience with automated reruns on failed workflows (handling Notion API limits). โ Improve decision-making with AI-powered competitor and sentiment analysis. ๐ Key Workflow Features ๐ค AI-Powered Transcript Analysis Uses AI to identify use cases, objections, competitors, and customer pain points. Categorizes insights for sales, marketing, and product teams. ๐ AI Agent Breakdown ๐น Sales AI Agent โ Extracts customer objections, pain points, competitors, and next steps. ๐น Marketing AI Agent โ Identifies recurring topics, keyword trends, and content opportunities. ๐น Product AI Agent โ Captures feature requests and AI/ML-related references. ๐ Structured Output Processing Sales Data Processor* โ Stores insights in *Notion & Salesforce** for sales tracking. Marketing Data Processor* โ Extracts *SEO & content strategy insights** for marketing teams. Product AI Data Processor* โ Logs *customer feedback* to prioritize *feature development**. ๐ก Competitor & Integration Analysis Tracks competing products mentioned in calls**. Identifies integration needs**, flagging workarounds used by prospects. ๐ข Real-Time Slack Notifications Alerts teams on workflow progress** and completed call analyses. ๐ Failure Resilience & Automated Re-Runs If a Notion API limit is reached, the process resumes automatically. ๐ How This Works ๐ 1. Trigger & Call Data Processing The workflow retrieves Gong call transcripts and metadata. Normalizes data**, correcting common mispronunciations like "n8n." ๐ค 2. AI Agents Analyze the Call Sales Agent** โ Extracts actionable insights for sales follow-ups. Marketing Agent* โ Identifies *recurring themes* and *keyword trends**. Product Agent* โ Captures *feature requests and AI/ML usage mentions**. ๐ก 3. Data is Stored in Notion & Salesforce Logs AI-extracted insights* in *Notion** for structured tracking. Pushes sales-related data* to *Salesforce** for team accessibility. ๐ 4. Slack Alerts for Teams Notifies sales, marketing, and product teams** about extracted insights. CallForge - 01 - Filter Gong Calls Synced to Salesforce by Opportunity Stage CallForge - 02 - Prep Gong Calls with Sheets & Notion for AI Summarization CallForge - 03 - Gong Transcript Processor and Salesforce Enricher CallForge - 04 - AI Workflow for Gong.io Sales Calls CallForge - 05 - Gong.io Call Analysis with Azure AI & CRM Sync CallForge - 06 - Automate Sales Insights with Gong.io, Notion & AI CallForge - 07 - AI Marketing Data Processing with Gong & Notion CallForge - 08 - AI Product Insights from Sales Calls with Notion ๐ Sample Output Data 1๏ธโฃ Sales Insights { "UseCases": [ { "Summary": "A manufacturing company wants to automate inventory tracking and reduce manual entry delays.", "DepartmentTags": ["Operations"], "IndustryTags": ["Manufacturing"], "ImplementationStatus": "Evaluating" } ], "Objection": { "ObjectionTags": ["Feature Limitation"], "Nature": "The prospect wanted a deeper integration with their ERP system, which n8n currently lacks." }, "CallSummary": "The call focused on automation for supply chain processes. The prospect expressed interest but wanted confirmation on ERP integration capabilities.", "NextSteps": ["Schedule a follow-up demo for ERP integration."] } 2๏ธโฃ Marketing Insights { "MarketingInsights": [ { "Tag": "Workflow Template Request", "Summary": "The prospect requested a template for automating CRM lead tracking." } ], "RecurringTopics": [ { "Topic": "CRM Integration", "Mentions": 3, "Context": "Discussed how n8n could sync CRM data automatically." } ], "ActionableInsights": [ { "RecommendationType": "Tutorial", "Title": "How to Automate CRM Lead Tracking with n8n", "Topic": "CRM Integration", "Rationale": "The prospect expressed a need for CRM automation templates." } ] } 3๏ธโฃ Product Feedback { "ProductFeedback": [ { "Sentiment": "Positive", "Feedback": "The external speaker praised the simplicity of n8n's UI, making it easier for non-developers to automate tasks." }, { "Sentiment": "Negative", "Feedback": "The external speaker mentioned frustration over the lack of a dedicated ERP integration node." } ], "AI_ML_References": { "Exist": true, "Context": "The external speaker mentioned using AI for automating customer ticket categorization.", "Details": { "DevelopmentStatus": "Building", "Department": "Support", "RequiresAgents": true, "RequiresRAG": false, "RequiresChat": "Yes: External App (e.g., Slack)" } } } ๐ง How to Customize This Workflow ๐ก ๐ Change Data Storage โ Swap Notion for Airtable, HubSpot, or another CRM. ๐ก ๐ฉ Customize Slack Notifications โ Send alerts via email, webhook, or another channel. ๐ก ๐ Modify AI Processing โ Adjust AI models or processing prompts. ๐ก ๐ Add More Integrations โ Sync insights with Pipedrive, HubSpot, or another CRM. ๐ Why Use This Workflow? โ Automates Gong call transcript analysis, eliminating manual work. โ Improves collaboration by structuring insights for sales, marketing, and product teams. โ Boosts sales conversions by identifying objections and next steps. โ Enhances marketing and SEO strategy with AI-driven insights. โ Optimizes product roadmap decisions based on customer feedback. This workflow scales AI-powered sales intelligence for better decision-making, content strategy, and sales enablement. ๐
by Nskha
Overview The [n8n] YouTube Channel Advanced RSS Feeds Generator workflow facilitates the generation of various RSS feed formats for YouTube channels without requiring API access or administrative permissions. It utilizes third-party services to extract data, making it extremely user-friendly and accessible. Key Use Cases and Benefits Content Aggregation**: Easily gather and syndicate content from any public YouTube channel. No API Key Required**: Avoid the complexities and limitations of Google's API. Multiple Formats**: Supports ATOM, JSON, MRSS, Plaintext, Sfeed, and direct YouTube XML feeds. Flexibility**: Input can be a YouTube channel or video URL, ID, or username. Services/APIs Utilized This workflow integrates with: commentpicker.com**: For retrieving YouTube channel IDs. rss-bridge.org**: To generate various RSS formats. Configuration Instructions Start the Workflow: Activate the workflow in your n8n instance. Input Details: Enter the YouTube channel or video URL, ID, or username via the provided form trigger. Run the Workflow: Execute the workflow to receive links to 13 different RSS feeds, including community and video content feeds. Screenshots Additional Notes Customization**: You can modify the RSS feed formats or integrate additional services as needed. Support and Contributions For support, questions, or contributions, please visit the n8n community forum or the GitHub repository. We welcome contributions from the community!
by Eduard
Create, iterate, and share! Transform a single image through multiple scenes while maintaining consistency. โจ What this workflow does This template showcases FLUX.1 Kontext - Black Forest Labs' in-context image generation model that excels at maintaining character features across multiple transformations. Combined with the Upload Post community node for effortless multi-platform social media posting, you can create and share compelling visual stories instantly. The workflow demonstrates FLUX Kontext's core strength: character consistency across multiple image generations. Starting with a single input image, it: ๐ผ๏ธ Loads an initial character image (example: a cute animal mascot) ๐ Defines multiple scene transformation prompts ๐ Iteratively generates new scenes while preserving exact character features ๐ฏ Maintains visual consistency by reusing binary data from previous generations ๐ฑ Auto-posts the complete transformation series to multiple social platforms simultaneously ๐ Key Features: The Consistency Advantage Character Preservation**: FLUX Kontext's signature feature - maintains character features and style across transformations (requires specific prompting techniques) Iterative Context Building**: Each generation uses the previous image as context, creating visual continuity Binary Data Reuse**: Smart workflow design that feeds output from one generation as input to the next Multi-Scene Storytelling**: Transform your character across different environments while keeping them recognizable One-Click Multi-Platform Posting*: Upload Post eliminates the tedious process of posting to each platform individually ๐ฑ Why use Upload Post? Posting the same content to TikTok, Instagram, LinkedIn, YouTube, Facebook, X (Twitter), and Threads individually is time-consuming and error-prone. The Upload Post service* simplifies this process: โ Connect once, post everywhere: Link all your social media accounts to Upload Post โ Single API call: Post to multiple platforms with one simple node โ No more platform juggling: Skip the endless switching between apps and dashboards โ Consistent timing: All platforms get your content simultaneously โ Trusted by 3,751+ users: Proven solution for content creators and marketers Instead of spending 30+ minutes manually posting to each platform, Upload Post does it all in seconds with a single n8n node! ๐ ๏ธ Prerequisites Required Accounts: Black Forest Labs API: Create account at dashboard.bfl.ai Get your API key for FLUX Kontext Pro access Upload Post Account: Sign up at upload-post.com* Connect your social media profiles (TikTok, Instagram, LinkedIn, YouTube, Facebook, X/Twitter, Threads) Get API credentials for automated posting Free tier available: 10 uploads/month ๐ก Perfect For: Character Designers** maintaining brand character integrity across scenes Social Media Managers** creating engaging visual story series without manual posting Brand Marketers** ensuring character consistency across campaigns Storytellers** building visual narratives with consistent protagonists Agencies** managing multiple client accounts efficiently ๐ง Customization Options: Modify transformation prompts** to create your own character journey Adjust iteration steps** Change initial character image** Configure social platform targeting** (choose which platforms to post to) Customize post content** and formatting Experiment with different consistency scenarios** \ Affiliate link*
by Harshil Agrawal
Based on your use case, you might want to trigger a workflow if new data gets added to your database. This workflow allows you to send a message to Mattermost when new data gets added in Google Sheets. The Interval node triggers the workflow every 45 minutes. You can modify the timing based on your use case. You can even use the Cron node to trigger the workflow. If you wish to fetch new Tweets from Twitter, replace the Google Sheet node with the respective node. Update the Function node accordingly.
by Harshil Agrawal
This workflow allows you to get analytics of a website and store it Airtable. In this workflow, we get the analytics for the sessions grouped by the country. Based on your use-case, you can select different Dimensions and set different Metrics. You can use the Cron node or the Interval node to trigger the workflow on a particular interval and fetch the analytics data regularly. Based on your use-case, you might want to store the data returned by Google Analytics to a database or a Google Sheet. Replace the Airtable node with the appropriate node.