by Simon
This n8n workflow simplifies the process of removing backgrounds from images stored in Google Drive. By leveraging the PhotoRoom API, this template enables automatic background removal, padding adjustments, and output formatting, all while storing the updated images back in a designated Google Drive folder. This workflow is very useful for companies or individuals that are spending a lot of time into removing the background from product images. How it Works The workflow begins with a Google Drive Trigger node that monitors a specific folder for new image uploads. Upon detecting a new image, the workflow downloads the file and extracts essential metadata, such as the file size. Configurations are set for background color, padding, output size, and more, which are all customizable to match specific requirements. The PhotoRoom API is called to process the image by removing its background and adding padding based on the settings. The processed image is saved back to Google Drive in the specified output folder with an updated name indicating the background has been removed. Requirements PhotoRoom API Key Google Drive API Access Customizing the Workflow Easily adjust the background color, padding, and output size using the configuration node. Modify the output folder path in Google Drive or replace Google Drive with another storage service if needed. For advanced use cases, integrate further image processing steps, such as adding captions or analyzing content using AI.
by tanaypant
This workflow is the third of three. You can find the other workflkows here: Incident Response Workflow - Part 1 Incident Response Workflow - Part 2 Incident Response Workflow - Part 3 We have the following nodes in the workflow: Webhook node: This trigger node listens to the event when the Resolve button is clicked. PagerDuty node: This node changes the status of the incident report from Acknowledged to Resolved in PagerDuty. Jira Software node: This node moves the incident issue to Done. Mattermost node: This node publishes a message in the auxiliary channel mentioning that the incident has been marked as resolved in PagerDuty and Jira. Mattermost node: This node publishes a message in the specified Incidents channel that the incident has been resolved by the on-call team.
by Matheus Weckwerth
This workflow automates daily LinkedIn posts using Notion. It starts by fetching the day's post from a Notion database, processes and formats the content, including images, then publishes it on LinkedIn. Finally, it updates the post status in the Notion database. Set up Notion and LinkedIn credentials as required.
by Robert Breen
This n8n workflow dynamically generates a realistic sample dataset based on a single topic you provide. It uses OpenAI (via LangChain) and n8n’s built-in nodes to: Generate structured JSON data for 5 columns with 3–5 values each Flatten that data into a single text blob Infer meaningful column names via a second AI call Pivot, split, merge, and rename columns automatically Output a clean, labeled dataset ready for export or further processing ⚙️ Prerequisites OpenAI API Key Visit: https://platform.openai.com/account/api-keys Create a new key In n8n: Credentials → New → OpenAI API, paste key, name it “OpenAi account” LangChain nodes enabled in your n8n instance 🥇 Step 1: Set Up OpenAI Credential Go to OpenAI API Keys Create and copy your key In n8n: Credentials → New → OpenAI API → paste key as “OpenAi account” 🥈 Step 2: Manual Trigger Add Manual Trigger to start the workflow 🥉 Step 3: Set Topic Add a Set node named Set Topic to Search Field: Topic = n8n use cases (or any topic you choose) ✨ Step 4: Generate Structured Data LangChain Agent** node Generate Random Data Connect to OpenAI Chat Model1 and Tool: Inject Creativity1 System prompt: instruct AI to output 5 columns of realistic values in JSON 🔧 Step 5: Parse AI Output Structured Output Parser** to validate JSON 🔄 Step 6: Flatten Data Code** node Outpt all Data to One Field Joins all values into a comma-separated string for column naming 🧠 Step 7: Generate Column Names LangChain Agent** Generate Column Names Connect to OpenAI Chat Model2 Prompt: infer 5 column names from the string 🔢 Step 8: Pivot Names Row Code** node Pivot Column Names transforms array into { column1: name1, … } 🪓 Step 9: Split Columns 5 SplitOut nodes to break each array back into rows per column 🔗 Step 10: Merge Rows Merge** node Merge Columns together using combineByPosition 🏷️ Step 11: Rename Columns Set** node Rename Columns assigns the AI-generated names to each column 🔗 Step 12: Final Output Merge** Append Column Names combines data and header row 🏁 Done! You now have a fully AI-driven, labeled dataset generated from a single topic—no external services needed. Easily extend by adding a Google Sheets or HTTP node to export. 📬 Need Help or Want to Customize This? 📧 robert@ynteractive.com 🔗 LinkedIn
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically tracks email campaign performance metrics and triggers smart follow-up actions based on engagement data. It saves you time by eliminating the need to manually monitor campaign reports and provides intelligent re-engagement strategies for improving email marketing ROI. Overview This workflow automatically scrapes email service provider (ESP) reports to extract campaign performance metrics like open rates, click-through rates, and bounce rates. It uses AI to analyze the data and automatically sends targeted follow-up emails to re-engage subscribers who opened but didn't click, maximizing campaign effectiveness. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping ESP campaign reports without being blocked OpenAI**: AI agent for intelligent campaign data analysis and decision making Gmail**: For sending automated follow-up engagement emails How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Gmail: Connect your Gmail account for sending follow-up emails Customize: Set ESP report URLs and define engagement thresholds for triggering follow-ups Use Cases Email Marketing**: Automatically optimize campaign performance with smart follow-ups Marketing Automation**: Trigger re-engagement campaigns based on behavior data Performance Tracking**: Monitor email metrics without manual ESP login Customer Retention**: Re-engage subscribers who showed interest but didn't convert Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #emailmarketing #campaigntracking #brightdata #webscraping #emailautomation #n8nworkflow #workflow #nocode #emailcampaigns #marketingautomation #emailperformance #campaignanalysis #emailmetrics #reengagement #marketingdata #emailoptimization #campaignmonitoring #emailanalytics #digitalmarketing #performancetracking #emailstrategy #conversionoptimization #marketinganalytics #emailroi #campaigninsights #emailengagement #marketingefficiency #automatedemail
by NanaB
Description This n8n workflow automates the entire process of creating and publishing AI-generated videos, triggered by a simple message from a Telegram bot (YTAdmin). It transforms a text prompt into a structured video with scenes, visuals, and voiceover, stores assets in MongoDB, renders the final output using Creatomate, and uploads the video to YouTube. Throughout the process, YTAdmin receives real-time updates on the workflow’s progress. This is ideal for content creators, marketers, or businesses looking to scale video production using automation and AI. You can see a video demonstrating this template in action here: https://www.youtube.com/watch?v=EjI-ChpJ4xA&t=200s How it Works Trigger: Message from YTAdmin (Telegram Bot) The flow starts when YTAdmin sends a content prompt. Generate Structured Content A Mistral language model processes the input and outputs structured content, typically broken into scenes. Split & Process Content into Scenes The content is split into categorized parts for scene generation. Generate Media Assets For each scene: Images: Generated using OpenAI’s image model. Voiceovers: Created using OpenAI’s text-to-speech. Audio files are encoded and stored in MongoDB. Scene Composition Assets are grouped into coherent scenes. Render with Creatomate A complete payload is generated and sent to the Creatomate rendering API to produce the video. Progress messages are sent to YTAdmin. The flow pauses briefly to avoid rate limits. Render Callback Once Creatomate completes rendering, it sends a callback to the flow. If the render fails, an error message is sent to YTAdmin. If the render succeeds, the flow proceeds to post-processing. Generate Title & Description A second Mistral prompt generates a compelling title and description for YouTube. Upload to YouTube The rendered video is retrieved from Creatomate. It’s uploaded to YouTube with the AI-generated metadata. Final Update A success message is sent to YTAdmin, confirming upload completion. Set Up Steps (Approx. 10–15 Minutes)Step 1: Set Up YTAdmin Bot Create a Telegram bot via BotFather and get your API token. Add this token in n8n's Telegram credentials and link to the "Receive Message from YTAdmin" trigger. Step 2: Connect Your AI Providers Mistral: Add your API key under HTTP Request or AI Model nodes. OpenAI: Create an account at platform.openai.com and obtain an API key. Use it for both image generation and voiceover synthesis. Step 3: Configure Audio File Storage with MongoDB via Custom API Receives the Base64 encoded audio data sent in the request body. Connects to the configured MongoDB instance (connection details are managed securely within the API- code below). Uses the MongoDB driver and GridFS to store the audio data. Returns the unique _id (ObjectId) of the stored file in GridFS as a response. This _id is crucial as it will be used in subsequent steps to generate the download URL for the audio file. My API code can be found here for reference: https://github.com/nanabrownsnr/YTAutomation.git Step 4: Set Up Creatomate Create a Creatomate account, define your video templates, and retrieve your API key. Configure the HTTP request node to match your Creatomate payload requirements. Step 5: Connect YouTube In n8n, add OAuth2 credentials for your YouTube account. Make sure your Google Cloud project has YouTube Data API enabled. Step 6: Deploy and Test Send a message to YTAdmin and monitor the flow in n8n. Verify that content is generated, media is created, and the final video is rendered and uploaded. Customization Options Change the AI Prompts Modify the generation prompts to adjust tone, voice, or content type (e.g., news recaps, product videos, educational summaries). Switch Messaging Platform Replace Telegram (YTAdmin) with Slack, Discord, or WhatsApp by swapping out the trigger and response nodes. Add Subtitles or Effects Integrate Whisper or another speech-to-text tool to generate subtitles. Add overlay or transition effects in the Creatomate video payload. Use Local File Storage Instead of MongoDB Swap out MongoDB upload http nodes with filesystem or S3-compatible storage. Repurpose for Other Platforms Swap YouTube upload with TikTok, Instagram, or Vimeo endpoints for broader publishing. **Need Help or Want to Customize This Workflow? If you'd like assistance setting this up or adapting it for a different use case, feel free to reach out to me at nanabrownsnr@gmail.com. I'm happy to help!**
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically tracks customer satisfaction scores across multiple platforms and surveys to help improve customer experience and identify areas for enhancement. It saves you time by eliminating the need to manually check different feedback sources and provides comprehensive satisfaction analytics. Overview This workflow automatically scrapes customer satisfaction surveys, review platforms, and feedback forms to extract satisfaction scores and sentiment data. It uses Bright Data to access various feedback platforms without being blocked and AI to intelligently analyze satisfaction trends and identify improvement opportunities. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping satisfaction surveys and review platforms without being blocked OpenAI**: AI agent for intelligent satisfaction analysis and trend identification Google Sheets**: For storing satisfaction scores and generating analytics reports How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your satisfaction tracking spreadsheet Customize: Define feedback sources and satisfaction metrics you want to monitor Use Cases Customer Experience**: Monitor satisfaction trends across all customer touchpoints Product Teams**: Identify product features that impact customer satisfaction Support Teams**: Track satisfaction scores for support interactions Management**: Get comprehensive satisfaction reporting for strategic decisions Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #customersatisfaction #satisfactionscores #brightdata #webscraping #customerexperience #n8nworkflow #workflow #nocode #satisfactiontracking #csat #nps #customeranalytics #feedbackanalysis #customerinsights #satisfactionmonitoring #experiencemanagement #customermetrics #satisfactionsurveys #feedbackautomation #customerfeedback #satisfactiondata #customerjourney #experienceanalytics #satisfactionreporting #customersentiment #experienceoptimization #satisfactiontrends #customervoice
by Lukas Kunhardt
Intelligently Segment PDFs by Table of Contents This workflow empowers you to automatically process PDF documents, intelligently identify or generate a hierarchical Table of Contents (ToC), and then segment the entire document's content based on these ToC headings. It effectively breaks down a large PDF into its constituent sections, each paired with its corresponding heading and hierarchical level. Why It's Useful Unlock the true structure of your PDFs for granular access and advanced processing: AI Agent Tool:** A key use case is to provide this workflow as a tool to an AI agent. The agent can then use the segmented output to "read" and navigate to specific sections of a document to answer questions, extract information, or perform tasks with much greater accuracy and efficiency. Targeted Content Extraction:** Programmatically pull out specific chapters or subsections for focused analysis, summarization, reporting, or repurposing content. Enhanced RAG Systems:** Improve your Retrieval Augmented Generation (RAG) pipelines by feeding them well-defined, contextually relevant document sections instead of entire, monolithic PDFs. This leads to more precise AI-generated responses. Modular Document Processing:** Process different parts of a document using distinct logic in subsequent n8n workflows by acting on individual sections. Data Preparation:** Seamlessly convert lengthy PDFs into a structured format where each section (including its heading, level, and content in multiple formats) becomes a distinct, manageable item. How It Works Ingestion & Advanced Parsing: The workflow ingests a PDF (via a provided URL or a pre-set one for manual runs). It then utilizes Chunkr.ai to perform Optical Character Recognition (OCR) and parse the document into detailed structural elements, extracting text, HTML, and Markdown for each segment. AI-Powered Table of Contents Generation: A Google Gemini AI model analyzes the initial pages of the document (where a ToC often resides) along with section headers extracted by Chunkr as a fallback. This allows it to construct an accurate, hierarchical Table of Contents in a structured JSON format, even if the PDF lacks an explicit ToC or if it's poorly formatted. Precise Content Segmentation: Sophisticated custom code then meticulously maps the AI-generated ToC headings to their corresponding content within the parsed document from Chunkr. It intelligently determines the precise start and end of each section. Structured & Flexible Output: The primary output provides each identified section as an individual n8n item. Each item includes the heading text, its hierarchical level (e.g., 1, 1.1, 2), and the full content of that section in Text, HTML, and Markdown formats. Optionally, the workflow can also reconstruct the entire document into a single, navigable HTML file or a clean Markdown file. What You Need To run this workflow, you'll need: Input PDF:** When triggered by another workflow: A URL pointing to the PDF document. When triggered manually: The workflow uses a pre-configured sample PDF from Google Drive for demonstration (this can be customized). Chunkr.ai API Key:** Required for the initial parsing and OCR of the PDF document. You'll need to insert this into the relevant HTTP Request nodes. Google Gemini API Credentials:** Necessary for the AI model to intelligently generate the Table of Contents. This should be configured in the Google Gemini Chat Model nodes. Outputs The workflow primarily generates: Individual Document Sections:** A series of n8n items. Each item represents a distinct section of the PDF and contains: heading: The text of the section heading. headingLevel: The hierarchical level of the heading (e.g., 1 for H1, 2 for H2). sectionText: The plain text content of the section. sectionHTML: The HTML content of the section. sectionMarkdown: The Markdown content of the section. Alternatively, you can configure the workflow to output: Full Reconstructed Document:** A single HTML file representing the entire processed document. A single Markdown file representing the entire processed document. This workflow is ideal for anyone looking to deconstruct PDFs into meaningful, manageable parts for advanced automation, AI integration, or detailed content analysis.
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically monitors publicly available competitor financial data—funding rounds, earnings, and SEC filings—and alerts your team to significant changes. Gain an edge by reacting to financial moves faster. Overview Using Bright Data, the automation scrapes Crunchbase, press releases, and SEC Edgar filings. OpenAI extracts key figures (revenue, funding amount, valuation) and assesses the potential impact. Highlights are posted to Slack and stored in Airtable for long-term tracking. Tools Used n8n** – Drives the automation Bright Data** – Scrapes financial disclosure sites OpenAI** – Extracts numbers and generates insights Slack** – Sends real-time alerts Airtable** – Maintains a financial timeline database How to Install Import the Workflow into n8n. Configure Bright Data credentials. Set Up OpenAI API key. Authorize Slack & Airtable. Customize Competitor List & Thresholds in the Set node. Use Cases Competitive Intelligence**: Track rivals’ financial health. Investor Relations**: Benchmark against peers. Strategic Planning**: Identify acquisition targets. Sales Enablement**: Time outreach after funding events. Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #financialmonitoring #competitoranalysis #brightdata #openai #secfilings #fundingrounds #n8nworkflow #nocode
by Don Jayamaha Jr
A powerful sub-agent that collects real-time market structure data from Binance for any trading pair — including price, volume, order book depth, and candlestick snapshots across multiple timeframes (15m, 1h, 4h, 1d). 🎥 Watch Tutorial: 🎯 Purpose This workflow powers the Quant AI system with: ✅ Real-time price feed (/ticker/price) ✅ 24-hour stats (OHLC, % change, volume via /ticker/24hr) ✅ Live order book depth (/depth) ✅ Latest candlestick data (/klines) for all major intervals All outputs are parsed and formatted using GPT and returned to the parent agent (e.g., Financial Analyst Tool) as a Telegram-optimized summary. ⚙️ Workflow Architecture | Node | Role | | ------------------------------------ | ------------------------------------------------------------ | | 🔗 Execute Workflow Trigger | Accepts input from parent workflow | | 🧠 Simple Memory | Stores session + symbol info | | 🤖 Binance SM Market Agent | Parses prompt, routes tool calls | | 💡 OpenAI Chat Model (gpt-4o-mini) | Converts raw data into a clean, readable format for Telegram | | 🌐 getCurrentPrice | Gets latest price | | 🌐 get24hrStats | Gets OHLC/volume over past 24 hours | | 🌐 getOrderBook | Gets top 100 bids and asks | | 🌐 getKlines | Gets latest 15m, 1h, 4h, and 1d candles | 📥 Input Requirements This workflow is not called directly by the user. Instead, it is triggered by another workflow, such as: { "message": "BTCUSDT", "sessionId": "539847013" } 📤 Telegram Output Example 📊 BTCUSDT Market Overview 💰 Price: $63,220 📈 24h Change: +2.3% | Volume: 45,210 BTC 📉 Order Book • Top Bid: $63,190 • Top Ask: $63,230 🕰️ Latest Candles • 15m: O: $63,000 | C: $63,220 | Vol: 320 BTC • 1h : O: $62,700 | C: $63,300 | Vol: 980 BTC • 4h : O: $61,800 | C: $63,500 | Vol: 2,410 BTC • 1d : O: $59,200 | C: $63,220 | Vol: 7,850 BTC ✅ Use Cases | Scenario | Output Provided | | ---------------------------------- | ------------------------------------------------------------ | | “Show current BTC price and trend” | Price, 24h stats, candles, and order book in one message | | “Candles for SOL” | 15m, 1h, 4h, 1d candlesticks for SOLUSDT | | Triggered by Quant AI system | Clean Telegram-ready summary with all structure tools merged | 🧩 Toolchain Breakdown | Tool Name | Endpoint | Purpose | | ----------------- | ---------------------- | ------------------------------ | | getCurrentPrice | /api/v3/ticker/price | Latest trade price | | get24hrStats | /api/v3/ticker/24hr | 24h OHLC, % change, volume | | getOrderBook | /api/v3/depth | Top 100 bids and asks | | getKlines | /api/v3/klines | 1-candle snapshot across 4 TFs | 🚀 Installation Steps Import the JSON into your n8n instance Connect your OpenAI credentials for the Chat Model node No Binance API key needed — public endpoints Trigger this tool only via: Binance SM Financial Analyst Tool Binance Spot Market Quant AI Agent 🔐 Licensing & Attribution © 2025 Treasurium Capital Limited Company Architecture, prompts, and trade structure are IP-protected. No unauthorized rebranding permitted. 🔗 For support: Don Jayamaha – LinkedIn
by Angel Menendez
CallForge - AI Sales Call Processing & Insights Extraction Automate sales call analysis with AI-powered insights for sales, marketing, and product teams. Who is This For? This workflow is designed for: ✅ Sales teams looking to extract structured insights from Gong call transcripts. ✅ Marketing professionals seeking AI-driven customer pain points & content strategy. ✅ Product teams needing feedback from sales calls to prioritize feature development. 🔍 What Problem Does This Workflow Solve? Manually analyzing Gong.io sales call transcripts is slow, inconsistent, and lacks structured insights. With CallForge, you can: ✔ Extract AI-powered insights about use cases, objections, competitors, and next steps. ✔ Provide structured marketing & product intelligence to enhance strategy. ✔ Automatically store call insights in Notion and Salesforce for easy access. ✔ Ensure resilience with automated reruns on failed workflows (handling Notion API limits). ✔ Improve decision-making with AI-powered competitor and sentiment analysis. 📌 Key Workflow Features 🎤 AI-Powered Transcript Analysis Uses AI to identify use cases, objections, competitors, and customer pain points. Categorizes insights for sales, marketing, and product teams. 📌 AI Agent Breakdown 🔹 Sales AI Agent – Extracts customer objections, pain points, competitors, and next steps. 🔹 Marketing AI Agent – Identifies recurring topics, keyword trends, and content opportunities. 🔹 Product AI Agent – Captures feature requests and AI/ML-related references. 📊 Structured Output Processing Sales Data Processor* → Stores insights in *Notion & Salesforce** for sales tracking. Marketing Data Processor* → Extracts *SEO & content strategy insights** for marketing teams. Product AI Data Processor* → Logs *customer feedback* to prioritize *feature development**. 💡 Competitor & Integration Analysis Tracks competing products mentioned in calls**. Identifies integration needs**, flagging workarounds used by prospects. 📢 Real-Time Slack Notifications Alerts teams on workflow progress** and completed call analyses. 🔄 Failure Resilience & Automated Re-Runs If a Notion API limit is reached, the process resumes automatically. 🚀 How This Works 🛠 1. Trigger & Call Data Processing The workflow retrieves Gong call transcripts and metadata. Normalizes data**, correcting common mispronunciations like "n8n." 🤖 2. AI Agents Analyze the Call Sales Agent** – Extracts actionable insights for sales follow-ups. Marketing Agent* – Identifies *recurring themes* and *keyword trends**. Product Agent* – Captures *feature requests and AI/ML usage mentions**. 📡 3. Data is Stored in Notion & Salesforce Logs AI-extracted insights* in *Notion** for structured tracking. Pushes sales-related data* to *Salesforce** for team accessibility. 🔔 4. Slack Alerts for Teams Notifies sales, marketing, and product teams** about extracted insights. CallForge - 01 - Filter Gong Calls Synced to Salesforce by Opportunity Stage CallForge - 02 - Prep Gong Calls with Sheets & Notion for AI Summarization CallForge - 03 - Gong Transcript Processor and Salesforce Enricher CallForge - 04 - AI Workflow for Gong.io Sales Calls CallForge - 05 - Gong.io Call Analysis with Azure AI & CRM Sync CallForge - 06 - Automate Sales Insights with Gong.io, Notion & AI CallForge - 07 - AI Marketing Data Processing with Gong & Notion CallForge - 08 - AI Product Insights from Sales Calls with Notion 📊 Sample Output Data 1️⃣ Sales Insights { "UseCases": [ { "Summary": "A manufacturing company wants to automate inventory tracking and reduce manual entry delays.", "DepartmentTags": ["Operations"], "IndustryTags": ["Manufacturing"], "ImplementationStatus": "Evaluating" } ], "Objection": { "ObjectionTags": ["Feature Limitation"], "Nature": "The prospect wanted a deeper integration with their ERP system, which n8n currently lacks." }, "CallSummary": "The call focused on automation for supply chain processes. The prospect expressed interest but wanted confirmation on ERP integration capabilities.", "NextSteps": ["Schedule a follow-up demo for ERP integration."] } 2️⃣ Marketing Insights { "MarketingInsights": [ { "Tag": "Workflow Template Request", "Summary": "The prospect requested a template for automating CRM lead tracking." } ], "RecurringTopics": [ { "Topic": "CRM Integration", "Mentions": 3, "Context": "Discussed how n8n could sync CRM data automatically." } ], "ActionableInsights": [ { "RecommendationType": "Tutorial", "Title": "How to Automate CRM Lead Tracking with n8n", "Topic": "CRM Integration", "Rationale": "The prospect expressed a need for CRM automation templates." } ] } 3️⃣ Product Feedback { "ProductFeedback": [ { "Sentiment": "Positive", "Feedback": "The external speaker praised the simplicity of n8n's UI, making it easier for non-developers to automate tasks." }, { "Sentiment": "Negative", "Feedback": "The external speaker mentioned frustration over the lack of a dedicated ERP integration node." } ], "AI_ML_References": { "Exist": true, "Context": "The external speaker mentioned using AI for automating customer ticket categorization.", "Details": { "DevelopmentStatus": "Building", "Department": "Support", "RequiresAgents": true, "RequiresRAG": false, "RequiresChat": "Yes: External App (e.g., Slack)" } } } 🔧 How to Customize This Workflow 💡 🔗 Change Data Storage – Swap Notion for Airtable, HubSpot, or another CRM. 💡 📩 Customize Slack Notifications – Send alerts via email, webhook, or another channel. 💡 🛠 Modify AI Processing – Adjust AI models or processing prompts. 💡 📊 Add More Integrations – Sync insights with Pipedrive, HubSpot, or another CRM. 🚀 Why Use This Workflow? ✔ Automates Gong call transcript analysis, eliminating manual work. ✔ Improves collaboration by structuring insights for sales, marketing, and product teams. ✔ Boosts sales conversions by identifying objections and next steps. ✔ Enhances marketing and SEO strategy with AI-driven insights. ✔ Optimizes product roadmap decisions based on customer feedback. This workflow scales AI-powered sales intelligence for better decision-making, content strategy, and sales enablement. 🚀
by Nskha
Overview The [n8n] YouTube Channel Advanced RSS Feeds Generator workflow facilitates the generation of various RSS feed formats for YouTube channels without requiring API access or administrative permissions. It utilizes third-party services to extract data, making it extremely user-friendly and accessible. Key Use Cases and Benefits Content Aggregation**: Easily gather and syndicate content from any public YouTube channel. No API Key Required**: Avoid the complexities and limitations of Google's API. Multiple Formats**: Supports ATOM, JSON, MRSS, Plaintext, Sfeed, and direct YouTube XML feeds. Flexibility**: Input can be a YouTube channel or video URL, ID, or username. Services/APIs Utilized This workflow integrates with: commentpicker.com**: For retrieving YouTube channel IDs. rss-bridge.org**: To generate various RSS formats. Configuration Instructions Start the Workflow: Activate the workflow in your n8n instance. Input Details: Enter the YouTube channel or video URL, ID, or username via the provided form trigger. Run the Workflow: Execute the workflow to receive links to 13 different RSS feeds, including community and video content feeds. Screenshots Additional Notes Customization**: You can modify the RSS feed formats or integrate additional services as needed. Support and Contributions For support, questions, or contributions, please visit the n8n community forum or the GitHub repository. We welcome contributions from the community!