by Jimleuk
This n8n template demonstrates the beginnings of building your own n8n-powered WhatsApp chatbot! Under the hood, utilise n8n's powerful AI features to handle different message types and use an AI agent to respond to the user. A powerful tool for any use-case! How it works Incoming WhatsApp Trigger provides a way to get messages into the workflow. The message received is extracted and sent through 1 of 4 branches for processing. Each processing branch uses AI to analyse, summarize or transcribe the message so that the AI agent can understand it. The supported types are text, image, audio (voice notes) and video. The AI Agent is used to generate a response generally and uses a wikipedia tool for more complex queries. Finally, the response message is sent back to the WhatsApp user using the WhatsApp node. How to use Once you have setup and configured your WhatsApp account, you'll need to activate your workflow to start processing messages. Good to know: Large media files may negatively impact workflow performance. Requirements WhatsApp Buisness account Google Gemini for LLM. Gemini is used specifically because it can accept audio and video files whereas at time of writing, many other providers like OpenAI's GPT, do not. Customising this workflow For performance reasons, consider detecting large audio and video before sending to the LLM. Pre-processing such files may allow your agent to perform better. Go beyond and create rich and engagement customer experiences by responding using images, audio and video instead of just text!
by Elegant Biztech
Automated QuickBooks Invoice to Custom PDF & Email Tired of the standard, boring invoices from QuickBooks Online? This workflow completely automates the process of creating beautiful, custom-branded PDF invoices and emailing them directly to your clients, saving you time and elevating your brand's professionalism. The moment you create an invoice in QuickBooks, this workflow triggers, fetches all the necessary data, and generates a lavish, multi-page-aware PDF invoice complete with your company logo and signature. Key Features Fully Automated:** Runs instantly when a new invoice is created in QuickBooks. Custom Branding:** Automatically fetches your company logo and signature from a URL to place on the invoice. Modern & Professional Design:** Uses a premium, multi-column HTML template that is clean, easy to read, and far superior to the default QBO templates. Multi-Page Ready:** If an invoice has many line items, the template will intelligently create multiple pages and add a "Page X of Y" footer automatically. Smart Layout:** The totals and summary block are designed to never break across pages, ensuring a professional look no matter the length. Automatic Emailing:** The final PDF is attached to a beautifully formatted email and sent directly to the customer's email address on file. Prerequisites Before you start, you will need a few things: A running n8n instance. A QuickBooks Online account with API access. A running Gotenberg instance. This is a powerful, open-source tool for converting HTML to PDF. This workflow is designed to connect to its API. You can learn more about it here. Publicly accessible URLs for your company logo and signature image (e.g., hosted on your website or a service like Imgur). Setup Guide Follow these steps carefully to configure the workflow for your own use. Nodes that need your attention are marked with a [!!] prefix. Step 1: Configure the QuickBooks Webhook The workflow starts with a webhook. You need to tell QuickBooks to send information to this webhook. Open the [!!] Listen for New QuickBooks Invoice node. You will see a Webhook URL. Copy the Production URL. Go to your QuickBooks Developer dashboard, select your app, and navigate to the Webhooks section. Paste the n8n URL into the Endpoint URL field and select the Invoice event to subscribe to. Step 2: Connect Your QuickBooks Account Open the [!!] Get Invoice Data from QuickBooks node. In the "Credentials" field, select your existing QuickBooks Online credentials or create a new set. Step 3: Add Your Branding Open the [!!] Fetch Company Logo Image node. In the URL field, replace the placeholder with the public URL of your company's logo. Open the [!!] Fetch Company Signature Image node. In the URL field, replace the placeholder with the public URL of your signature image. Step 4: Update the PDF Generation Service Open the [!!] Generate PDF via Gotenberg node. In the URL field, replace the placeholder http://YourGotenBergInstanceURL/... with the real URL of your running Gotenberg instance. Step 5: Configure Your Email Open the [!!] Email PDF Invoice to Customer node. In the "Credentials" field, select your SMTP or email service credentials. Customize the From Email and Subject fields. You can also edit the beautiful HTML email body to match your company's tone of voice. Step 6: Activate Your Workflow You're all set! Save the workflow and activate it using the toggle at the top-right of the screen. Now, when you create a new invoice in QuickBooks, this automation will handle the rest. A Note from the Creator Thank you for using this workflow! I believe that professional and automated invoicing is a cornerstone of a great business. This tool was designed to save you time and help you put your best foot forward with every client interaction. If you have any questions or need assistance, feel free to reach out. Website:** https://www.elegantbiztech.com/ Email:** sales@elegantbiztech.com
by Cyril Nicko Gaspar
📌 AI Agent Template with Bright Data MCP Tool Integration This template enables natural-language-driven automation using Bright Data MCP tools. It extracts all available tools from MCP, processes the user’s query through an AI agent, then dynamically selects and executes the appropriate tool. ❓ Problem It Solves Traditional automation often requires users to understand APIs, interfaces, or scripts to perform backend tasks. The Bright Data MCP integration solves this by allowing natural language interaction, intelligently classifying user intent, and managing context-aware execution of complex operations—ideal for data extraction, customer support, and workflow orchestration. 🧰 Pre-requisites Before deploying this template, make sure you have: An active N8N instance (self-hosted or cloud). A valid OpenRouter API key (or another compatible AI model). Telegram bot and its API token Access to the Bright Data MCP API with credentials. Basic familiarity with N8N workflows and nodes. ⚙️ Setup Instructions Setup and obtain API token and other necessary information from Bright Data In your Bright Data account, obtain the following information: API token Web Unlocker zone name (optional) Browser Zone name (optional) Host SSE server from STDIO command The methods below will allow you to receive SSE (Server-Sent Events) from Bright Data MCP via a local Supergateway or Smithery ** Method 1: Run Supergateway in a separate web service (Recommended) This method will work for both cloud version and self-hosted N8N. Signup to any cloud services of your choice (DigitalOcean, Heroku, Hetzner, Render, etc.). For NPM based installation: Create a new web service. Choose Node.js as runtime environment and setup a custom server without repository. In your server’s settings to define environment variables or .env file, add: `API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_ZONE=optional_browser_zone_name` Paste the following text as a start command: npx -y supergateway --stdio "npx -y @brightdata/mcp" --port 8000 --baseUrl http://localhost:8000 --ssePath /sse --messagePath /message Deploy it and copy the web server URL, then append /sse into it. Your SSE server should now be accessible at: https://your_server_url/sse For Docker based installation: Create a new web service. Choose Docker as the runtime environment. Set up your Docker environment by pulling the necessary images or creating a custom Dockerfile. In your server’s settings to define environment variables or .env file, add: `API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_AUTH=optional_browser_auth` Use the following Docker command to run Supergateway: `docker run -it --rm -p 8000:8000 supercorp/supergateway \ --stdio "npx -y @brightdata/mcp /" \ --port 8000` Deploy it and copy the web server URL, then append /sse into it. Your SSE server should now be accessible at: https://your_server_url/sse For more installation guides, please refer to https://github.com/supercorp-ai/supergateway.git. ** Method 2: Run Supergateway in the same web service as the N8N instance This method will only work for self-hosted N8N. a. Set Required Environment Variables In your server's settings to define environment variables or .env file, add: API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_ZONE=optional_browser_zone_name b. Run Supergateway in Background npx -y supergateway --stdio "npx -y @brightdata/mcp" --port 8000 --baseUrl http://localhost:8000 --ssePath /sse --messagePath /message Use the command above to execute it through the cloud shell or set it as a pre-deploy command. Your SSE server should now be accessible at: http://localhost:8000/sse For more installation guides, please refer to https://github.com/supercorp-ai/supergateway.git. * *Method 3: Configure via Smithery.ai* (Easiest) If you don't want additional setup and want to test it right away, follow these instructions: Visit https://smithery.ai/server/@luminati-io/brightdata-mcp/tools to: Signup (if you are new to Smithery) Create an API key Define environment variables via a profile Retrieve your SSE server HTTP URL Import the Workflow Open N8N. Import the JSON workflow file included with this template. Update any nodes referencing external services (e.g., OpenRouter, Telegram). Setup Telegram Integration If you haven't setup a bot in Telegram, below is the instruction how to create one using BotFather: Search for @BotFather in Telegram and start a conversation with it. Send the command /newbot to create a new bot. You'll be prompted to enter a name and a unique username for your bot. BotFather will provide you with an access token, which you'll need to use to interact with the bot's API. Edit the HTTP Request node in the workflow. Configure the URL as follows: https://api.telegram.org/bot+your_telegram_bot_token+/setWebhook?url=+your_webhook_url Replace +your_telegram_bot_token+ with your actual Telegram bot token. Replace +your_webhook_url+ with the URL from the Webhook Trigger node in the workflow. This will set up Telegram to forward messages to your n8n agent. 🔄 Workflow Functionality (Summary) The user submits a message via chat. Memory** nodes retain context for multi-turn conversations. The mapped tool is executed and results are returned contextually. 🧠 Optional memory buffers and memory manager nodes keep the interaction context-aware. 🧩 Use Cases Data Scraping on Demand**: Launch scraping tasks via chat. Lead Generation Bots**: Enrich or validate leads with MCP tools. AI-Powered Customer Support**: Classify and answer queries with real-time data tools. Workflow Assistants**: Let teams run backend processes like lookups or report generation using plain language. 🛠️ Customization Classifier Prompt & Logic**: Tweak the AI’s prompt and tool-matching schema to better fit your use case. Memory Configuration**: Adjust retention policies and context depth. Tool Execution Sub-Workflow**: Extend for retries, logging, or chaining actions. Omni-Channel Support**: Connect via webhooks to chat interfaces like Slack, WhatsApp, Telegram, or custom UIs. ✅ Summary This template equips you with a powerful no-code/low-code AI agent that translates conversation into real-world action. Using Bright Data’s MCP tools through natural language, it enables teams to automate and scale data-driven tasks effortlessly.
by Lucía Maio Brioso
🧑💼 Who is this for? This workflow is for anyone with two YouTube channels who wants to copy playlists from one to the other — no technical skills required. Whether you're a content creator, hobbyist, educator, or just someone managing multiple channels, this workflow helps you save time and avoid the manual work of recreating playlists video by video. 🧠 What problem is this workflow solving? YouTube doesn't provide an option to transfer or duplicate playlists between accounts or channels. That means if you want the same playlists in two places, you're stuck: Creating new playlists manually Searching for each video again Copy-pasting links one by one This workflow automates the entire process for you — accurately, quickly, and with no manual work. ⚙️ What this workflow does Retrieves all playlists from a source YouTube channel (excluding private ones) For each playlist: Gets all its videos Filters out private or unavailable videos Creates a new playlist in the target channel with the same title Adds the videos to the new playlist Continues smoothly even if some videos fail to copy (e.g., if they’re restricted or deleted) 🛠️ Setup Create two YouTube OAuth2 credentials in n8n: One for your source channel One for your target channel Assign the credentials to the correct nodes as indicated in the sticky notes: Source nodes → source credentials Target nodes → target credentials Click “Test workflow” to run it. > ⚠️ Note: If you have many playlists or videos, you may hit YouTube’s API quota. You can request a quota increase in your Google Cloud Console if needed. 🧩 How to customize this workflow to your needs ✂️ Copy only specific playlists Use a Filter node after the playlist fetch to include only certain titles or IDs. 📝 Change the title of the copied playlists Modify the title in the Create playlist node (e.g., add “(Copy)” or a prefix). 🔄 Automate it regularly Replace the Manual Trigger with a Cron node if you want to run this periodically. 🧪 Test safely If you're unsure, use a secondary channel as your test target before applying changes to your main account.
by Nasser
For Who? Content Creators Youtube Automation Marketing Team How it works? 1 - Every week, retrieve the keywords you want to track 2 - Thanks to Apify, scrape videos from YouTube Search related to these keywords, filtered by relevance 3 - Wait until the dataset is completed 4 - Get the information contained in the dataset 5 - For each video, clean and summarize the script 6 - Upload everything to your Airtable database 📺 YouTube Video Tutorial: Setup (~5min) Scheduled Trigger: Select the frequency you want. If you change it, update the data accordingly in the "Create Videos Dataset" HTTP Request node in Body ➡️ JSON ➡️ dateFilter. Setup Keywords: Enter keywords related to the niche you want. If you change the number of keywords, update the data accordingly in the "Create Videos Dataset" HTTP Request node in Body ➡️ JSON ➡️ searchQueries. Create Videos Dataset: Refer to the Apify documentation for more: https://docs.apify.com/api/v2/getting-started APIs: For all HTTP Request nodes in the URL field, replace [YOUR_API_TOKEN] with your API token. 👨💻 More Workflows : https://n8n.io/creators/nasser/
by MattF
This workflow generates a weekly performance summary from Google Search Console, focused on brand-level SEO metrics and week-over-week trends. It provides a structured view of how each brand segment is performing, with clean formatting for quick insights. Key Features Sends a weekly email with a table showing clicks, impressions, CTR, and position — along with % change vs. the previous week. Highlights both brand and non-brand clicks separately. Color-coded % changes make it easy to spot wins (green) and losses (red) at a glance. It’s designed to give SEO teams a consistent overview of performance by brand, helping to track directional shifts and support deeper analysis when needed. How it works Runs weekly (e.g. every Monday) to compare “Last Week” vs. “2 Weeks Ago” from GSC data. Includes both brand + non-brand click breakdown. Calculates raw values and week-over-week % change for clicks, impressions, CTR, and position. Outputs a clean, formatted table with labeled rows and color-coded changes. Sends the table as part of a scheduled email (can also be adapted for Slack or other channels). Setup steps Requires connected Google Search Console data (per brand segment). Email delivery is included by default (customizable to other platforms). Update brand segmentation logic to match your tracking needs (e.g. domain, label, or custom filters). Typical setup time: ~5-10 minutes with structured input data.
by Abhishek Patoliya
This n8n automation lets you build a complete AI-powered task management system that integrates Telegram, Google Sheets, and GPT-4o mini to help users easily manage to-do lists and receive daily task reminders. Users can interact with the system via Telegram, while the AI assistant (powered by GPT-4o mini) processes commands naturally, updates a central Google Sheet, and ensures scheduled reminders are sent for pending tasks. ✨ Key Features ✅ Add, list, update, complete, or delete tasks via Telegram ✅ AI-powered conversational responses using GPT-4o ✅ All tasks stored and synced in Google Sheets ✅ Daily scheduled task summary and pending reminders sent to Telegram ✅ Friendly, human-like assistant responses ✅ Fully configurable and easy to set up 🛠️ Workflow Functionality Breakdown 1. User Interacts on Telegram Sends commands like: add buy groceries list tasks complete submit report delete dentist appointment 2. AI-Powered Processing A GPT-4o agent processes user messages Ensures clear, friendly responses Determines task intent: add, update, delete, list, complete 3. Google Sheets Sync Every operation is logged to Google Sheets Google Sheets acts as the master task database Sheet structure includes: Task Status (pending or done) Created At (timestamp) Due Date (optional) Notes (optional) 4. Scheduled Daily Task Summary At 9 PM daily, the workflow: Fetches tasks from Google Sheets Generates a warm, conversational summary via GPT-4o Sends the summary to the user on Telegram 5. Automated Reminders Checks for pending tasks due today Sends reminder messages to Telegram ✅ Prerequisites Before setting up the workflow, ensure you have: ✔️ An n8n instance (Cloud or self-hosted) ✔️ A Telegram Bot Token ✔️ Access to Google Sheets API (OAuth2 credentials) ✔️ An OpenAI API Key with GPT-4o access ✔️ A Google Sheet structured as per the specification below 📝 Google Sheet Structure Your Google Sheet should have the following columns: | Column Name | Description | | -------------- | ---------------------------------------------------- | | Task | Short task description | | Status | pending or done | | Created At | Date & time task was created (YYYY-MM-DD HH:mm:ss) | | Due Date | (Optional) When task is due (YYYY-MM-DD HH:mm:ss) | | Notes | (Optional) Additional details | Important: The first row should be the header row with these exact column names. 🔧 Setup Instructions 1. Telegram Bot Setup Create a bot via BotFather Obtain the Bot Token Connect Telegram Trigger and Telegram Send nodes using your Bot Token in n8n 2. Google Sheets API Setup Follow n8n Google Sheets integration guide Set up OAuth2 credentials Provide access to your task Google Sheet 3. OpenAI API Setup Obtain an API key from OpenAI Ensure GPT-4o mini access is enabled Add OpenAI credentials to relevant nodes 4. Sheet Linking Replace the Google Sheet ID in the workflow with your own Confirm sheet names and column structure match exactly 5. Schedule Configuration (Optional) Adjust the daily summary time (Schedule Trigger node) as desired ⚙️ Configuration Options 🔧 Adjust AI prompt instructions for tone/style 🔧 Change reminder times in the schedule trigger 🔧 Customize Google Sheet columns if needed (update mappings accordingly) 🔧 Add multi-user support with chat ID checks (advanced) 📂 Files Included Full n8n JSON workflow ready to import 💡 Tips You can extend this with Slack, WhatsApp, or Email reminders Combine with Notion, ClickUp, or CRM integrations for more powerful task management Consider adding a "Priority" column for advanced sorting Ready to stay organized with AI-powered task management? Import this workflow, link your accounts, and your Telegram assistant is good to go! 🚀
by Darien Kindlund
If you have multiple users managing workflows, there may come a time where a user “accidentally” turns off a workflow. Or, if you have workflows that automatically turn off other workflows, that code might “accidentally” turn off the wrong one. In either case, here’s a workflow that can attempt to “auto-start” accidentally disabled workflows: How it works: When activated, then every 4 hours, the workflow will search all other workflows that have the auto_resume:true tag present. If any other workflow has auto_resume:true set but is currently turned off, then this workflow will turn it back on. Of course, this watchdog won’t work if the watchdog workflow is turned off. That said, we’ve found this useful in recovering from accidental actions that cause production workflows to be turned off.
by Ranjan Dailata
Who this is for The Crunchbase B2B Lead Discovery Pipeline is designed for sales teams, B2B marketers, business analysts, and data operations teams who need a reliable way to extract, structure, and summarize company information from Crunchbase to fuel lead generation and market intelligence. This workflow is ideal for: Sales Development Reps (SDRs) - Needing structured leads from Crunchbase Marketing Analysts - Generating segmented outreach lists Growth Teams - Identifying trending B2B startups RevOps Teams - Automating company research pipelines Data Teams - Consolidating insights into Google Sheets for dashboards What problem is this workflow solving? Manual extraction of company data from Crunchbase is time-consuming, inconsistent, and often lacks the contextual summary required for sales enablement or growth targeting. This workflow automates the extraction, transformation, summarization, and delivery of Crunchbase company data into structured formats, making it instantly usable for B2B targeting and analysis. It solves: The difficulty of scaling lead discovery from Crunchbase The need to summarize raw textual content for quick insights The lack of integration between web scraping, LLM processing, and storage What this workflow does Markdown to Textual Data Extractor**: Takes raw scraped markdown from Crunchbase and converts it into readable plain text using a basic LLM chain Structured Data Extraction**: Applies a parsing model (OpenAI) to extract structured fields such as company name, funding rounds, industry tags, location, and founding year Summarization Chain**: Generates an executive summary from the raw Crunchbase text using a summarization prompt template Send to Google Sheets**: Adds the structured data and summary into a Google Sheet for team access and further processing Persist to Disk**: Saves both raw and structured data files locally for archiving or further use Webhook Notification**: Sends a structured payload to a webhook endpoint (e.g., Slack, CRM, internal tools) with lead insights Pre-conditions You need to have a Bright Data account and do the necessary setup as mentioned in the "Setup" section below. You need to have an OpenAI Account. Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. In n8n, Configure the Google Sheet Credentials with your own account. Follow this documentation - Set Google Sheet Credential In n8n, configure the OpenAi account credentials. Ensure the URL and Bright Data zone name are correctly set in the Set URL, Filename and Bright Data Zone node. Set the desired local path in the Write a file to disk node to save the responses. How to customize this workflow to your needs LLM Prompt Customization : Modify the extraction prompt to include additional fields like revenue, social links, leadership team Adjust summarization tone (e.g., executive summary, sales-focused snapshot or marketing digest) File Persistence Store raw markdown, extracted JSON, and summary text separately for audit/debug Webhook Notification Connect to CRM (e.g., HubSpot, Salesforce) via webhook to automatically create leads Send Slack notifications to alert sales reps when a new high-potential company is discovered
by Trung Tran
🎙️ VoiceScribe AI: Telegram Audio Message Auto Transcription with OpenAI Whisper > Automatically transcribe Telegram voice messages and store them as structured logs in Google Sheets, while backing up the audio in Google Drive. 🧑💼 Who’s it for Journalists, content creators, or busy professionals who often record voice memos or short interviews on the go. Anyone who wants to turn voice recordings into searchable, structured notes. ⚙️ How it works / What it does User sends a voice message to a Telegram bot. n8n checks if the message is an audio voice note. If valid, it downloads the audio file and: Transcribes it using OpenAI Whisper (or your LLM of choice). Uploads the original audio to Google Drive for safekeeping. The transcript and audio metadata are merged. The workflow: Logs the data into a Google Sheet. Sends a formatted confirmation message to the user via Telegram. If the input is not audio, the bot politely informs the user that only voice messages are accepted. ✅ Features Accepts only Telegram voice messages. Transcribes via OpenAI Whisper. Logs DateTime, Duration, Transcript, and Audio URL to Google Sheets. Sends user feedback message via Telegram with download + transcript link. 🚀 How to set up Prerequisites Telegram Bot connected to n8n (via Telegram Trigger) Google Drive & Google Sheets credentials configured OpenAI or Whisper API credentials (for transcription) Steps Telegram Trigger Start the flow when a new message is sent to your bot. Check Message Type Use a conditional node to confirm it's a voice message. Download Voice Message Download the .oga file from Telegram. Transcribe Audio Send the binary audio to OpenAI Whisper or your transcription service. Upload to Google Drive Backup the original audio file. Merge Outputs Combine transcription with Drive metadata. Transform to Row Format Prepare structured JSON for Google Sheets. Append to Google Sheet Store the transcript log (DateTime, Duration, Transcript, AudioURL). Send Confirmation to User Inform the user via Telegram with their transcript and download link. Unsupported Message Handler Reply to users who send non-audio messages. 📄 Example Output in Google Sheet | DateTime | Duration | Transcript | AudioURL | |-----------------------|----------|--------------------------------------------|------------------------------------------------------------| | 2025-08-07T13:12:19Z | 27 | Dự án Outlet Activation là... | https://drive.google.com/uc?id=xxxx&export=download | 🧠 How to customize the workflow Swap Whisper with Deepgram, AssemblyAI, or other providers. Add speaker name detection or prompt-based tagging via GPT. Route transcripts into Notion, Airtable, or CRM systems. Add multi-language support or summarization steps. 📦 Requirements | Component | Required | |---------------------|----------| | Telegram API | ✅ | | Google Drive API | ✅ | | Google Sheets API | ✅ | | OpenAI Whisper API | ✅ | | n8n Cloud or Self-hosted | ✅ | Created with ❤️ using n8n
by vinci-king-01
Smart Supplier Health Monitor with ScrapeGraphAI Risk Detection and Multi-Channel Alerts 🎯 Target Audience Procurement managers and directors Supply chain risk analysts CFOs and financial controllers Vendor management teams Enterprise risk managers Operations managers Contract administrators Business continuity planners 🚀 Problem Statement Manual supplier monitoring is reactive and time-consuming, often missing early warning signs of financial distress that could disrupt your supply chain. This template solves the challenge of proactive supplier health surveillance by automatically monitoring financial indicators, news sentiment, and market conditions to predict supplier risks before they impact your business operations. 🔧 How it Works This workflow automatically monitors your critical suppliers' financial health using AI-powered web scraping, analyzes multiple risk factors, identifies alternative suppliers when needed, and sends intelligent alerts through multiple channels to ensure your procurement team can act quickly on emerging risks. Key Components Weekly Health Check Scheduler - Automated trigger based on supplier criticality levels Supplier Database Loader - Dynamic supplier portfolio management with risk-based monitoring frequency ScrapeGraphAI Website Analyzer - AI-powered extraction of financial health indicators from company websites Financial News Scraper - Intelligent monitoring of financial news and sentiment analysis Advanced Risk Scorer - Industry-adjusted risk calculation with failure probability modeling Alternative Supplier Finder - Automated identification and ranking of backup suppliers Multi-Channel Alert System - Email, Slack, and API notifications with escalation rules 📊 Risk Analysis Specifications The template performs comprehensive financial health analysis with the following parameters: | Risk Factor | Weight | Score Impact | Description | |-------------|--------|--------------|-------------| | Financial Issues | 40% | +0-24 points | Revenue decline, debt levels, cash flow problems | | Operational Risks | 30% | +0-18 points | Management changes, restructuring, capacity issues | | Market Risks | 20% | +0-12 points | Industry disruption, regulatory changes, competition | | Reputational Risks | 10% | +0-6 points | Negative news, legal issues, public sentiment | Industry Risk Multipliers: Technology: 1.1x (Higher volatility) Manufacturing: 1.0x (Baseline) Energy: 1.2x (Regulatory risks) Financial: 1.3x (Market sensitivity) Logistics: 0.9x (Generally stable) Risk Levels & Actions: Critical Risk**: Score ≥ 75 (CEO/CFO escalation, immediate transition planning) High Risk**: Score ≥ 55 (Procurement director escalation, backup activation) Medium Risk**: Score ≥ 35 (Manager review, increased monitoring) Low Risk**: Score < 35 (Standard monitoring) 🏢 Supplier Management Features | Feature | Critical Suppliers | High Priority | Medium Priority | |---------|-------------------|---------------|-----------------| | Monitoring Frequency | Weekly | Bi-weekly | Monthly | | Risk Threshold | 35+ points | 40+ points | 50+ points | | Alert Recipients | C-Level + Directors | Directors + Managers | Managers only | | Alternative Suppliers | 3+ pre-qualified | 2+ identified | 1+ researched | | Transition Timeline | 24-48 hours | 1-2 weeks | 1-3 months | 🛠️ Setup Instructions Estimated setup time: 25-30 minutes Prerequisites n8n instance with community nodes enabled ScrapeGraphAI API account and credentials Gmail account for email alerts (or alternative email service) Slack workspace with webhook or bot token Supplier database or CRM system API access Basic understanding of procurement processes Step-by-Step Configuration 1. Configure ScrapeGraphAI Credentials Sign up for ScrapeGraphAI API account Navigate to Credentials in your n8n instance Add new ScrapeGraphAI API credentials with your API key Test the connection to ensure proper functionality 2. Set up Email Integration Add Gmail OAuth2 credentials in n8n Configure sender email and authentication Test email delivery with sample message Set up email templates for different risk levels 3. Configure Slack Integration Create Slack webhook URL or bot token Add Slack credentials to n8n Configure target channels for different alert types Customize Slack message formatting and buttons 4. Load Supplier Database Update the "Supplier Database Loader" node with your supplier data Configure supplier categories, contract values, and criticality levels Set monitoring frequencies based on supplier importance Add supplier website URLs and contact information 5. Customize Risk Parameters Adjust industry risk multipliers for your business context Modify risk scoring thresholds based on risk tolerance Configure economic factor adjustments Set failure probability calculation parameters 6. Configure Alternative Supplier Database Populate the alternative supplier database in the "Alternative Supplier Finder" node Add supplier ratings, capacities, and specialties Configure geographic coverage and certification requirements Set suitability scoring parameters 7. Set up Procurement System Integration Configure the procurement system webhook endpoint Add API authentication credentials Test webhook payload delivery Set up automated data synchronization 8. Test and Validate Run test scenarios with sample supplier data Verify ScrapeGraphAI extraction accuracy Check risk scoring calculations and thresholds Confirm all alert channels are working properly Test alternative supplier recommendations 🔄 Workflow Customization Options Modify Risk Analysis Add custom risk indicators specific to your industry Implement sector-specific economic adjustments Configure contract-specific risk factors Add ESG (Environmental, Social, Governance) scoring Extend Data Sources Integrate credit rating agency APIs (Dun & Bradstreet, Experian) Add financial database connections (Bloomberg, Reuters) Include social media sentiment analysis Connect to government regulatory databases Enhance Alternative Supplier Management Add automated supplier qualification workflows Implement dynamic pricing comparison Create supplier performance scorecards Add geographic risk assessment Advanced Analytics Implement predictive failure modeling Add supplier portfolio optimization Create supply chain risk heatmaps Generate automated compliance reports 📈 Use Cases Supply Chain Risk Management**: Proactive monitoring of supplier financial stability Procurement Optimization**: Data-driven supplier selection and management Business Continuity Planning**: Automated backup supplier identification Financial Risk Assessment**: Early warning system for supplier defaults Contract Management**: Risk-based contract renewal and negotiation Vendor Diversification**: Strategic supplier portfolio management 🚨 Important Notes Respect ScrapeGraphAI API rate limits and terms of service Implement appropriate delays between supplier assessments Keep all API credentials secure and rotate them regularly Monitor API usage to manage costs effectively Ensure compliance with data privacy regulations (GDPR, CCPA) Regularly update supplier databases and contact information Review and adjust risk parameters based on market conditions Maintain confidentiality of supplier financial information 🔧 Troubleshooting Common Issues: ScrapeGraphAI extraction errors: Check API key validity and rate limits Email delivery failures: Verify Gmail credentials and permissions Slack notification failures: Check webhook URL and channel permissions False positive alerts: Adjust risk scoring thresholds and industry multipliers Missing supplier data: Verify website URLs and accessibility Alternative supplier errors: Check supplier database completeness Monitoring Best Practices: Set up workflow execution monitoring and error alerts Regularly review and update supplier information Monitor API usage and costs across all integrations Validate risk scoring accuracy with historical data Test disaster recovery and backup procedures Support Resources: ScrapeGraphAI documentation and API reference n8n community forums for workflow assistance Procurement best practices and industry standards Financial risk assessment methodologies Supply chain management resources and tools
by Nima Salimi
Description🔍 This n8n workflow is a complete marketing automation system that connects to your CDP (Customer Data Platform), selects which flows to send, and delivers personalized emails using Brevo. It's modular and extensible — you can also add SMS, push notifications, Telegram messages, or other channels. To build a full marketing automation system, you need four key components: Workflow Automation – using n8n (this workflow) CDP – store and manage user data (e.g., NocoDB, Metabase, Power BI, etc.) Database – track transactions, templates, and send statuses (e.g., NocoDB) BI / Analytics – monitor performance by flows, journeys, and sent events This workflow represents the Workflow Automation layer. You can connect it to your own data stack or use the included example databases (cdp-ecrm, n8n-templates-ecrm, and n8n-transaction-ecrm) to get started quickly. 👤 Who’s it for? Growth & CRM teams managing user engagement flows Ecommerce marketers running time-sensitive email journeys Marketing automation pros using low-code CRM stacks Data teams building custom campaign triggers from CDPs ✅ Features 🔁 Two modular flows: "Insert user_id" and "Sending Email" 🧠 Select flow using flow_id from templates in NocoDB ✏️ Insert user data into n8n-transaction-ecrm with processing status 🔍 Filter duplicate users by user_id to avoid over-sending 📧 Validate email fields and flag disposables 📨 Send personalized emails using Brevo template parameters 📊 Track delivery with sent_result, sent_at, and status updates 🕒 Runs every 30 minutes via schedule trigger 🛠 How to Use Set your flow In the Setup Flow node, change the flow_id to match a row in your n8n-templates-ecrm table. Prepare your tables in NocoDB cdp-ecrm: contains users (user_id, email, first_name, phone_number) n8n-templates-ecrm: contains flows with metadata n8n-transaction-ecrm: stores and updates user send status Configure credentials NocoDB API Token Brevo (Sendinblue) API Key Trigger the flows Run “Insert user_id” manually or on a schedule to prepare users “Sending Email” runs automatically every 30 minutes 📌 Notes Disposable email domains are filtered using regex Status: 0-processing → just inserted 1-sending → ready to send 2-sent → email sent successfully 3-no-email → missing email address 4-disposal-email → disposable or banned email Easily duplicate the "Insert user_id" flow to add more campaigns