by Marcial Ambriz
Remixed Backup your workflows to GitHub from Solomon's work. Check out his templates. How it works This workflow will backup your workflows to GitHub. It uses the n8n API node to export all workflows. It then loops over the data, checks in GitHub to see if a file exists that uses the credential's ID. Once checked it will: update the file on GitHub if it exists; create a new file if it doesn't exist; ignore if it's the same. In addition, it also checks if any workflows have been deleted from n8n. If a workflow no longer exists in n8n, the corresponding file will be removed from the repository to keep everything in sync. Who is this for? People wanting to backup their workflows outside the server for safety purposes or to migrate to another server.
by Aayushman Sharma
Automatically create Google Tasks from new Gmail emails labeled "To-Do". Who is this for? This template is perfect for individuals and teams who want to boost their productivity by automatically converting important emails into actionable tasks in Google Tasks. What problem is this workflow solving? Manually managing emails and creating tasks can be tedious. This workflow ensures you never miss a follow-up by instantly turning important emails into tasks without switching between apps. What this workflow does? Watches for new emails in Gmail with the label "To-Do". Creates a new Google Task with the email subject as the task title and the email snippet as notes. Sets the task due date to 24 hours after the email is received. Setup Create a label "To-Do" in your Gmail account if it doesn't already exist. Connect your Gmail and Google Tasks accounts to n8n using OAuth2 credentials. Import the workflow into n8n and activate it. How to customize this workflow to your needs? Change the Gmail label to a different one (e.g., "Important", "Follow-up"). Modify the due date logic in the expression if you want more/less time to complete tasks: {{ $now.add(2, 'days').toISOString() }} Add additional Gmail filters (like only unread emails) to refine which emails create tasks.
by Oneclick AI Squad
In this guide, we’ll walk you through setting up an AI-driven workflow that automatically fetches daily sales, food waste, and customer feedback data from Google Sheets, generates actionable insights using AI, merges them into a comprehensive report, and sends it as an email draft. Ready to automate your restaurant’s daily insights? Let’s dive in! What’s the Goal? Automatically retrieve daily sales data, food waste records, and customer feedback from Google Sheets. Use AI to analyze data and generate insights, including top performers, waste reduction recommendations, and feedback summaries. Merge the insights into a structured daily report. Send the report as an AI-generated email draft for review or sending. Enable scheduled automation for daily insights delivery. By the end, you’ll have a self-running system that delivers daily restaurant insights effortlessly. Why Does It Matter? Manual data analysis and reporting are time-consuming and error-prone. Here’s why this workflow is a game-changer: Zero Human Error**: AI ensures accurate and consistent insights. Time-Saving Automation**: Instantly process data and draft reports, boosting efficiency. Scheduled Delivery**: Receive insights daily without manual effort. Actionable Insights**: Empower your team with data-driven decisions. Think of it as your tireless data analyst that keeps your restaurant informed. How It Works Here’s the step-by-step magic behind the automation: Step 1: Trigger the Workflow Initiate the workflow daily using the Daily Report Scheduler node (e.g., every day at a set time). Step 2: Fetch Daily Sales Data Retrieve sales data from the Google Sheet using the Fetch Daily Sales Data node. Step 3: Fetch Daily Food Waste Records Retrieve food waste data from the Google Sheet using the Fetch Daily Food Waste Records node. Step 4: Fetch Customer Feedback Retrieve customer feedback from the Google Sheet using the Fetch Customer Feedback node. Step 5: Normalize Sales Records Process and standardize sales data for AI analysis. Step 6: Normalize Waste Data Process and standardize food waste data for AI analysis. Step 7: Normalize Feedback Data Process and standardize customer feedback data for AI analysis. Step 8: AI Sales Insights Generator Use AI (e.g., Google Chat Model) to analyze sales data, identify top performers, and provide recommendations. Step 9: AI Waste Reduction Insights Generator Use AI to analyze waste data and suggest reduction strategies. Step 10: AI Feedback Summary Use AI to summarize customer feedback and identify common themes. Step 11: Format Sales Output Structure the sales insights into a readable format. Step 12: Format Waste Output Structure the waste reduction insights into a readable format. Step 13: Format Feedback AI Output Structure the feedback summary into a readable format. Step 14: Merge & Create Email Combine all formatted insights into a single daily report email draft. Step 15: Prepare Email Content Finalize the email content for sending. Step 16: Send Daily Report Send the AI-generated daily summary email via Gmail. How to Use the Workflow? Importing a workflow in n8n is a straightforward process that allows you to use pre-built workflows to save time. Below is a step-by-step guide to importing the Restaurant Daily Insights Automation workflow in n8n. Steps to Import a Workflow in n8n Obtain the Workflow JSON Source the Workflow: Workflows are shared as JSON files or code snippets, e.g., from the n8n community, a colleague, or exported from another n8n instance. Format: Ensure you have the workflow in JSON format, either as a file (e.g., workflow.json) or copied text. Access the n8n Workflow Editor Log in to n8n (via n8n Cloud or self-hosted instance). Navigate to the Workflows tab in the n8n dashboard. Click Add Workflow to create a blank workflow. Import the Workflow Option 1: Import via JSON Code (Clipboard): Click the three dots (⋯) in the top-right corner to open the menu. Select Import from Clipboard. Paste the JSON code into the text box. Click Import to load the workflow. Option 2: Import via JSON File: Click the three dots (⋯) in the top-right corner. Select Import from File. Choose the .json file from your computer. Click Open to import. Setup Notes Google Sheet Columns**: Sales Data Sheet: Date, Item Name, Quantity Sold, Revenue, Cost, Profit. Food Waste Records Sheet: Date, Item Name, Waste Quantity, Reason, Timestamp. Customer Feedback Sheet: Date, Customer Name, Feedback Text, Rating, Timestamp. Google Sheets Credentials**: Configure OAuth2 settings in the fetch nodes with your Google Sheet ID and credentials. AI Models**: Set up the AI nodes (e.g., Google Chat Model) with appropriate API credentials. Gmail Integration**: Authorize the Send Daily Report node with Gmail API credentials to send emails. Scheduling**: Adjust the Daily Report Scheduler node to your preferred time (e.g., daily at 9 AM).
by Guillaume
🟡 DESCRIPTION This n8n workflow automates the daily generation of comprehensive analytics reports from multiple websites, processes them using OpenAI's powerful language models, and then delivers the insights directly to your email and Slack channels. Designed for digital marketers, website administrators, and data analysts, this template streamlines the tedious process of manually collecting and interpreting website performance data, offering immediate, actionable intelligence. The core problem this workflow solves is the fragmentation and manual effort involved in monitoring numerous website analytics. Instead of logging into multiple Google Analytics accounts or properties daily, this workflow centralizes data extraction, enriches it with AI-powered summaries, and distributes it to your preferred communication platforms. This significantly reduces the time spent on routine reporting, allowing you to focus on strategy and decision-making. Users who would benefit most from this template include: Digital Marketing Agencies: Efficiently generate and share client reports across multiple properties. E-commerce Businesses: Monitor performance of various product sites or regional domains. Content Creators: Track the success of different blogs or content platforms. Anyone managing a portfolio of websites: Gain a unified view of their digital assets' performance. 🟡 SETUP Before importing and running this workflow, ensure you have the following prerequisites in place: ► Google Analytics API Setup: Go to the Google Cloud Console (https://console.cloud.google.com/). Create a new project or select an existing one. Navigate to "APIs & Services" > "Enabled APIs & services." Search for and enable the "Google Analytics Data API" (for GA4) or "Google Analytics Reporting API" (for Universal Analytics, if still in use). Go to "APIs & Services" > "Credentials." Create "OAuth client ID" credentials for a "Web application." Copy your client ID and client secret. In n8n, add a new Google Analytics credential using the "OAuth2" authentication method. Provide the Client ID and Client Secret. You will need to authorize n8n to access your Google Analytics data. Scopes https://www.googleapis.com/auth/analytics.readonly (read-only) or https://www.googleapis.com/auth/analytics (manage) OAuth Redirect URL Already pre-filled by n8n (copy this when creating the Google client) Click Sign in with Google, choose the GA account, and approve the requested scopes. Click Save – n8n tests the token and stores it encrypted. For each yourwebsite.com node, select these credentials and specify the correct Google Analytics View ID (for Universal Analytics) or Property ID (for GA4). ► Email Credentials (SMTP): In n8n, add a new SMTP credential. Provide your SMTP host, port, username, and password. This will be used to send the aggregated reports. ► Slack Credentials: In n8n, add a new Slack credential. Follow the instructions to create a Slack App and generate a Bot User OAuth Token. This token will allow n8n to post messages to your Slack channels. ► OpenAI Credentials: Sign up for an OpenAI account (https://platform.openai.com/). Generate an API Key from your account dashboard. In n8n, add a new OpenAI credential and paste your API Key. ► Website Configuration: The workflow currently includes nodes for yourwebsite.com (1 to 8). For each "Get Report" node (e.g., yourwebsite.com, yourwebsite.com 2, etc.), duplicate or remove them as needed to match the number of Google Analytics properties you wish to monitor. Replace the names and number of websites inside the report maker agent. 🟡 Usage Once you have configured the credentials, you can customize the workflow to fit your specific needs, including leveraging various alternative nodes for sending notifications and reports. This flexibility allows you to integrate seamlessly with your preferred communication tools such as Gmail, Telegram, Outlook, or any other service supported by n8n's extensive node library.
by Ayham Joumran
How It Works This template is a complete, hands-on tutorial for building a RAG (Retrieval-Augmented Generation) pipeline. In simple terms, you'll teach an AI to become an expert on a specific topic—in this case, the official n8n documentation—and then build a chatbot to ask it questions. Think of it like this: instead of a general-knowledge AI, you're building an expert librarian. 🔧 Workflow Overview The workflow is split into two main parts: Part 1: Indexing the Knowledge (📚 Building the Library) This is a one-time process you run manually. The workflow will: Automatically scrape all pages of the n8n documentation. Break them down into small, digestible chunks. Use an AI model to create a numerical representation (an embedding) for each chunk. Store these embeddings in n8n's built-in Simple Vector Store. > This is like a librarian reading every book and creating a hyper-detailed index card for every paragraph. > ⚠️ Important: This in-memory knowledge base is temporary. It will be erased if you restart your n8n instance. You'll need to run the indexing process again in that case. Part 2: The AI Agent (🧠 The Expert Librarian) This is the chat interface. When you ask a question: The AI agent doesn't guess the answer. It searches the knowledge base to find the most relevant “index cards” (chunks). It feeds those chunks to a language model (Gemini) with strict instructions: > “Answer the user's question using ONLY this information.” This ensures answers are accurate, factual, and grounded in your documents. 🚀 Setup Steps > Total setup time: ~2 minutes > Indexing time: ~15–20 minutes This template uses n8n’s built-in tools, so no external database is needed. 1. Configure OpenAI Credentials You’ll need an OpenAI API key (for GPT models). In your n8n workflow: Go to any of the three OpenAI nodes (e.g., OpenAI Chat Model). Click the Credential dropdown → + Create New Credential. Enter your OpenAI API key and save. 2. Apply Credentials to All Nodes Your new credential is now saved. Go to the other two OpenAI nodes (e.g., OpenAI Embeddings) and select the newly created credential from the dropdown. 3. Build the Knowledge Base Find the Start Indexing manual trigger node (top-left of the workflow). Click the Execute Workflow button to start indexing. > ⚠️ Be patient: This takes 15–20 minutes to scrape and process the full documentation. > You only need to do this once per n8n session. 4. Chat With Your Expert Agent After indexing completes, activate the entire workflow (toggle at the top). Open the RAG Chatbot chat trigger node (bottom-left). Copy its Public URL. Open it in a new tab and ask questions about n8n! Example questions: "How does the IF node work?" "What is a sub-workflow?" 👤 Credits All credits go to Lucas Peyrin 🔗 lucaspeyrin on n8n.io
by Le Nguyen
Turn your blog into a set-and-forget content engine: every new article is instantly repurposed into channel-specific social posts with visuals, keeping your brand visible on LinkedIn, X, and Reddit without extra copywriting time. Perfect for lean marketing teams who want consistent, always-on distribution from a single source of content. How it works • Watches your blog RSS feed (or receives a single URL) and detects new articles. • Saves each post in Postgres so every article is only processed once. • Fetches the article HTML, extracts the main body content and sends it to OpenAI (GPT-4.1). • OpenAI creates platform-optimized copy: 1 LinkedIn post, 1 X/Twitter post, 1 Reddit post + image prompts. • Generates on-brand images with OpenAI and publishes everything automatically to LinkedIn, X, and Reddit. • You can also trigger it manually or via webhook whenever you want to push a specific campaign. Setup Steps • Time: around 20–40 minutes for someone familiar with n8n and the platforms. • Create a Postgres table “rss_items” with fields: guid (PRIMARY KEY), title, link, published_at. • Add credentials in n8n for: – Postgres – OpenAI – LinkedIn OAuth2 – X/Twitter OAuth2 + OAuth1 (for media upload) – Reddit OAuth2 • In the RSS node, set your blog feed URL (for example: https://yourblog.com/feed). • In the webhook node, confirm the URL/path you want external tools or other workflows to call with a “link” field. • Run the manual trigger with one test blog URL to verify: – Article content is extracted correctly. – AI returns LinkedIn/X/Reddit posts and image prompts. – Posts and images appear correctly on all social accounts. • Once tests look good, enable the Schedule Trigger so Blog2Social AI runs automatically at your chosen interval.
by Luan Correia
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This comprehensive RAG workflow enables your AI agents to answer user questions with contextual knowledge pulled from your own documents — using metadata-rich embeddings stored in Supabase. 🔧 Key Features: RAG Agents powered by GPT-4.5 or GPT-3.5 via OpenRouter or OpenAI. Supabase Vector Store to store and retrieve document embeddings. Cohere Reranker to improve response relevance and quality. Metadata Agent to enrich vectorized data before ingestion. PDF Extraction Flow to automatically parse and upload documents with metadata. ✅ Setup Steps: Connect your Supabase Vector Store. Use OpenAI Embeddings (e.g. text-embedding-3-small). Add API keys for OpenAI and/or OpenRouter. Connect a reranker like Cohere. Process documents with metadata before embedding. Start chatting — your AI agent now returns context-rich answers from your own knowledge base! Perfect for building AI assistants that can reason, search and answer based on internal company data, academic papers, support docs, or personal notes.
by Incrementors
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 📦 Multi-Platform Price Finder: Scraping Prices with Bright Data & Telegram An intelligent n8n automation that fetches real-time product prices from marketplaces like Amazon, Wayfair, Lowe's, and more using Bright Data's dataset, and sends promotional messages via Telegram using AI—perfect for price tracking, deal alerts, and affiliate monetization. 📋 Overview This automation tracks product prices across top e-commerce platforms using Bright Data and sends out alerts via Telegram based on the best available deals. The workflow is designed for affiliate marketers, resellers, and deal-hunting platforms who want real-time competitive pricing. ✨ Key Features 🔎 Multi-Platform Scraping: Supports Amazon, Wayfair, Lowe's, and more ⚡ Bright Data Integration: Access to structured product snapshots 📢 AI-Powered Alerts: Generates Telegram-ready promo messages using AI 🧠 Lowest Price Logic: Filters and compares products across sources 📈 Data Merge & Processing: Combines multiple sources into a single stream 🔄 Keyword-Driven Search: Searches using dynamic keywords from form input 📦 Scalable Design: Built for multiple platform processing simultaneously 🧼 Clean Output: Strips unnecessary formatting before publishing 🎯 What This Workflow Does Input Search Keywords**: User-defined keyword(s) from a form trigger Platform Sources**: Wayfair, Lowe's, Amazon, etc. Bright Data API Key**: Needed for authenticated scraping Processing Steps User Input via n8n form trigger (keyword-based) Bright Data API Trigger for each marketplace Status Polling: Wait until scraping snapshot is ready Data Retrieval: Fetches JSON results from Bright Data snapshot Data Cleaning & Normalization: Price, title, and URL are extracted Merging Products from all platforms Find Lowest Price Product using custom JS logic AI Prompt Generation via Claude/Anthropic Telegram Formatting and alert message creation Output 🛍️ Product Title 💰 Final Price 🔗 Product URL ✉️ Promotional Message (for Telegram/notifications) 🚀 Setup Instructions Step 1: Import Workflow Open n8n > Workflows > + Add Workflow Import the provided JSON file Step 2: Configure Bright Data Add credentials under Credentials → Bright Data API Set the appropriate dataset_id for each platform Ensure dataset includes title, price, and url fields Step 3: Enable Keyword Trigger Use the built-in Form Trigger node Input: Single keyword field (SearchHere) Step 4: Telegram or AI Integration Modify prompt node for your language or tone Add Telegram webhook or integration where needed 📖 Usage Guide Adding Keywords Trigger the form with a product keyword like iPhone 15 Wait for workflow to fetch best deals and generate Telegram message Understanding AI-Powered Output AI creates a short, engaging message like: > "🔥 Deal Alert: Get the iPhone 15 for just ₹74,999! Limited stock—Check it out: [link]" Debugging Output Output node shows cleaned JSON with title, price, url, and message If no valid results, debug message is returned with sample structure info 🔧 Customization Options Add More Marketplaces Clone any HTTP Request node (e.g., for Wayfair) Update dataset_id and required output fields Modify Price Logic Update the Code1 node to change comparison (e.g., highest price instead of lowest) Change Message Format Edit the AI Agent prompt to customize tone/language Add emoji, CTAs, or markdown formatting as needed 🧪 Test & Activation Add a few sample keywords via form trigger Run manually or set as a webhook for external app input Check final AI-generated message in output node 🚨 Troubleshooting | Issue | Solution | |-------|----------| | No Data Returned | Ensure keyword matches real products | | Status Not 'Ready' | Bright Data delay; add Wait nodes | | Invalid API Key | Check Bright Data credentials | | AI Errors | Adjust prompt or validate input fields | 📊 Use Cases 💰 Affiliate Campaigns: Show best deals across platforms 🛒 Deal Pages: Post live offers with product links 🧠 Competitor Analysis: Track cross-platform pricing 🔔 Alert Bots: Send real-time alerts to Telegram or Slack ✅ Quick Setup Checklist [x] Bright Data API credentials configured [x] n8n form trigger enabled [x] Claude or AI model connected [x] All HTTP requests working [x] AI message formatting verified 🌐 Example Output { "title": "Apple iPhone 15 Pro Max", "price": 1199, "url": "https://amazon.com/iphone-15", "message": "🔥 Grab the Apple iPhone 15 Pro Max for just $1199! Limited deal—Check it out: https://amazon.com/iphone-15" } 📬For any questions or support, please contact: 📧 <info@incrementors.com> or fill out this form: https://www.incrementors.com/contact-us/
by mustafa kendigüzel
How it works This automated workflow discovers trending Instagram posts and creates similar AI-generated content. Here's the high-level process: 1. Content Discovery & Analysis Scrapes trending posts from specific hashtags Analyzes visual elements using AI Filters out videos and duplicates 2. AI Content Generation Creates unique images based on trending content Generates engaging captions with relevant hashtags Maintains brand consistency while being original 3. Automated Publishing Posts content directly to Instagram Monitors publication status Sends notifications via Telegram Set up steps Setting up this workflow takes approximately 15-20 minutes: 1. API Configuration (7-10 minutes) Instagram Business Account setup Telegram Bot creation API key generation (OpenAI, Replicate, Rapid Api) 2. Database Setup (3-5 minutes) Create required database table Configure PostgreSQL credentials 3. Workflow Configuration (5-7 minutes) Set scheduling preferences Configure notification settings Test connection and permissions Detailed technical specifications and configurations are available in sticky notes within the workflow.
by Sina
👔 Who is this for? Entrepreneurs and startup founders preparing for investors Business consultants drafting complete client plans Strategy teams building long-term business models Accelerators, incubators, or pitch trainers ❓ What problem does this workflow solve? Writing a full business plan takes days of work, multiple tools, and often gets stuck in messy docs or slides. This template automates every major section, generating a clean, detailed, and professional business plan with AI in just minutes. ⚙️ What this workflow does Starts with a chat message asking for your business idea or startup concept Passes the idea through 83 intelligent agents, each handling a full business plan chapter: Executive Summary Problem & Solution Product Description Market Research Competitor Analysis Business Model Marketing Strategy (includes guerrilla ideas) Operational Plan Financial Plan Team & Advisors Roadmap Conclusion & Next Steps Each section uses tailored prompts and business logic Combines all outputs into a structured, professional Markdown file Final result: a ready-to-export business plan in seconds 🛠️ Setup Import this template into your n8n instance Replace the “LLM Chat Model” node with your preferred model (Ollama, GPT-4, etc.) Start from the chat input node — describe your startup or idea Wait for all agents to finish Download the final Business plan file 🤖 LLM Flexibility (Choose Your Model) Supports: OpenAI (GPT-4 / GPT-3.5) Ollama (LLaMA 3.1, Mistral, DeepSeek, etc.) Any compatible N8N chat model To change the model, just replace the “Language Model” node — no other updates required 📌 Notes All nodes are clearly named by function (e.g., “Market Research Generator”) Sticky notes included for clarity Generates high-quality plans suitable for VCs or accelerators Modular: you can turn off or reorder any chapter 📩 Need help? Email: sinamirshafiee@gmail.com Happy to support setup, LLM switching, or custom section development.
by Constantine Kissel
Generate research-backed article with n8n Who’s it for Content marketers, SEO teams, and founders who need fast, research-grounded blog posts or long-form articles—multi-language included. Works well for teams that want citations, outlines, and section-by-section drafting with minimal manual effort. How it works / What it does Use a Form to collects domain, keywords, and target language. The workflow refines keywords, finds recent articles and authoritative citations, then synthesizes a master outline and loops through each section: generate search queries → fetch web results → summarize findings → write the section with the advanced model. Finally, it aggregates all sections into a clean Markdown article. Optional delivery nodes (Email, Telegram) and an AI Agent are included but disabled by default.  How to set up Import the workflow JSON into n8n. Add your OpenAI credential Set simple_model / advanced_model in LLM Params Requirements n8n instance with outbound internet access. OpenAI API access (Responses API + web_search_preview). (Optional) Email/Telegram credentials if you want to deliver results async How to customize the workflow Edit prompts in LLM Params and Section Prompts to match tone, structure, and SEO style Tweak recency and source rules in Search Articles / Search Citations.  Insert a human review step before “Write Section”, enable delivery nodes Change working/output languages in the Language node
by Cyril Nicko Gaspar
📌 HubSpot Lead Enrichment with Bright Data MCP This template enables natural-language-driven automation using Bright Data's MCP tools, triggered directly by new leads in HubSpot. It dynamically extracts and executes the right tool based on lead context—powered by AI and configurable in N8N. ❓ What Problem Does This Solve? Manual lead enrichment is slow, inconsistent, and drains valuable time. This solution automates the process using a no-code workflow that connects HubSpot, Bright Data MCP, and an AI agent—without requiring scripts or technical skills. Perfect for marketing, sales, and RevOps teams. 🧰 Prerequisites To use this template, you’ll need: A self-hosted or cloud instance of N8N A Bright Data MCP API token A valid OpenAI API key (or compatible AI model) A HubSpot account Either a Private App token or OAuth credentials for HubSpot Basic familiarity with N8N workflows ⚙️ Setup Instructions 1. Set Up Authentication in HubSpot 🔐 Option 1: Use a Private App Token (Simple Setup) Log in to your HubSpot account. Navigate to Settings → Integrations → Private Apps. Create a new Private App with the following scopes: crm.objects.contacts.read crm.objects.contacts.write crm.schemas.contacts.read crm.objects.companies.read (optional) Copy the Access Token. In N8N, create a credential for HubSpot App Token and paste the app token in the field. Go back to Hubspot Private App settings to setup a webhook. Copy the url in your workflow's Webhook node and paste it here. 🔁 Option 2: Use OAuth (Advanced + Secure) In HubSpot, go to Settings → Integrations → Apps → Create App. Set your Redirect URL to match your N8N OAuth2 redirect path. Choose scopes like: crm.objects.companies.read crm.objects.contacts.read crm.objects.deals.read crm.schemas.companies.read crm.schemas.contacts.read crm.schemas.deals.read crm.objects.contacts.write (conditionally required) Note the Client ID and Client Secret. Copy the App ID and the developer API key In N8N, create a credential for HubSpot Developer API and paste those info from previous step. Attach these credentials to the HubSpot node in N8N. 2. Setup and obtain API token and other necessary information from Bright Data In your Bright Data account, obtain the following information: API token Web Unlocker zone name (optional) Browser API username and password string separated by colon (optional) 3. Host SSE server from STDIO command The methods below will allow you to receive SSE (Server-Sent Events) from Bright Data MCP via a local Supergateway or Smithery ** Method 1: Run Supergateway in a separate web service (Recommended) This method will work for both cloud version and self-hosted N8N. Signup to any cloud services of your choice (DigitalOcean, Heroku, Hetzner, Render, etc.). For NPM based installation: Create a new web service. Choose Node.js as runtime environment and setup a custom server without repository. In your server’s settings to define environment variables or .env file, add: `API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_AUTH=optional_browser_auth` Paste the following text as a start command: npx -y supergateway --stdio "npx -y @brightdata/mcp" --port 8000 --baseUrl http://localhost:8000 --ssePath /sse --messagePath /message Deploy it and copy the web server URL, then append /sse into it. Your SSE server should now be accessible at: https://your_server_url/sse For Docker based installation: Create a new web service. Choose Docker as the runtime environment. Set up your Docker environment by pulling the necessary images or creating a custom Dockerfile. In your server’s settings to define environment variables or .env file, add: `API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_ZONE=optional_browser_zone_name` Use the following Docker command to run Supergateway: `docker run -it --rm -p 8000:8000 supercorp/supergateway \ --stdio "npx -y @brightdata/mcp /" \ --port 8000` Deploy it and copy the web server URL, then append /sse into it. Your SSE server should now be accessible at: https://your_server_url/sse For more installation guides, please refer to https://github.com/supercorp-ai/supergateway.git. ** Method 2: Run Supergateway in the same web service as the N8N instance This method will only work for self-hosted N8N. a. Set Required Environment Variables In your server's settings to define environment variables or .env file, add: API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_ZONE=optional_browser_zone_name b. Run Supergateway in Background npx -y supergateway --stdio "npx -y @brightdata/mcp" --port 8000 --baseUrl http://localhost:8000 --ssePath /sse --messagePath /message Use the command above to execute it through the cloud shell or set it as a pre-deploy command. Your SSE server should now be accessible at: http://localhost:8000/sse For more installation guides, please refer to https://github.com/supercorp-ai/supergateway.git. * *Method 3: Configure via Smithery.ai* (Easiest) If you don't want additional setup and want to test it right away, follow these instructions: Visit https://smithery.ai/server/@luminati-io/brightdata-mcp/tools to: Signup (if you are new to Smithery) Create an API key Define environment variables via a profile Retrieve your SSE server HTTP URL 4. Connect Google Sheets to N8N Ensure your Google Sheet: Contains columns like row_id, first_name, last_name, email, and status. Is shared with your N8N service account (or connected via OAuth) In N8N: Add a Google Sheets Trigger node Set it to watch for new rows in your lead sheet 5. Import and Configure the N8N Workflow Import the provided JSON workflow into N8N Update nodes with your credentials: Hubspot: Add your API key or connect it via OAuth. Google Sheets Trigger: Link to your actual sheet OpenAI Node: Add your API key Bright Data Tool Execution: Add Bright Data token and SSE URL 🔄 How It Works New contact in Hubspot or a new row is added to the Google Sheet N8N triggers the workflow AI agent classifies the task (e.g., “Find LinkedIn”, “Get company info”) The relevant MCP tool is called Results are appended back to the sheet or routed to another destination Rerun the specific record by specifying status "needs more enrichment", or leaving it blank. 🧩 Use Cases B2B Lead Enrichment** – Add missing fields (title, domain, social profiles) Email Intelligence** – Validate and enrich based on email Market Research** – Pull company or contact data on demand CRM Auto-fill** – Push enriched leads to tools like HubSpot or Salesforce 🛠️ Customization Prompt Tuning** – Adjust how the AI interprets input data Column Mapping** – Customize which fields to pull from the sheet Tool Logic** – Add retries, fallback tools, or confidence-based routing Destination Output** – Integrate with CRMs, Slack, or webhook endpoints ✅ Summary This template turns a Google Sheet into an AI-powered lead enrichment engine. By combining Bright Data’s tools with a natural language AI agent, your team can automate repetitive tasks and scale lead ops—without writing code. Just add a row, and let the workflow do the rest.