by WeblineIndia
This n8n workflow automatically fetches all interview events scheduled for today from a specified Google Calendar and sends a personalized email to each interviewer. The email contains a formatted HTML table listing their interview details, including meeting times, Google Meet links, and attendees with their response status. This ensures all interviewers are informed daily at 8:00 AM IST without any manual coordination. Who’s it for Interviewers** who want a quick morning packet instead of opening multiple calendar tabs. Recruiters/coordinators** who need a reliable, zero‑friction daily brief for interviewers. Teams** that paste CV/notes links directly into calendar events (no file search required). How it works Cron triggers daily at 08:00 IST. Google Calendar reads today’s events from the Interviews calendar. A Code step parses each event to identify interviewers and extract candidate details, meeting link, and any CV: / Notes: links from the description and create a table to share via Gmail. A grouping step compiles a per‑interviewer list for the day. Email (Gmail) sends one digest to interviewer. How to set up Ensure all interviews are scheduled on the Google Calendar named Interviews and that interviewers are added as attendees. Add CV: <url> and Notes: <url> in the event description when available. Import the workflow and add credentials: Google Calendar (OAuth) SMTP/Gmail for sending the email digests Keep the default 08:00 IST schedule in the Cron node or adjust as needed. Requirements Google Workspace account with access to the Interviews calendar. Gmail sender account for digests (App Password if using 2FA). n8n instance (cloud or self‑hosted). Steps Trigger Schedule Node:** Schedule Trigger Purpose:* Starts the workflow daily at *8:00 AM**. Fetch Interview Events Node:** Google Calendar(Fetch Interview Events) Purpose:** Retrieves all events (interviews) from the configured calendar. Output:** Event details including summary, time, and organizer email. Group & Format Schedule Node:** HTML Table (JavaScript Code Node) Purpose:** Groups events by interviewer email and generates an HTML schedule table. Output:** Formatted fields: interviewer_email subject htmlContent Send Personalized Emails Node:** Gmail Purpose:** Sends the formatted interview schedule to each interviewer’s email address. Send To:** Dynamically set using ={{ $json.interviewer_email }} Subject:** "Interview Reminder" Body:** ={{ $json.htmlContent }} (HTML) How to customize the workflow Parsing rules:** If your event titles follow a pattern (e.g., Onsite – {Candidate} – {Role}), tweak the regex in the Code node. Attendee logic:** Refine how interviewers are detected (e.g., filter only accepted responses, or include tentative). Digest format:** Adjust table columns/order, or add role/team labels from the title. Schedule:** Duplicate the Cron for regional time zones or add a midday refresh. Add-ons to level up the Workflow with additional nodes Reminder pings:** Add 10‑minute pre‑interview reminders via Email or Slack/Teams. Conflict alerts:** Flag if an interviewer is double‑booked within a 15‑minute window. Feedback follow‑up:** After the scheduled time, DM or email a standardized feedback form link. Drive search (optional):** If you later want auto‑attach CVs, add a Google Drive search step (by candidate name) in a designated folder. Common troubleshooting points No events found:* Confirm the calendar name is *Interviews* and that events exist *today**. Missing links:** If CV/notes links aren’t in the email, add CV:/Notes: links to the event description. Email not sent:** Verify SMTP credentials, from‑address permissions, and any sending limits. Time mismatch:* Confirm workflow timezone and Calendar times are set to *Asia/Kolkata** (or adjust). A short note If you need help tailoring the parsing rules, adjusting the schedule or troubleshooting any step, feel free to reach out we will happy to help.
by Omer Fayyaz
This workflow automates news aggregation and summarization by fetching relevant articles from Gnews.io and using AI to create concise, professional summaries delivered via Slack What Makes This Different: Real-Time News Aggregation** - Fetches current news articles from Gnews.io API based on user-specified topics AI-Powered Summarization** - Uses GPT-4.1 to intelligently select and summarize the most relevant articles Professional Formatting** - Generates clean, readable summaries with proper dates and article links Form-Based Input** - Simple web form interface for topic specification Automated Delivery** - Sends summarized news directly to Slack for immediate consumption Intelligent Filtering** - AI selects the top 15 most relevant articles from search results Key Benefits of Automated News Summarization: Time Efficiency** - Transforms hours of news reading into minutes of focused summaries Comprehensive Coverage** - AI ensures all important financial and business developments are captured Professional Quality** - Generates publication-ready summaries with proper formatting Real-Time Updates** - Always delivers the latest news on any topic Centralized Access** - All news summaries delivered to one Slack channel Customizable Topics** - Search for news on any subject matter Who's it for This template is designed for business professionals, financial analysts, content creators, journalists, and anyone who needs to stay updated on specific topics without spending hours reading through multiple news sources. It's perfect for professionals who want to stay informed about industry developments, market trends, or any specific subject matter while maintaining productivity. How it works / What it does This workflow creates an automated news summarization system that transforms topic searches into professional news summaries. The system: Receives topic input through a simple web form interface Fetches news articles from Gnews.io API based on the specified topic Maps article data to prepare for AI processing Uses AI to select the 15 most relevant articles related to financial advancements, tools, research, and applications Generates professional summaries with clear, readable language and proper formatting Includes article links and current date for complete context Delivers summaries via Slack notification for immediate review Key Innovation: Intelligent News Curation - Unlike basic news aggregators, this system uses AI to intelligently filter and summarize only the most relevant articles, saving time while ensuring comprehensive coverage of important developments. How to set up 1. Configure Form Trigger Set up n8n form trigger with "topic" field (required) Configure form title as "News Search" Test form submission functionality Ensure proper data flow to subsequent nodes 2. Configure Gnews.io API Get your API key**: Sign up at gnews.io and obtain your API key from the dashboard Add API key to workflow**: In the "Get GNews articles" HTTP Request node, replace "ADD YOUR API HERE" with your actual Gnews.io API key Example configuration**: { "q": "{{ $json.topic }}", "lang": "en", "apikey": "your-actual-api-key-here" } Configure search parameters**: Ensure language is set to "en" for English articles Test API connectivity**: Run a test execution to verify news articles are returned correctly 3. Configure OpenAI API Set up OpenAI API credentials in n8n Ensure proper API access and quota limits Configure the GPT-4.1 Model node for AI summarization Test AI model connectivity and response quality 4. Configure Slack Integration Set up Slack API credentials in n8n Configure Slack channel ID for news delivery Set up proper message formatting for news summaries Test Slack notification delivery 5. Test the Complete Workflow Submit test form with sample topic (e.g., "artificial intelligence") Verify Gnews.io returns relevant articles Check that AI generates appropriate news summaries Confirm Slack notification contains formatted news summary Requirements n8n instance** with form trigger and HTTP request capabilities OpenAI API** access for AI-powered news summarization Gnews.io API** credentials for news article fetching Slack workspace** with API access for news delivery Active internet connection** for real-time API interactions How to customize the workflow Modify News Search Parameters Adjust the number of articles to summarize (currently set to 15) Add more search depth options or date ranges Implement language filtering for different regions Add news source filtering or preferences Enhance AI Capabilities Customize AI prompts for specific industries or niches Add support for multiple languages Implement different summary styles (brief, detailed, bullet points) Add content quality scoring and relevance filtering Expand News Sources Integrate with additional news APIs (NewsAPI, Bing News, etc.) Add support for RSS feed integration Implement trending topic detection Add competitor news monitoring Improve News Delivery Add email notifications alongside Slack Implement news scheduling capabilities Add news approval workflows Implement news performance tracking Business Features Add news analytics and engagement metrics Implement A/B testing for different summary formats Add news calendar integration Implement team collaboration features for news sharing Key Features Real-time news aggregation** - Fetches current news articles from Gnews.io API AI-powered summarization** - Uses GPT-4.1 to intelligently select and summarize relevant articles Professional formatting** - Generates clean, readable summaries with proper dates and links Form-based input** - Simple interface for topic specification Automated workflow** - End-to-end automation from topic input to news delivery Intelligent filtering** - AI selects the most relevant articles from search results Slack integration** - Centralized delivery of news summaries Scalable news processing** - Handles multiple topic searches efficiently Technical Architecture Highlights AI-Powered News Summarization OpenAI GPT-4.1 integration** - Advanced language model for intelligent news summarization Content filtering** - AI selects the 15 most relevant articles from search results Professional formatting** - Generates clean, readable summaries with proper structure Quality consistency** - Maintains professional tone and formatting standards News API Integration Gnews.io API** - Comprehensive news search with article extraction Real-time data** - Access to current, relevant news articles Content mapping** - Efficiently processes article data for AI analysis Search optimization** - Efficient query construction for better news results Form-Based Input System n8n form trigger** - Simple, user-friendly input interface for topic specification Data validation** - Ensures required topic field is properly filled Parameter extraction** - Converts form data to search parameters Error handling** - Graceful handling of incomplete or invalid inputs News Delivery System Slack integration** - Professional news summary delivery Formatted output** - Clean, readable summaries with dates and article links Centralized access** - All news summaries delivered to one location Real-time delivery** - Immediate notification of news summaries Use Cases Financial analysts** needing to stay updated on market developments and industry news Business professionals** requiring daily news summaries on specific topics Content creators** needing current information for articles and social media posts Journalists** requiring comprehensive news coverage on specific subjects Research teams** needing to track developments in their field of expertise Investment professionals** requiring real-time updates on market trends Academic researchers** needing to stay informed about industry developments Corporate communications** teams requiring news monitoring for crisis management Business Value Time Efficiency** - Reduces news reading time from hours to minutes Cost Savings** - Eliminates need for manual news monitoring and summarization Comprehensive Coverage** - AI ensures all important developments are captured Scalability** - Handles unlimited topic searches without additional resources Quality Assurance** - AI ensures professional-quality summaries every time Real-Time Updates** - Always delivers the latest news on any topic Research Integration** - Incorporates current information for relevant, timely insights This template revolutionizes news consumption by combining AI-powered summarization with real-time news aggregation, creating an automated system that delivers professional-quality news summaries on any topic from a simple form submission.
by Avkash Kakdiya
How it works This workflow automates the generation of ad-ready product images by combining product and influencer photos with AI styling. It runs on a scheduled trigger, fetches data from Google Sheets, and retrieves product and influencer images from Google Drive. The images are processed with OpenAI and OpenRouter to generate enhanced visuals, which are then saved back to Google Drive. Finally, the result is logged into Google Sheets with a ready-to-publish status. Step-by-step 1. Trigger & Data preparation Schedule Trigger** – Runs workflow automatically on a set schedule. Google Sheets (Get the Raw)** – Retrieves today’s product and model URLs. Google Drive (Download Product Image)** – Downloads the product image. Google Drive (Download Influencer Image)** – Downloads the influencer image. Extract from File (Binary → Base64)** – Converts both product and model images for AI processing. 2. AI analysis & image generation OpenAI (Analyze Image)** – Creates an ad-focused visual description (lighting, mood, styling). HTTP Request (OpenRouter Gemini)** – Generates an AI-enhanced image combining product + influencer. Code Node (Cleanup)** – Cleans base64 output to remove extra prefixes. Convert to File** – Transforms AI output into a proper image file. 3. Save & update Google Drive (Upload Image)** – Uploads generated ad image to target folder. Google Sheets (Append/Update Row)** – Stores the Drive link and updates publish status. Why use this? Automates the entire ad image creation process without manual design work. Ensures product visuals are consistent, styled, and ad-ready. Saves final creatives in Google Drive for easy access and sharing. Keeps campaign tracking organized by updating Google Sheets automatically. Scales daily ad production efficiently for multiple products or campaigns.
by Christian Lutz
How it works This workflow automates the delivery of personalized, AI-generated reports or roadmaps for new leads. When someone submits their information through a form, the workflow: Captures and stores the lead data. Uses an AI model to generate a customized report or roadmap. Formats the output into a professional, email-ready HTML document. Sends the report automatically to the lead via email. Optionally sends internal notifications (e.g., via chat or email) for tracking and follow-up. The process eliminates manual work and ensures every lead receives instant, high-quality output tailored to their input. Setup steps Webhook – Connect your form or website to the webhook endpoint to receive lead data. Data Table – Create or link a table to store incoming leads and track delivery status. AI Configuration – Add your OpenAI (or compatible) API credentials and customize prompts for your desired output. Email Setup – Configure SMTP credentials and define sender/recipient addresses for report delivery. Notifications – Optionally connect a chat or messaging service (e.g., Telegram) for internal alerts. Activation – Test the workflow, confirm the data flow and email delivery, then activate it for live use. This workflow transforms manual lead engagement into a fully automated, AI-driven experience that delivers instant, personalized value to every new contact.
by Takuya Ojima
Who’s it for Remote and distributed teams that schedule across time zones and want to avoid meetings landing on public holidays—PMs, CS/AM teams, and ops leads who own cross-regional calendars. What it does / How it works The workflow checks next week’s Google Calendar events, compares event dates against public holidays for selected country codes, and produces a single Slack digest with any conflicts plus suggested alternative dates. Core steps: Workflow Configuration (Set) → Fetch Public Holidays (via a public holiday API such as Calendarific/Nager.Date) → Get Next Week Calendar Events (Google Calendar) → Detect Holiday Conflicts (compare dates) → Generate Reschedule Suggestions (find nearest business day that isn’t a holiday/weekend) → Format Slack Digest → Post Slack Digest. How to set up Open Workflow Configuration (Set) and edit: countryCodes, calendarId, slackChannel, nextWeekStart, nextWeekEnd. Connect your own Google Calendar and Slack credentials in n8n (no hardcoded keys). (Optional) Adjust the Trigger to run daily or only on Mondays. Requirements n8n (Cloud or self-hosted) Google Calendar read access to the target calendar Slack app with permission to post to the chosen channel A public-holiday API (no secrets needed for Nager.Date; Calendarific requires an API key) How to customize the workflow Time window: Change nextWeekStart/End to scan a different period. Holiday sources: Add or swap APIs; merge multiple regions. Suggestion logic: Tweak the look-ahead window or rules (e.g., skip Fridays). Output: Post per-calendar messages, DM owners, or create tentative reschedule events automatically.
by vanhon
Split Test AI Prompts Using Supabase & Langchain Agent This workflow allows you to A/B test different prompts for an AI chatbot powered by Langchain and OpenAI. It uses Supabase to persist session state and randomly assigns users to either a baseline or alternative prompt, ensuring consistent prompt usage across the conversation. 🧠 Use Case Prompt optimization is crucial for maximizing the performance of AI assistants. This workflow helps you run controlled experiments on different prompt versions, giving you a reliable way to compare performance over time. ⚙️ How It Works When a message is received, the system checks whether the session already exists in the Supabase table. If not, it randomly assigns the session to either the baseline or alternative prompt. The selected prompt is passed into a Langchain Agent using the OpenAI Chat Model. Postgres is used as chat memory for multi-turn conversation support. 🧪 Features Randomized A/B split test per session Supabase database for session persistence Langchain Agent + OpenAI GPT-4o integration PostgreSQL memory for maintaining chat context Fully documented with sticky notes 🛠️ Setup Instructions Create a Supabase table named split_test_sessions with the following columns: session_id (text) show_alternative (boolean) Add credentials for: Supabase OpenAI PostgreSQL (for chat memory) Modify the "Define Path Values" node to set your baseline and alternative prompts. Activate the workflow. Send messages to test both prompt paths in action. 🔄 Next Steps Add tracking for conversions or feedback scores to compare outcomes. Modify the prompt content or model settings (e.g. temperature, model version). Expand to multi-variant tests beyond A/B. 📚 Learn More How This Workflow Uses Supabase + OpenAI for Prompt Testing
by Robert Breen
Use this template to upload an image, run a first-pass OpenAI Vision analysis, then re-attach the original file (binary/base64) to the next step using a Merge node. The pattern ensures your downstream AI Agent (or any node) can access both the original file (data) and the first analysis result (content) at the same time. ✅ What this template does Collects an image file* via *Form Trigger** (binary field labeled data) Analyzes the image* with *OpenAI Vision* (GPT-4o) using *base64** input Merges* the original upload and the analysis result (combine by position) so the next node has *both** Re-analyzes/uses* the image alongside the first analysis in an *AI Agent** step 🧩 How it works (Node-by-node) Form Trigger Presents a simple upload form and emits a binary/base64 field named data. Analyze image (OpenAI Vision) Reads the same data field as base64 and runs image analysis with GPT-4o. The node outputs a text content (first-pass analysis). Merge (combine by position) Combines the two branches so the next node receives both the original upload (data) and the analysis (content) on the same item. AI Agent Receives data + content together. Prompt includes the original image (=data) and the first analysis ({{$json.content}}) to compare or refine results. OpenAI Chat Model Provides the language model for the Agent (wired as ai_languageModel). 🛠️ Setup Instructions (from the JSON) > Keep it simple: mirror these settings and you’re good to go. 1) Form Trigger (n8n-nodes-base.formTrigger) Path:* d6f874ec-6cb3-46c7-8507-bd647c2484f0 *(you can change this) Form Title:** Image Document Upload Form Description:** Upload a image document for AI analysis Form Fields:** Label: data Type: file Output:* emits a binary/base64 field named *data**. 2) Analyze image (@n8n/n8n-nodes-langchain.openAi) Resource:** image Operation:** analyze Model:** gpt-4o Text:* =data *(use the uploaded file field) Input Type:** base64 Credentials:* OpenAI (use your stored *OpenAI API** credential) 3) Merge (n8n-nodes-base.merge) Mode:** combine Combine By:** combineByPosition Connect Form Trigger → Merge (input 2) Connect Analyze image → Merge (input 1) This ensures the original file (data) and the analysis (content) line up on the same item. 4) AI Agent (@n8n/n8n-nodes-langchain.agent) Prompt Type:** define Text:** System Message:** analyze the image again and see if you get the same result. Receives:** merged item containing data + content. 5) OpenAI Chat Model (@n8n/n8n-nodes-langchain.lmChatOpenAi) Model:** gpt-4.1-mini Wiring:* connect as *ai_languageModel* to the *AI Agent** Credentials:** same OpenAI credential as above > Security Note: Store API keys in Credentials (do not hardcode keys in nodes). 🧠 Why “Combine by Position” fixes the binary issue Some downstream nodes lose access to the original binary once a branch processes it. By merging the original branch (with data) and the analysis branch (with content) by position, you restore a single item with both fields—so the next step can use the image again while referencing earlier analysis. 🧪 Test Tips Upload a JPG/PNG and execute the workflow from the Form Trigger preview. Confirm Merge output contains both data (binary/base64) and content (text). In the AI Agent, log or return both fields to verify availability. 🔧 Customize Swap GPT-4o for another Vision-capable model if needed. Extend the AI Agent to extract structured fields (e.g., objects detected, text, brand cues). Add a Router after Merge to branch into storage (S3, GDrive) or notifications (Slack, Email). 📝 Requirements n8n (cloud or self-hosted) with web UI access OpenAI** credential configured (Vision support) 🩹 Troubleshooting Binary missing downstream?* Ensure *Merge* receives *both** branches and is set to combineByPosition. Wrong field name?* The *Form Trigger* upload field must be labeled *data** to match node expressions. Model errors?* Verify your *OpenAI* credential and that the chosen model supports *image analysis**. 💬 Sticky Note (included in the workflow) > “Use Binary Field after next step” — This workflow demonstrates how to preserve and reuse an uploaded file (binary/base64) after a downstream step by using a Merge node (combineByPosition). A user uploads an image via Form Trigger → the image is analyzed with OpenAI Vision → results are merged back with the original upload so the next AI Agent step can access both the original file (data) and the first analysis (content) at the same time. 📬 Contact Need help customizing this (e.g., filtering by campaign, sending reports by email, or formatting your PDF)? 📧 rbreen@ynteractive.com 🔗 https://www.linkedin.com/in/robert-breen-29429625/ 🌐 https://ynteractive.com
by Bohdan Saranchuk
This n8n template automates your customer support workflow by connecting Gmail, OpenAI, Supabase, and Slack. It listens for new incoming emails, classifies them using AI, routes them to the appropriate Slack channel based on category (e.g., support or new requests), logs each thread to Supabase for tracking, and marks the email as read once processed. Good to know • The OpenAI API is used for automatic email classification, which incurs a small per-request cost. See OpenAI Pricing for up-to-date info. • You can easily expand the categories or connect more Slack channels to fit your workflow. • The Supabase integration ensures you don’t process the same thread twice. How it works Gmail Trigger checks for unread emails. Supabase Get Row verifies if the thread already exists. If it’s a new thread, the OpenAI node classifies the email into categories such as “support” or “new-request.” The Switch node routes messages to the correct Slack channel based on classification. Supabase Create Row logs thread details (sender, subject, IDs) to your database. Finally, the Gmail node marks the message as read to prevent duplication. How to use • The workflow uses a manual Gmail trigger by default, but you can adjust the polling frequency. • Modify category names or Slack channels to match your internal setup. • Extend the workflow to include auto-replies or ticket creation in your CRM. Requirements • Gmail account (with OAuth2 credentials) • Slack workspace (with channel access) • OpenAI account for classification • Supabase project for storing thread data Customizing this workflow Use this automation to triage incoming requests, route sales leads to specific teams, or even filter internal communications. You can add nodes for auto-responses, CRM logging, or task creation in Notion or ClickUp.
by Wessel Bulte
Generate Weekly n8n Execution Report and Email Summary Description: How it works Automatically runs every 7 days to pull all n8n workflow executions from the past week Merges execution data with workflow information to provide context Generates a professional HTML report with execution statistics (errors, successes, waiting status) Sends the formatted report with Outlook or Gmail Set up steps 1. Configure n8n API Credential Go to your n8n instance → Settings → API Create a new API token with read access to workflows and executions In this workflow, add a new "n8n" credential and paste your API token This credential is used by two nodes: "Get all Workflows" and "Get all previous executions" 2. Connect Email Services Configure your Outlook credential in the "Send a message outlook" node Configure your Gmail credential in the "Send a message gmail" node Set your preferred email recipients in both nodes 3. Adjust Schedule (Optional) By default, the workflow runs every 7 days Edit the "Schedule Trigger" node to change the interval if needed Key features Tracks workflow execution status and runtime metrics Calculates average and total runtime for each status type Provides visual HTML report with color-coded status indicators Dual email delivery (Outlook + Gmail options) Requires only n8n API credentials (no external API keys needed) Need Help 🔗 LinkedIn – Wessel Bulte
by Ilyass Kanissi
📋Instant Proposal Generator Automatically convert sales call transcripts into professional client proposals by extracting key details with AI, dynamically populating Google Slides templates, and tracking progress in Airtable, all in one seamless workflow. 🎯 What does this workflow do? This end-to-end automation creates client-ready proposals by: Taking call transcripts via chat interface The AI analyzes the transcript to extract key details like company name, goals, budget, and requirements, then structures this data as JSON for seamless workflow integration. Generating customized documents using Google Slides template with dynamic variables, Auto populating {Company_Name}, {Budget}, etc. from extracted data. Delivering finished proposals: Sharing final document with client, and Updating CRM status automatically. ⚙️ How it works User input: Paste call transcript in chat trigger node AI analysis: OpenAI node processes text to extract structured JSON, Identifies company name, goals, budget, requirements, etc. Document copy: it copies the file from Google Drive, and name it {company name} proposal. Variables replacement: Replaces all template variables ({Company_Name}, {Budget}, etc.) with extracted data from ChatGPT. Delivery & tracking: Shares finalized proposal with client via email, an Updates Airtable "Lead Status" to "Proposal Sent". 🔑 Required setup OpenAI API Key: Create a key from here Google Cloud Credentials: Setup here Required scopes: Google Slides edit + file creation Airtable Access Token: Create one from here
by Mohammad
Telegram ticket intake and status tracking with Postgres Who’s it for Anyone running support requests through Telegram, Email, Webhooks, and so on who needs a lightweight ticketing system without paying Zendesk prices. Ideal for small teams, freelancers, or businesses that want tickets logged in a structured database (Postgres) and tracked automatically. I'm using Telegram since it's the most convenient one. How it works / What it does This workflow turns (Telegram) into a support desk: Receives new requests via a Telegram bot command. Creates a ticket in a Postgres database with a correlation ID, requester details, and status. Auto-confirms back to the requester with the ticket ID. Provides ticket updates (status changes, resolutions) when queried. Keeps data clean using dedupe keys and controlled input handling. How to set up Create a Telegram bot using @BotFather and grab the token. Connect your Postgres database to n8n and create a tickets table: CREATE TABLE tickets ( id BIGSERIAL PRIMARY KEY, correlation_id UUID, source TEXT, external_id TEXT, requester_name TEXT, requester_email TEXT, requester_phone TEXT, subject TEXT, description TEXT, status TEXT, priority TEXT, dedupe_key TEXT, chat_id TEXT, created_at TIMESTAMP DEFAULT NOW(), updated_at TIMESTAMP DEFAULT NOW() ); Add your Telegram and Postgres credentials in n8n (via the Credentials tab, not hardcoded). Import the workflow JSON and replace the placeholder credentials with yours. Test by sending /new in Telegram and follow the prompts. Requirements n8n (latest version recommended) Telegram bot token Postgres instance (local, Docker, or cloud) How to customize the workflow Change database fields if you need more requester info. Tweak the Switch node and Comands for multiple status types. Extend with Slack, Discord, or email nodes for broader notifications. Integrate with external systems (CRM, project management) by adding more branches.
by Rohit Dabra
Google Sheets → Stripe Payment Automation Workflow 📌 Overview This workflow automates the end-to-end process of generating and sending client payment links using Google Sheets and Stripe. Whenever a new or updated entry is added to the Google Sheet, the workflow will: Fetch client and invoice details. Create a Stripe Product and Price. Generate a Stripe Payment Link. Store the link back in the Google Sheet. Upload a copy of the invoice to Google Drive. Send a professional, formatted email with the payment link to the client using Gmail. 🔗 Demo Video: Watch on YouTube ⚡️ Workflow Steps Trigger – The workflow is triggered on any update in the Google Sheet. Filter – Ensures only relevant rows (e.g., PENDING invoices) proceed. Stripe Automation Create Stripe Product Create Stripe Price Generate Stripe Payment Link Google Drive – Store invoice files (if required). Google Sheets – Update the sheet with the generated Stripe Payment Link and timestamp. Gmail – Send a client-facing email with the invoice details and payment link. 🛠 Setup Guide Prerequisites n8n account** Google Sheets & Google Drive credentials** Gmail API credentials** Stripe API Key** Steps Clone/Import Workflow Import the workflow JSON file into your n8n instance. Configure Google Sheets Create a Google Sheet with columns: Order ID, Client Name, Client Email, Items Description, Due Date, Amount, Currency, Invoice Status, Invoice Link, Stripe Payment Link, Last Updated Connect your Google Sheets node to this sheet. Set Up Stripe Obtain your Stripe Secret Key from Stripe Dashboard. Add it in the Stripe nodes for Product, Price, and Payment Link creation. Google Drive Configure to store invoice backups (optional). Gmail Authorize Gmail and set up the Send Email node. Customize the email template with client details and the Stripe link. Test the Workflow Add a sample row in Google Sheets. Run the workflow manually or update the sheet to trigger automatically. Verify that the Stripe link is created, updated in the sheet, and emailed to the client. ✅ Now your workflow is ready to automatically manage client invoices and payments with Stripe + Google Sheets + Gmail + Google Drive.