by Yang
👥 Who is this for? This workflow is ideal for virtual assistants, researchers, developers, automation specialists, and data analysts who need to regularly extract and organize structured product information (like books) from a website. It’s especially useful for those working with catalog-based websites who want to automate extraction and delivery of clean, sorted data. 🧩 What problem is this solving? Manually copying product listings like book titles and prices from a website into a spreadsheet is slow and repetitive. This automation solves that problem by scraping content using Dumpling AI, extracting the right data using CSS selectors, and formatting it into a clean CSV file that is sent to your email—all triggered automatically when a new URL is added to Google Sheets. ⚙️ What this workflow does This template automates an entire content scraping and delivery process: Watches a Google Sheet for new URLs Scrapes the HTML content of the given webpage using Dumpling AI Uses CSS selectors in the HTML node to extract each book from the page Splits the HTML array into individual items Extracts the book title and price from each HTML block Sorts the books in descending order based on price Converts the sorted data to a CSV file Sends the CSV via email using Gmail 🛠️ Setup Google Sheets Create a sheet titled something like URLs Add your product listing URLs (e.g., http://books.toscrape.com) Connect the Google Sheets trigger node to your sheet Ensure you have proper credentials connected Dumpling AI Create an account at Dumpling AI) - Generate your API key Set the HTTP Method to POST and pass the URL dynamically from the Google Sheet Use Header Auth to include your API key in the request header Make sure "cleaned": "True" is included in the body for optimized HTML output HTML Node The first HTML node extracts the main book container blocks using: .row > li The second HTML node parses out the individual fields: title: h3 > a (via the title attribute) price: .price_color Sort Node Sorts books by price in descending order Note: price is extracted as a string, ensure it's parsable if you plan to use numeric filtering later Convert to CSV The JSON data is passed into a Convert node and transformed into a CSV file Gmail Sends the CSV as an attachment to a designated email 🔄 How to customize this workflow Extract more data**: Add more CSS selectors in the second HTML node to pull fields like author, availability, or product links Switch destinations**: Replace Gmail with Slack, Google Drive, Dropbox, or another platform Adjust sorting**: Sort alphabetically or based on another extracted value Use a different source**: As long as the site structure is consistent, this can scrape any listing-like page Trigger differently**: Use a webhook, form submission, or schedule trigger instead of Google Sheets ⚠️ Dependencies and Notes This workflow uses Dumpling AI to perform the web scraping. This requires an API key and uses credits per request. The HTML node depends on valid CSS selectors. If the site layout changes, the selectors may need to be updated. Ensure you’re not scraping content from websites that prohibit automated scraping.
by Paulo Ramirez
Upload your CRM contacts to telli and schedule AI voice-agent calls Introduction to telli and AI Voice-Agent Calls telli is an innovative platform that provides AI-powered voice agents capable of making calls and performing tasks tailored to specific customer use cases. These AI voice-agents can handle a wide range of communication tasks, from appointment scheduling to customer support, with remarkable efficiency and natural conversation flow. This template is designed for businesses and organizations looking to automate their outbound calling processes using telli's AI voice-agents in conjunction with Airtable as their CRM. It solves the problem of manual call scheduling and data transfer between your CRM and calling system, saving time and reducing human error. Prerequisites telli account Airtable base with contact information n8n instance Step-by-Step Setup Guide n8n Setup: Create a new workflow in n8n. Add the Airtable node to connect to your CRM table. telli API Configuration: Log in to your telli dashboard. Locate and copy your API key under telli - Settings - API/Webhooks. Workflow Configuration: Add two HTTP Request nodes to your n8n workflow. Set the "Authorization" header in both POST requests, replacing the value with your telli API key. Configure the first request to use the /add-contact endpoint. Set up the second request to use the /schedule-call endpoint. Data Mapping: Map the relevant fields from your Airtable node to the telli API requests. Testing and Activation: Run a test execution of your workflow. Once satisfied with the results, activate the workflow. API Endpoint Details Add Contact Endpoint URL**: https://api.telli.com/v1/add-contact Method**: POST Headers**: Authorization: YOUR-API-KEY Content-Type: application/json Payload**: { "external_contact_id": "string", "salutation": "string", "first_name": "string", "last_name": "string", "phone_number": "string", "email": "jsmith@example.com", "contact_details": {}, "timezone": "string" } Schedule Call Endpoint URL**: https://api.telli.com/v1/schedule-call Method**: POST Headers**: Authorization: YOUR-API-KEY Content-Type: application/json Payload**: { "contact_id": TELLI-CONTACT-ID, "agent_id": "string", "max_retry_days": 123, "call_details": { "message": "Hello, this is your friendly reminder!", "questions": [ { "fieldName": "email", "neededInformation": "email of the customer", "exampleQuestion": "What is your email address?", "responseFormat": "email string" } ] }, "override_from_number": "string" } Use Cases This template is versatile and can be applied to various scenarios, including: Lead Qualification*: Automatically schedule calls to new leads entered in your CRM. Appointment Reminders*: Set up calls to remind clients of upcoming appointments. Customer Feedback*: Schedule follow-up calls after product deliveries or service completions. Uploading Multiple Contacts For bulk operations, you have two options: Loop Node: Include a Loop node in your n8n workflow to process multiple contacts sequentially. Batch Endpoints: Instead of /add-contact and /schedule-call, use telli's batch endpoints: /add-contacts-batch: Add multiple contacts within an array. /schedule-calls-batch: Schedule multiple calls at once. Example of batch endpoint usage: { "contacts": [ {"name": "John Doe", "phone": "+1234567890"}, {"name": "Jane Smith", "phone": "+1987654321"} ] } By leveraging this template, you can seamlessly integrate your Airtable CRM with telli's powerful AI voice-agents, automating your outbound calling process and enhancing your customer communication strategy.
by JPres
👥 Who Is This For? Sales and marketing teams seeking efficient, hands‑free generation of personalized slide decks for each prospect from CSV lead lists. 🛠 What Problem Does This Solve? Manually editing presentation decks for large lead lists is slow and error‑prone. This workflow fully automates: Importing and parsing CSV lead data Logging leads and outputs in Google Sheets Duplicating a master Slides template per lead Injecting lead‑specific variables into slides 🔄 Node‑by‑Node Breakdown | Step | Node | Purpose | | ---- | ---------------------------------------- | -------------------------------------------------------- | | 1 | New Leads Arrived | Detect new CSV uploads in Drive | | 2 | File Type? | Filter for .csv files only | | 3 | Download by ID | Download the CSV content | | 4 | Create new Sheet | Create a Google Sheet to record lead data | | 5 | Combine Empty New Document with CSV Data | Structure each lead record for slide creation | | 6 | Merge Data for new Lead Document | Map template placeholders to lead values | | 7 | Get all Leads | Retrieve sheet rows to iterate through each lead | | 8 | MoveToLeadListFolder | Move processed CSV to an archive folder | | 9 | Copy Slides Template | Make a copy of the master Slides deck | | 10 | Create Custom Presentation | Replace placeholders in the copied deck with lead data | | 11 | Add Presentation ID to Lead | Write the generated presentation URL back into the Sheet | ⚙️ Pre‑conditions / Requirements n8n with Google Drive, Sheets, and Slides credentials A master Google Slides deck with placeholder tokens (e.g. {{Name}}, {{Company}}) A Drive folder for incoming CSV lead files ⚙️ Setup Instructions Import this workflow into your n8n instance. Configure the New Leads Arrived node to watch your CSV folder. Enter your Google credentials in the Drive, Sheets, and Slides nodes. Specify the master Slides template ID in the Copy Slides Template node. In Create Custom Presentation, map slide tokens to sheet column names. Disable “Keep Binary Data” in Copy Slides Template to conserve memory. Upload a sample CSV (with headers like Name, Company, Metric) to test. 🎨 How to Customize Add or remove variables by editing the CSV headers and updating the mapping in Merge Data for new Lead Document. Insert an AI/natural‑language node before slide creation to generate more advanced and personalized text blocks. Use SplitInBatches to throttle API calls and avoid rate‑limit errors. Add error‑handling branches to capture and log failed operations. 🔐 Security and Privacy The workflow uses placeholder variables for file and folder IDs, so no actual IDs are exposed in the template. Ensure OAuth scopes are limited to only the required Google APIs.
by JPres
👥 Who Is This For? Content creators, marketing teams, and channel managers who want a simple, hands‑off solution to upload videos and automatically generate optimized metadata from video transcripts. 🛠 What Problem Does This Solve? Manual video uploads with proper metadata creation is time‑consuming and repetitive. This workflow fully automates: Monitoring a specific Google Drive folder for new video uploads Seamless YouTube upload processing Transcript extraction for context understanding AI‑powered generation of titles, descriptions, and tags Metadata application to uploaded videos without manual intervention 🔄 Node‑by‑Node Breakdown | Step | Node Purpose | |------|---------------------------------------------------------------------| | 1 | New Video? (Trigger) – Monitors specified Google Drive folder | | 2 | Download New Video – Retrieves the video file from Google Drive | | 3 | Upload to YouTube – Uploads the video to YouTube with initial settings | | 4 | Get Transcript – Extracts transcript from the uploaded video | | 5 | Adjust Transcript Format – Formats raw transcript for processing | | 6 | Create Description – Generates SEO‑optimized description | | 7 | YT Tags (Message Model) – Creates relevant tags based on content | | 8 | YT Title (Message Model) – Generates compelling title | | 9 | Define File Path Upload Format (Optional) – Structures data paths | | 10 | Update Video’s Metadata – Applies generated title, description, tags| ⚙️ Pre‑conditions / Requirements n8n with Google Drive and YouTube API credentials configured (stored as n8n credentials/variables; no hard‑coded IDs) Dedicated Google Drive folder for video uploads YouTube channel with proper upload permissions AI service access for transcript processing and metadata generation Sufficient storage for temporary video handling ⚙️ Setup Instructions Import this workflow into your n8n instance. Configure Google Drive credentials; reference folder ID via n8n variable (do not hard‑code). Set up YouTube API credentials with upload and edit permissions. Specify the target Google Drive folder ID in the New Video? trigger node (via variable). Configure AI service credentials for transcript and metadata generation. Adjust message templates for title, description, and tag creation. Test with a small video file before production use. 🎨 How to Customize Modify AI prompts to match your channel’s tone and style. Add conditional logic based on video categories or naming conventions. Implement notification systems to alert when uploads complete. Create custom metadata templates for different content types. Include timestamps or chapter markers based on transcript analysis. Add social media sharing nodes to announce new uploads. ⚠️ Important Notes Video quality is preserved through the upload process. Consider YouTube API quotas when handling multiple uploads. Transcript quality affects metadata generation results. Videos are initially uploaded without visibility adjustments. Processing time depends on video length and transcript complexity. 🔐 Security and Privacy Store API credentials and folder IDs as n8n Credentials/Variables—remove any hard‑coded tokens or IDs. Video files are processed temporarily and not stored permanently. Limit Google Drive folder access to authorized users only. Manage YouTube upload permissions carefully (use OAuth/service accounts). Ensure compliance with organizational data‑handling policies.
by Yang
Who is this for? This template is for sales teams, agencies, or local service providers who want to quickly generate cold outreach lists and automatically call local businesses with a Vapi AI assistant. It’s perfect for automating cold calls from scraped local listings with no manual dialing or research. What problem is this workflow solving? Finding leads and initiating outreach calls can be time-consuming. This workflow automates the process: it scrapes business listings from Google Maps using Dumpling AI, extracts phone numbers, filters out incomplete data, formats the numbers, and uses Vapi to make outbound AI-powered calls. Every call is logged in Google Sheets for follow-up and tracking. What this workflow does Starts manually and pulls search queries (e.g., "plumbers in Austin") from Google Sheets. Sends each query to Dumpling AI’s Google Maps scraping endpoint. Splits the returned business data into individual leads. Extracts key info like business name, website, and phone number. Filters to only keep leads with valid phone numbers. Formats phone numbers for Vapi dialing (adds +1). Calls each business using Vapi AI. Logs each successful call in a Google Sheet. Setup Google Sheets Setup Create a sheet with business search queries in the first column (e.g., best+restaurants+in+Chicago) Make sure the tab name is set and authorized in your credentials. Connect your Google Sheets account in the Get Search Keywords from Google Sheets node. Dumpling AI Setup Go to dumplingai.com Generate an API Key and connect it as a header token in the Scrape Google Map Businesses using Dumpling AI node Vapi Setup Sign into Vapi and create an assistant Get your assistantId and phoneNumberId Insert these into the JSON payload of the Initiate Vapi AI Call to Business node Add your Vapi API key to the credentials section Call Logging Create another tab in your sheet (e.g., “leads”) with these headers: company name phone number website This will be used in the Log Called Business Info to Sheet node How to customize this workflow to your needs Modify the business search terms in your Google Sheet to target specific industries or locations. Add filters to exclude certain businesses based on ratings, keywords, or location. Update your Vapi assistant script to match the type of outreach or pitch you’re using. Add additional integrations (e.g., CRM logging, Slack notifications, follow-up emails). Change the trigger to run on a schedule or webhook instead of manually. Nodes and Functions Breakdown Start Workflow Manually: Initiates the automation manually for testing or controlled runs. Get Search Keywords from Google Sheets: Reads search phrases from the spreadsheet. Scrape Google Map Businesses using Dumpling AI: Sends each search query to Dumpling AI and receives matching local business data. Split Each Business Result: Breaks the returned array of businesses into individual records for processing. Extract Business Name, Phone and website: Extracts title, phone, and website from each business record. Filter Valid Phone Numbers Only: Ensures only entries with a phone number move forward. Format Phone Number for Calling: Adds a +1 country code and strips non-numeric characters. Initiate Vapi AI Call to Business: Uses the business name and number to initiate a Vapi AI outbound call. Log Called Business Info to Sheet: Appends business details into a Google Sheet for tracking. Notes You must have valid API keys and authorized connections for Dumpling AI, Google Sheets, and Vapi. Make sure to handle API rate limits if you're running the workflow on large datasets. This workflow is optimized for US-based leads (+1 country code); adjust the formatting node if calling internationally.
by Don Jayamaha Jr
🕒 Evaluate Tesla (TSLA) price action and market structure on the 1-hour timeframe using 6 real-time indicators. This sub-agent is designed to feed mid-term technical insights into the Tesla Financial Market Data Analyst Tool. It uses GPT-4.1 to interpret Alpha Vantage indicator data delivered via secure webhooks. ⚠️ This workflow is not standalone and is executed via Execute Workflow. 🔌 Requires: Tesla Quant Technical Indicators Webhooks Tool Alpha Vantage Premium API Key 🔧 Connected Indicators This tool fetches and analyzes the latest 20 datapoints for: RSI (Relative Strength Index)** MACD (Moving Average Convergence Divergence)** BBANDS (Bollinger Bands)** SMA (Simple Moving Average)** EMA (Exponential Moving Average)** ADX (Average Directional Index)** 📋 Sample Output { "summary": "TSLA is gaining strength on the 1-hour chart. RSI is rising, MACD has crossed bullish, and BBANDS are widening.", "timeframe": "1h", "indicators": { "RSI": 62.1, "BBANDS": { "upper": 176.90, "lower": 169.70, "middle": 173.30, "close": 176.30 }, "SMA": 174.20, "EMA": 175.60, "ADX": 27.5, "MACD": { "macd": 0.84, "signal": 0.65, "histogram": 0.19 } } } 🧠 Agent Components | Component | Role | | ------------------------------ | -------------------------------------------------- | | 1hour Data | Pulls Alpha Vantage indicator data via webhook | | Tesla 1hour Indicators Agent | Interprets signals using structured GPT-4.1 prompt | | OpenAI Chat Model | GPT-4.1 LLM performs analysis | | Simple Memory | Maintains session context | 🛠️ Setup Instructions Import Workflow into n8n Name it: Tesla_1hour_Indicators_Tool Install the Webhook Fetcher Tool 👉 Required: Tesla_Quant_Technical_Indicators_Webhooks_Tool This agent expects webhook /1hourData to return pre-cleaned data Add Credentials Alpha Vantage Premium API Key (via HTTP Query Auth) OpenAI GPT-4.1 credentials Configure for Sub-Agent Use Triggered only via Execute Workflow from: 👉 Tesla Financial Market Data Analyst Tool Inputs: message (optional) sessionId (required for memory linkage) 📌 Sticky Notes Overview 🟢 Trigger Setup – Activated only by the parent agent 📊 1h Webhook Fetcher – Calls Alpha Vantage via secured endpoint 🧠 AI Agent Summary – Interprets trend/momentum from indicator data 🔗 GPT Model Notes – GPT-4.1 parses and explains technical alignment 📘 Documentation Sticky – Embedded in canvas with full walkthrough 🔐 Licensing & Support © 2025 Treasurium Capital Limited Company This tool is part of a proprietary multi-agent AI architecture. No commercial reuse or redistribution permitted. 🔗 Author: Don Jayamaha 🔗 Templates: https://n8n.io/creators/don-the-gem-dealer/ 🚀 Detect TSLA trend shifts and validate setups with 1-hour technical clarity—powered by Alpha Vantage + GPT-4.1. This tool is required by the Tesla Financial Market Data Analyst Tool.
by Don Jayamaha Jr
📅 Analyze Tesla’s daily trading structure with AI using 6 Alpha Vantage indicators. This tool evaluates long-term trend health, volatility patterns, and potential reversal signals at the 1-day timeframe. Designed for use within the Tesla Financial Market Data Analyst Tool, this agent helps swing and position traders anchor macro sentiment. ⚠️ Not standalone. Must be executed via Execute Workflow 🔌 Requires: Tesla Quant Technical Indicators Webhooks Tool Alpha Vantage Premium API Key OpenAI GPT-4.1 credentials 🔍 What It Does This tool queries a secured webhook (/1dayData) to retrieve real-time, trimmed JSON data for: RSI (Relative Strength Index)** BBANDS (Bollinger Bands)** SMA (Simple Moving Average)** EMA (Exponential Moving Average)** ADX (Average Directional Index)** MACD (Moving Average Convergence Divergence)** These values are then passed to a LangChain AI Agent powered by GPT-4.1, which returns: A 2–3 sentence market condition summary Structured indicator values Timeframe tag ("1d") 📋 Sample Output { "summary": "TSLA shows consolidation on the daily chart. RSI is neutral, BBANDS are contracting, and MACD is flattening.", "timeframe": "1d", "indicators": { "RSI": 51.3, "BBANDS": { "upper": 192.80, "lower": 168.20, "middle": 180.50, "close": 179.90 }, "SMA": 181.10, "EMA": 179.75, "ADX": 15.8, "MACD": { "macd": -0.25, "signal": -0.20, "histogram": -0.05 } } } 🧠 Agent Components | Component | Description | | ----------------------------- | -------------------------------------------------- | | 1day Data (HTTP Node) | Pulls latest data from secured /1dayData webhook | | OpenAI Chat Model | GPT-4.1 powers the analysis logic | | Tesla 1day Indicators Agent | LangChain agent performing interpretation | | Simple Memory | Short-term session continuity | 🛠️ Setup Instructions Import Workflow into n8n Name: Tesla_1day_Indicators_Tool Add Required Credentials Alpha Vantage Premium (via HTTP Query Auth) OpenAI GPT-4.1 (Chat Model) Install Webhook Fetcher Required: Tesla Quant Technical Indicators Webhooks Tool Endpoint /1dayData must be active Execution Context This tool is only triggered via: 👉 Tesla Financial Market Data Analyst Tool Inputs expected: message: optional context sessionId: session memory linkage 📌 Sticky Notes Overview 📘 Tesla 1-Day Indicators Tool – Purpose and integration 📡 Webhook Fetcher – Pulls daily Alpha Vantage data via HTTPS 🧠 GPT-4.1 Model – Reasoning for trend classification 🔗 Sub-Agent Trigger – Used only by Financial Market Analyst 🧠 Memory Buffer – Ensures consistent session logic 🔒 Licensing & Support © 2025 Treasurium Capital Limited Company This workflow—including prompts, logic, and formatting—is protected IP. 🔗 Don Jayamaha – LinkedIn 🔗 Creator Profile 🚀 Evaluate long-term Tesla price behavior with AI-enhanced technical analysis—critical for swing trading strategy. Required by the Tesla Financial Market Data Analyst Tool.
by Jimleuk
This n8n template demonstrates one approach to customer authentication via chat agents. Unlike approaches where you have to authenticate users prior to interacting with the agent, this approach allows guest users to authenticate at any time during the session or not at all. Note about Security: this template is for illustration purposes only and requires much more work to be ready for production! How it works A conversational agent is used for this demonstration. The key component is the Redis node just after the chat trigger which acts as the session context. For guests, the session item is blank. for customers, the session item is populated with their customer profile. The agent is instructed to generate a unique login URL only for guests when appropriate or upon request. This login URL redirects the guest user to a simple n8n form also hosted in this template. The login URL has the current sessionID as a query parameter as the way to pass this data to the form. Once login is successful, the matching session item by sessionId is populated with the customer profile. The user can now return to the chat window. Back to the agent, now when the user sends their next message, the Redis node will pick up the session item and the customer profile associated with it. The system prompt is updated with this data which let's the agent know the user is now a customer. How to use You'll need to update the "auth URL" tool to match the URL of your n8n instance. Better yet, copy the production URL of your form from the trigger. Activate the workflow to turn on production mode which is required for this workflow. Implement the authentication logic in step 3. This could be sending the user and pass to a postgreSQL data for validation. Requirements OpenAI for LLM (feel free to swap to any provider) Redis for Cache/Sessions (again, feel free to swap this out for postgresql or other database) Customising this workflow Consider not populating the session item with the user data as it can become stale. Instead, just add the userId and instruct the agent to query using tools. Extend the Login URL idea by experimenting with signup URLs or single-use Urls.
by bangank36
This workflow backup Squarespace website header and footer injections into Github How It Works The Squarespace injections are fetched when an URL is placed Setup Instructions First, edit HTTP Request's URL to put your Squarespace site URL there Next, to configure the Github, update the Globals node with the following values: repo.owner – Your GitHub username repo.name – The name of your GitHub repository storing the workflows repo.path – The folder path within the repository where workflows are stored For example, if your GitHub username is john-doe, your repository is named n8n-backups, and injections are stored in a squarespace-backup/ folder, you would set: repo.owner → john-doe repo.name → n8n-backups repo.path → squarespace-backup/ Each site's injections will be added into seperate folder Required Credentials GitHub API – Access to your repository Who Is This For? This template is made for Squarespace users who want to backup their header and footer injections at interval to or on demand Check out my other templates: 👉 My n8n Templates
by Airtop
Recursive Web Scraping Use Case Automating web scraping with recursive depth is ideal for collecting content across multiple linked pages—perfect for content aggregation, lead generation, or research projects. What This Automation Does This automation reads a list of URLs from a Google Sheet, scrapes each page, stores the content in a document, and adds newly discovered links back to the sheet. It continues this process for a specified number of iterations based on the defined scraping depth. Input Parameters: Seed URL: The starting URL to begin the scraping process. Example: https://example.com/ Links must contain: Restricts the links to those that contain this specified string. Example: https://example.com/ Depth: The number of iterations (layers of links) to scrape beyond the initial set. Example: 3 How It Works Starts by reading the Seed URL from the Google Sheet. Scrapes each page and saves its content to the specified document. Extracts new links from each page that match the Links must contain string, appends them to the Google Sheet. Repeats steps 2–3 for the number of times specified by Depth - 1. Setup Requirements Airtop API Key — free to generate. Credentials set up for Google Docs (requires creating a project on Google Console). Read how to. Credentials set up for Google Spreadsheet. Next Steps Add Filtering Rules**: Filter which links to follow based on domain, path, or content type. Combine with Scheduler**: Run this automation on a schedule to continuously explore newly discovered pages. Export Structured Data**: Extend the process to store extracted data in a CSV or database for analysis. Read more about website scraping for LLMS
by Yaron Been
🤖 AI Cart Recovery Agent: Smart Abandoned Checkout Assistant Transform abandoned carts into recovered sales with intelligent automation. This sophisticated n8n workflow monitors checkout abandonment, implements smart waiting periods, and sends AI-generated personalized recovery emails only when needed - maximizing conversions while respecting customer experience. 🔄 How It Works This intelligent 7-step recovery system recovers lost sales automatically: Step 1: Initial Abandonment Detection The workflow fetches current abandoned checkout data from your e-commerce platform (Shopify, WooCommerce, etc.), identifying customers who added items but didn't complete their purchase. Step 2: Strategic Grace Period Instead of immediately sending recovery emails, the system waits 1 hour (customizable), giving customers natural time to complete their purchase without pressure or interruption. Step 3: Smart Re-verification After the waiting period, the workflow rechecks the abandonment status by fetching updated checkout data, ensuring accuracy before taking action. Step 4: Intelligent Decision Logic Advanced conditional logic compares initial and updated abandonment lists, determining if customers are still abandoned or have completed their purchase during the grace period. Step 5: AI-Powered Email Generation For customers still showing abandonment, GPT generates personalized recovery emails featuring: Customer's actual name for personal connection Specific products left in their cart Friendly, non-pushy messaging tone Optional discount incentives Compelling call-to-action to complete purchase Step 6: Automated Email Delivery Personalized recovery emails are sent directly to abandoned customers via Gmail or your preferred email service, maintaining professional branding and deliverability. Step 7: Comprehensive Activity Logging All recovery attempts are logged in Google Sheets for tracking, including customer details, email content, and campaign performance analytics. ⚙️ Setup Steps Prerequisites E-commerce platform with API access (Shopify, WooCommerce, BigCommerce) OpenAI API key for personalized email generation Gmail or SMTP email service for delivery Google Sheets for activity tracking and analytics n8n instance (cloud or self-hosted) E-commerce Platform Configuration Shopify Setup: API Endpoint: https://your-store.myshopify.com/admin/api/2023-10/checkouts.json Authentication: X-Shopify-Access-Token header Required Permissions: Read checkouts, Read customers Parameters: status=abandoned WooCommerce Setup: API Endpoint: https://your-site.com/wp-json/wc/v3/orders Authentication: Consumer Key/Secret or JWT Parameters: status=pending, status=failed Required Plugins: WooCommerce REST API Configuration Steps 1. Credential Setup E-commerce API**: Store admin API access tokens or keys OpenAI API Key**: GPT-4 access for intelligent email generation Gmail OAuth2**: Professional email delivery service Google Sheets OAuth2**: Activity logging and performance tracking 2. Abandonment Detection Configuration Monitoring Frequency**: Set workflow trigger schedule (hourly, daily) Grace Period Duration**: Customize wait time (default: 1 hour) Platform Integration**: Configure API endpoints for your specific platform Data Filtering**: Set criteria for what constitutes abandonment 3. AI Email Customization Default email generation includes: Personalization Level**: Customer name, product specifics, cart value Tone Customization**: Friendly, urgent, helpful, or premium Discount Integration**: Optional percentage or fixed amount offers Brand Voice**: Maintain consistent company messaging and style 4. Recovery Campaign Settings Email Timing**: Optimal sending times based on customer time zones Frequency Limits**: Prevent over-emailing with cooldown periods Segmentation Rules**: Different approaches for high-value vs standard carts Follow-up Sequences**: Multi-email recovery campaigns with escalating incentives 5. Performance Tracking Setup Analytics Dashboard**: Google Sheets with recovery metrics and ROI Success Tracking**: Monitor completion rates and revenue recovered A/B Testing**: Compare different email approaches and timing Customer Journey**: Track from abandonment through recovery completion 🚀 Use Cases E-commerce Retailers Fashion & Apparel**: Recover high-value clothing and accessory purchases Electronics**: Target abandoned tech purchases with technical support offers Home & Garden**: Remind customers about seasonal or home improvement items Beauty & Cosmetics**: Recover abandoned skincare and makeup purchases Subscription & SaaS Businesses Software Trials**: Convert abandoned trial signups into paid subscriptions Membership Sites**: Recover incomplete membership purchases Online Courses**: Re-engage learners who abandoned course purchases Digital Services**: Follow up on abandoned service bookings or consultations B2B E-commerce Office Supplies**: Recover bulk order abandonments with volume discounts Industrial Equipment**: Follow up on high-value equipment quote requests Professional Services**: Re-engage businesses that abandoned service bookings Software Licenses**: Recover enterprise software purchase abandonments Specialty Retailers Luxury Goods**: Provide white-glove service for high-value abandoned purchases Custom Products**: Follow up on personalized or custom order abandonments Seasonal Items**: Time-sensitive recovery for holiday or event-specific products Limited Edition**: Create urgency for exclusive or limited availability items Service-Based Businesses Travel & Hospitality**: Recover abandoned hotel, flight, or package bookings Event Tickets**: Re-engage customers who abandoned concert or event purchases Professional Services**: Follow up on abandoned consultation or service bookings Fitness & Wellness**: Recover abandoned membership or class package purchases 🔧 Advanced Customization Options Multi-Platform Integration Extend beyond single platform monitoring: Shopify Plus: Advanced checkout analytics and customer segmentation WooCommerce: Custom post-purchase and abandonment tracking Magento: Enterprise-level cart recovery with customer journey mapping BigCommerce: API-driven recovery with advanced personalization Custom Platforms: Webhook-based abandonment detection and recovery Intelligent Email Sequencing Create sophisticated recovery campaigns: Progressive Incentives**: Escalating discounts over multiple touchpoints Behavioral Triggers**: Different emails based on cart value, customer history Seasonal Campaigns**: Holiday-specific recovery messaging and offers Win-Back Sequences**: Long-term customer re-engagement beyond immediate recovery Advanced Personalization Enhance AI-generated content with: Purchase History Analysis**: Reference previous purchases and preferences Browsing Behavior**: Include recently viewed items and categories Geographic Personalization**: Local offers, shipping options, or store locations Demographic Targeting**: Age, gender, or interest-based messaging customization Performance Optimization Implement advanced tracking and optimization: Revenue Attribution**: Track exact recovery amounts and ROI calculations Customer Lifetime Value**: Prioritize high-value customer recovery efforts Conversion Funnel Analysis**: Identify optimal timing and messaging strategies Predictive Analytics**: Use ML to predict recovery likelihood and optimize approaches 📊 Recovery Email Examples Fashion Retailer Example: Subject: You left something stylish behind, Sarah! Hi Sarah, I noticed you were checking out those gorgeous items in your cart earlier - the Bohemian Summer Dress and Classic Leather Handbag have been waiting for you! I completely understand if you got busy or needed time to think it over. These pieces are still available and ready to ship to you today. Since you showed such great taste in selecting these items, I'd love to offer you 10% off your order to make the decision easier. Just use code WELCOME10 at checkout. Your cart includes: • Bohemian Summer Dress (Size M) - $89.99 • Classic Leather Handbag (Brown) - $156.99 Complete your purchase now and get free shipping to your door! [Complete My Purchase] Best regards, The StyleHub Team P.S. These items are popular and inventory is limited - don't wait too long! Software/SaaS Example: Subject: Your ProductivityPro trial is waiting, Mike Hi Mike, You were just one step away from unlocking the full power of ProductivityPro for your team at TechStartup Inc. I noticed you explored our Premium Plan features - the advanced reporting and team collaboration tools that could streamline your workflow and boost productivity by up to 40%. Since you invested time exploring our platform, I'd like to offer you an exclusive 25% discount on your first year. This offer is valid for the next 48 hours. Your selected plan: • ProductivityPro Premium (5 users) - $99/month • With 25% discount: $74/month (Save $300/year!) Ready to transform your team's productivity? [Activate My Account] Questions? Reply to this email or schedule a quick 15-minute demo call. Best regards, David Chen Customer Success Manager, ProductivityPro High-Value B2B Example: Subject: Your equipment quote is ready for approval, Jennifer Hi Jennifer, Thank you for your interest in our Industrial Packaging System for ManuCorp's new facility expansion. I understand that equipment investments of this scale require careful consideration and stakeholder alignment. Your configured system includes: • Model X5000 Packaging Line - $45,000 • Installation & Training Package - $8,000 • Extended 3-Year Warranty - $3,500 Total Investment: $56,500 Given the scope of your project, I'd like to extend our Q1 promotion pricing, which provides: 15% discount on equipment ($6,750 savings) Free installation supervision ($2,000 value) Expedited 6-week delivery This brings your total to $48,750 - a savings of $7,750. I'm available for a brief call to address any technical questions or help facilitate internal approvals. [Accept Quote & Proceed] Best regards, Robert Martinez Senior Sales Engineer Industrial Solutions Inc. Direct: (555) 123-4567 🛠️ Troubleshooting & Best Practices Common Issues & Solutions API Rate Limiting Implement exponential backoff for API requests Stagger workflow execution times across different stores Monitor API usage and upgrade plans as needed Cache frequently accessed data to reduce API calls Email Deliverability Challenges Use authenticated SMTP services with proper SPF/DKIM setup Monitor sender reputation and email engagement metrics Implement opt-out mechanisms and respect unsubscribe requests Segment email lists and avoid over-emailing customers False Positive Recoveries Extend grace periods for complex checkout processes Implement more sophisticated abandonment detection logic Add customer behavior analysis before triggering recovery Create exception rules for technical checkout failures Optimization Strategies Recovery Timing Optimization A/B test different grace period durations (30 min, 1 hour, 3 hours) Analyze customer behavior patterns to optimize sending times Consider time zone differences for global customer bases Implement seasonal timing adjustments for holidays and events Content Personalization Enhancement Continuously refine AI prompts based on successful recoveries Implement dynamic discount strategies based on cart value Create customer segment-specific messaging approaches Add urgency elements for time-sensitive or limited inventory items Performance Measurement Track recovery rates, revenue impact, and customer satisfaction Implement cohort analysis for long-term customer value impact Monitor email engagement metrics and optimize accordingly Calculate true ROI including customer acquisition costs and lifetime value 📈 Success Metrics Recovery Performance Indicators Recovery Rate**: Percentage of abandoned carts successfully recovered Revenue Recovery**: Total dollar amount recovered from abandoned purchases Email Engagement**: Open rates, click rates, and conversion rates Time to Recovery**: Average time from abandonment to completed purchase Business Impact Measurements ROI Calculation**: Revenue recovered vs workflow operational costs Customer Retention**: Impact on long-term customer relationships Average Order Value**: Effect on overall purchase values post-recovery Operational Efficiency**: Automation savings vs manual recovery efforts 📞 Questions & Support Need help implementing your AI Cart Recovery Agent? 📧 E-commerce Automation Expert Support Email**: Yaron@nofluff.online Response Time**: Within 24 hours on business days Specialization**: E-commerce automation, cart recovery optimization, AI email personalization 🎥 Comprehensive Implementation Resources YouTube Channel**: https://www.youtube.com/@YaronBeen/videos Complete setup guides for major e-commerce platforms Advanced AI email personalization techniques Recovery campaign optimization strategies Integration tutorials for Shopify, WooCommerce, and custom platforms Performance tracking and analytics implementation 🤝 E-commerce Automation Community LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Connect for ongoing e-commerce automation support and consulting Share your cart recovery success stories and ROI achievements Access exclusive templates for different industry verticals Join discussions about e-commerce automation trends and innovations 💬 Support Request Guidelines Include in your support message: Your e-commerce platform and current cart abandonment rates Average order values and customer segments you serve Current recovery processes and conversion challenges Integration requirements with existing marketing tools Specific technical errors or workflow execution issues
by Dvir Sharon
Automated Content Idea Generation and Expansion with Google Gemini and Google Sheets This n8n workflow automates the process of generating content ideas based on a user-defined topic, then expands each idea into a more detailed content piece (like a blog post) using Google Gemini, and finally saves all the generated data (idea title, description, and full content) into a Google Sheet. It's a powerful tool for streamlining content creation workflows. This workflow includes: Generation of multiple content ideas from a single topic. Expansion of each idea into detailed content using AI. Storage of ideas and generated content in a structured Google Sheet. Sticky Notes within the workflow for inline documentation and setup guidance. Prerequisites n8n Instance: You need a running n8n instance (self-hosted or cloud). Google AI Account: Access to Google AI (Gemini). You will need an API key. Google Account: Access to Google Sheets. You will need to create or use an existing spreadsheet with specific column headers. Installation and Setup Import the Workflow: Copy the entire JSON code provided. In your n8n instance, go to "Workflows". Click "New" -> "Import from JSON". Paste the JSON code and click "Import". Configure Credentials: Google AI (Gemini): Find the "Google Gemini Chat Model for Content Idea Generator" node and the "Google Gemini Chat Model for Content Generation" node. Click on the "Credentials" field in both nodes (it will likely show a placeholder name like "Google Gemini(PaLM) Api account"). Click "Create New". Select "Google AI API". Enter your Google AI API Key. Save the credential. (You can reuse the same credential for both nodes). Google Sheets: Find the "Google Sheets" node. Click on the "Credentials" field (it will likely show a placeholder name like "Google Sheets account"). Click "Create New". Select "Google Sheets OAuth2 API". Follow the steps to connect your Google Account and grant n8n access to Google Sheets. Save the credential. Configure Google Sheets Node: Open the "Google Sheets" node settings. Spreadsheet ID: Replace the placeholder value with the actual ID of your Google Sheet. You can find the Spreadsheet ID in the URL of your Google Sheet (it's the long string of characters between /d/ and /edit). Sheet Name: Select or enter the name or GID of the sheet within your spreadsheet where you want to save the data (e.g., Sheet1 or gid=0). Columns: Ensure your Google Sheet has columns named title, description, and content. The node is configured to map the generated data to these specific column headers. Save the node settings. Review Sticky Notes: Look at the Sticky Notes placed around the workflow canvas. They provide helpful context and reminders for setup, required Google Sheet columns, and the AI models used. How to Use Activate the Workflow: Toggle the workflow switch to "Active". Trigger the Workflow: Since this workflow uses a "When clicking ‘Execute workflow’" node as the trigger, you can run it directly from the n8n editor. Click the "Execute Workflow" button. The workflow will start automatically. Set the Topic: Open the "Set the input fields" node. Modify the topic value to the subject you want to generate content ideas about. Save the node settings. Monitor Execution: Watch the workflow execute. The nodes will light up as they process. The "Loop Over Items" node will show multiple executions as it processes each generated idea. Check Results: The generated content ideas (title, description) and the expanded content will be written as new rows in the Google Sheet you configured. Each row will correspond to one generated idea and its content. This workflow provides a robust starting point for AI-assisted content creation. You can customize the AI prompts in the "Content Idea Generator" and "LLM Content Generator" nodes to refine the output style and format, or integrate additional steps like sending notifications or further processing the generated content.