by Incrementors
🛒 Lead Workflow: Yelp & Trustpilot Scraping + OpenAI Analysis via BrightData > Description: Automated lead generation workflow that scrapes business data from Yelp and Trustpilot based on location and category, analyzes credibility, and sends personalized outreach emails using AI. > ⚠️ Important: This template requires a self-hosted n8n instance to run. 📋 Overview This workflow provides an automated lead generation solution that identifies high-quality prospects from Yelp and Trustpilot, analyzes their credibility through reviews, and sends personalized outreach emails. Perfect for digital marketing agencies, sales teams, and business development professionals. ✨ Key Features 🎯 Smart Location Analysis** AI breaks down cities into sub-locations for comprehensive coverage 🛍 Yelp Integration** Scrapes business details using BrightData's Yelp dataset ⭐ Trustpilot Verification** Validates business credibility through review analysis 📊 Data Storage** Automatically saves results to Google Sheets 🤖 AI-Powered Outreach** Generates personalized emails using Claude AI 📧 Automated Sending** Sends emails directly through Gmail integration 🔄 How It Works User Input: Submit location, country, and business category through a form AI Location Analysis: Gemini AI identifies sub-locations within the specified area Yelp Scraping: BrightData extracts business information from multiple locations Data Processing: Cleans and stores business details in Google Sheets Trustpilot Verification: Scrapes reviews and company details for credibility check Email Generation: Claude AI creates personalized outreach messages Automated Outreach: Sends emails to qualified prospects via Gmail 📊 Data Output | Field | Description | Example | |---------------|----------------------------------|----------------------------------| | Company Name | Business name from Yelp/Trustpilot | Best Local Restaurant | | Website | Company website URL | https://example-restaurant.com | | Phone Number | Business contact number | (555) 123-4567 | | Email | Business email address | demo@example.com | | Address | Physical business location | 123 Main St, City, State | | Rating | Overall business rating | 4.5/5 | | Categories | Business categories/tags | Restaurant, Italian, Fine Dining | 🚀 Setup Instructions ⏱️ Estimated Setup Time: 10–15 minutes Prerequisites n8n instance (self-hosted or cloud) Google account with Sheets access BrightData account with Yelp and Trustpilot datasets Google Gemini API access Anthropic API key for Claude Gmail account for sending emails Step 1: Import the Workflow Copy the JSON workflow code In n8n: Workflows → + Add workflow → Import from JSON Paste JSON and click Import Step 2: Configure Google Sheets Integration Create two Google Sheets: Yelp data: Name, Categories, Website, Address, Phone, URL, Rating Trustpilot data: Company Name, Email, Phone Number, Address, Rating, Company About Copy Sheet IDs from URLs In n8n: Credentials → + Add credential → Google Sheets OAuth2 API Complete OAuth setup and test connection Update all Google Sheets nodes with your Sheet IDs Step 3: Configure BrightData Set up BrightData credentials in n8n Replace API token with: BRIGHT_DATA_API_KEY Verify dataset access: Yelp dataset: gd_lgugwl0519h1p14rwk Trustpilot dataset: gd_lm5zmhwd2sni130p Test connections Step 4: Configure AI Models Google Gemini (Location Analysis)** Add Google Gemini API credentials Configure model: models/gemini-1.5-flash Claude AI (Email Generation)** Add Anthropic API credentials Configure model: claude-sonnet-4-20250514 Step 5: Configure Gmail Integration Set up Gmail OAuth2 credentials in n8n Update "Send Outreach Email" node Test email sending Step 6: Test & Activate Activate the workflow Test with sample data: Country: United States Location: Dallas Category: Restaurants Verify data appears in Google Sheets Check that emails are generated and sent 📖 Usage Guide Starting a Lead Generation Campaign Access the form trigger URL Enter your target criteria: Country: Target country Location: City or region Category: Business type (e.g., restaurants) Submit the form to start the process Monitoring Results Yelp Data Sheet:** View scraped business information Trustpilot Sheet:** Review credibility data Gmail Sent Items:** Track outreach emails sent 🔧 Customization Options Modifying Email Templates Edit the "AI Generate Email Content" node to customize: Email tone and style Services mentioned Call-to-action messages Branding elements Adjusting Data Filters Modify rating thresholds Set minimum review counts Add geographic restrictions Filter by business size Scaling the Workflow Increase batch sizes Add delays between requests Use parallel processing Add error handling 🚨 Troubleshooting Common Issues & Solutions 1. BrightData Connection Failed Cause: Invalid API credentials or dataset access Solution: Verify credentials and dataset permissions 2. No Data Extracted Cause: Invalid location or changed page structure Solution: Verify location names and test other categories 3. Gmail Authentication Issues Cause: Expired OAuth tokens Solution: Re-authenticate and check permissions 4. AI Model Errors Cause: API quota exceeded or invalid keys Solution: Check usage limits and API key Performance Optimization Rate Limiting:** Add delays Error Handling:** Retry failed requests Data Validation:** Check for malformed data Memory Management:** Process in smaller batches 📈 Use Cases & Examples 1. Digital Marketing Agency Lead Generation Goal:** Find businesses needing marketing Target:** Restaurants, retail stores Approach:** Focus on good-rated but low-online-presence businesses 2. B2B Sales Prospecting Goal:** Find software solution clients Target:** Growing businesses Approach:** Focus on recent positive reviews 3. Partnership Development Goal:** Find complementary businesses Target:** Established businesses Approach:** Focus on reputation and satisfaction scores ⚡ Performance & Limits Expected Performance Processing Time:** 5–10 minutes/location Data Accuracy:** 90%+ Success Rate:** 85%+ Daily Capacity:** 100–500 leads Resource Usage API Calls:** ~10–20 per business Storage:** Minimal (Google Sheets) Execution Time:** 3–8 minutes/10 businesses Network Usage:** ~5–10MB/business 🤝 Support & Community Getting Help n8n Community Forum:** community.n8n.io Docs:** docs.n8n.io BrightData Support:** Via dashboard Contributing Share improvements Report issues and suggestions Create industry-specific variations Document best practices > 🔒 Privacy & Compliance: Ensure GDPR/CCPA compliance. Always respect robots.txt and terms of service of scraped sites. 🎯 Ready to Generate Leads! This workflow provides a complete solution for automated lead generation and outreach. Customize it to fit your needs and start building your pipeline today! For any questions or support, please contact: 📧 info@incrementors.com or fill out this form: Contact Us
by Kumar Shivam
Complete AI Product Description Generator Transforms product images into high-converting copy with GPT-4o Vision + Claude 3.5 The Shopify AI Product Description Factory is a production-grade n8n workflow that converts product images and metadata into refined, SEO-aware descriptions—fully automated and region-agnostic. It blends GPT-4o vision for visible attribute extraction, Claude 3.5 Sonnet for premium copy, Perplexity research for verified brand context, Google Sheets for orchestration and audit trails, plus automated daily sales analytics enrichment. Link-header pagination and structured output enforcement ensure reliable scale. To refine according to your usecase connect via my profile @connect Key Advantages Vision-first copywriting Uses gpt-4o to identify only visible physical attributes (closure, heel, materials, sole) from product images—no guesses. Premium copy generation anthropic/claude-3.5-sonnet crafts concise, benefit-led descriptions with consistent tone, length control, and clean formatting. Research-assisted accuracy perplexityTool verifies vendor/brand context from official sources to avoid speculation or fabricated claims. Pagination you can trust Automates Shopify REST pagination via Link headers and persists page_info for resumable runs. Google Sheets orchestration Centralized staging, status tracking, and QA in Products, with ProcessingState for batch/page markers, and Error_log for diagnostics. Bulletproof error feedback errorTrigger + AI diagnosis logs clear, non-technical and technical explanations to Error_log for fast recovery. Automated sales analytics Daily sales tracking automatically captures and enriches total sales data for comprehensive business intelligence and performance monitoring. How It Works Intake and filtering httpRequest fetches /admin/api/2024-04/products.json?limit=200&{page_info} code filters only items with: Image present Empty body_html The currSeas:SS2025 tag Extracts tag metadata such as x-styleCode, country_of_origin, and gender when available Pagination controller code parses Link headers for rel="next" and extracts page_info googleSheets updates ProcessingState with page_info_next and increments the batch number for resumable polling Generation pipeline googleSheets pulls rows with Status = Ready for AI Description; limit throttles batch size openAi Analyze image (model gpt-4o) returns strictly visible features lmChatOpenRouter (Claude 3.5) composes the SEO description, optionally blending verified vendor context from perplexityTool outputParserStructured guarantees strict JSON: product_id, product_title (normalized), generated_description, status googleSheets writes results back to Products for review/publish Sales analytics enrichment Schedule Trigger** runs daily at 2:01 PM to capture previous day's sales httpRequest fetches paid orders from Shopify REST API with date range filtering splitOut and summarize nodes calculate total daily sales Automatic Google Sheets logging with date stamps and totals Zero-sale days are properly recorded for complete analytics continuity Reliability and insight errorTrigger routes failures to an AI agent that explains the root cause and appends a concise note to Error_log. What's Inside (Node Map) Data + API httpRequest (Shopify REST 2024-04 for products and orders) googleSheets (multiple sheet operations) googleSheetsTool (error logging) AI models openAi (gpt-4o vision analysis) lmChatOpenRouter (anthropic/claude-3.5-sonnet for content generation) AI Agent** (intelligent error diagnosis) Analytics & Processing splitOut (order data processing) summarize (sales totals calculation) set nodes (data field mapping) Tools and guards perplexityTool (brand research) outputParserStructured (JSON validation) memoryBufferWindow (conversation context) Control & Scheduling scheduleTrigger (multiple time-based triggers) cron (periodic execution) limit (batch size control) if (conditional logic) code (custom filtering and pagination logic) Observability errorTrigger + AI diagnosis to Error_log Processing state tracking Sales analytics logging Content & Compliance Rules Locale-agnostic copy**; brand voice is configurable per store Only image-verifiable attributes** (no guesses); clean HTML suitable for Shopify themes Optional normalization rules (e.g., color/branding cleanup, title sanitization) Style code inclusion supported when x-styleCode is present Gender-aware content generation when gender tag is present Strict JSON output** and schema consistency for safe downstream publishing Setup Steps Core integrations Shopify Access Token** — Products read + Orders read (REST 2024-04) OpenAI API** — gpt-4o vision OpenRouter API** — Claude Sonnet (3.5) Perplexity API** — vendor/market verification via perplexityTool Google Sheets OAuth** — Products, ProcessingState, Error_log, Sales analytics Configure sheets ProcessingState** with fields: batch number page_info_next Products** with: Product ID Product Title Product Type Vendor Image url Status country of origin x_style_code gender Generated Description Error_log** with: timestamp Reason of Error Sales Analytics Sheet** with: Date Total Sales Workflow Capabilities Discovery and staging Auto-paginate Shopify; stage eligible products in Sheets with reasons and timestamps. Vision-grounded copywriting Descriptions reflect only visible attributes plus verified brand context; concise, mobile-friendly structure with gender-aware tone. Metadata awareness Auto-injects x-styleCode, country_of_origin, and gender when present; natural SEO for brand and product type. Sales intelligence Automated daily sales tracking with Melbourne timezone support, handles zero-sale days, and maintains complete historical records. Error analytics Layman + technical diagnosis logged to Error_log to shorten MTTR. Safe output Structured JSON via outputParserStructured for predictable row updates. Credentials Required Shopify Access Token** (Products + Orders read permissions) OpenAI API Key** (GPT-4o vision) OpenRouter API Key** (Claude Sonnet) Perplexity API Key** Google Sheets OAuth** Ideal For E-commerce teams** scaling compliant, on-brand product copy with comprehensive sales insights Agencies and SEO specialists** standardizing image-grounded descriptions with performance tracking and analytics Stores** needing resumable pagination, auditable content operations, and automated daily sales reporting in Sheets Advanced Features Dual-workflow architecture**: Content generation + Sales analytics in one system Link-header pagination with page_info persistence in ProcessingState Title/content normalization (e.g., color removal) configurable per brand Gender-aware copywriting** based on product tags Memory windows (memoryBufferWindow) to keep multi-step prompts consistent Melbourne timezone support** for accurate daily sales cutoffs Zero-sales handling** ensures complete analytics continuity Structured Output enforcement for downstream safety AI-powered error diagnosis** with technical and layman explanations Time & Scheduling (Universal) The workflow includes two independent schedules: Content Generation**: Every 5 minutes (configurable) for product processing Sales Analytics**: Daily at 2:01 PM Melbourne time for previous day's sales For globally distributed teams, schedule triggers and timestamps can be standardized on UTC to avoid regional drift. Pro Tip Start with small batches (limit set to 10 or fewer) to validate both copy generation and sales tracking flows. The workflow handles dual operations independently - content generation failures won't affect sales analytics and vice versa. Monitor the Error_log sheet for any issues and use the ProcessingState sheet to track pagination progress.
by Gerald Denor
Overview This comprehensive n8n workflow automatically transforms trending Google search queries into engaging LinkedIn posts using AI. The system runs autonomously, discovering viral topics, researching content, and publishing professionally formatted posts to grow your social media presence. Workflow Description Automate your entire social media content pipeline - from trend discovery to publication. This workflow monitors Google Trends, selects high-potential topics, creates human-like content using advanced AI, and publishes across multiple social platforms with built-in tracking. Key Features Automated Trend Discovery**: Pulls trending topics from Google Trends API with customizable filters Intelligent Topic Selection**: AI chooses the most relevant trending topic for your niche Multi-AI Content Generation**: Combines Perplexity for research and OpenAI for content curation Human-Like Writing**: Advanced prompts eliminate AI detection markers LinkedIn Optimization**: Proper formatting with Unicode characters, emojis, and engagement hooks Multi-Platform Support**: Ready for LinkedIn, Twitter/X, and Facebook posting Automated Scheduling**: Configurable posting times (default: 6 AM & 6 PM daily) Performance Tracking**: Automatic logging to Google Sheets with timestamps and metrics Error Handling**: Built-in delays and retry mechanisms for API stability Technical Implementation Workflow Architecture Schedule Trigger: Automated execution at specified intervals Google Trends API: Fetches trending search queries with geographical filtering Data Processing: JavaScript code node filters high-volume keywords (30+ search volume) Topic Selection: OpenAI GPT-3.5 evaluates and selects optimal trending topic Content Research: Perplexity AI researches selected topic for current information Content Generation: Advanced prompt engineering creates LinkedIn-optimized posts Content Distribution: Multi-platform posting with platform-specific formatting Analytics Tracking: Google Sheets integration for performance monitoring Node Breakdown Schedule Trigger**: Configurable timing for automated execution HTTP Request (Google Trends)**: SerpAPI integration for trend data Set Node**: Structures trending data for processing Code Node**: JavaScript filtering for high-volume keywords OpenAI Node**: Intelligent topic selection based on relevance and trend strength HTTP Request (Perplexity)**: Advanced AI research with anti-detection prompts Wait Node**: Rate limiting and API respect Split Out**: Prepares content for multi-platform distribution LinkedIn Node**: Authenticated posting with community management Google Sheets Node**: Automated tracking and analytics Social Media Nodes**: Twitter/X, LinkedIn and Facebook ready for activation Use Cases Content Creators**: Maintain consistent posting schedules with trending content Marketing Agencies**: Scale content creation across multiple client accounts Business Development**: Build thought leadership with timely industry insights Personal Branding**: Establish authority by commenting on trending topics SEO Professionals**: Create content around high-search-volume keywords Configuration Requirements API Integrations SerpAPI**: Google Trends data access Perplexity AI**: Advanced content research capabilities OpenAI**: Content curation and topic selection LinkedIn Community Management API**: Professional posting access Google Sheets API**: Analytics and tracking Authentication Setup LinkedIn OAuth2 community management credentials Google Sheets OAuth2 integration HTTP header authentication for AI services Customization Options Industry Targeting**: Modify prompts for specific business verticals Posting Schedule**: Adjust timing based on audience activity Content Tone**: Customize voice and style through prompt engineering Platform Selection**: Enable/disable specific social media channels Trend Filtering**: Adjust search volume thresholds and geographic targeting Content Length**: Modify character limits for different platforms Advanced Features Anti-AI Detection**: Sophisticated prompts create human-like content Rate Limit Management**: Built-in delays prevent API throttling Error Recovery**: Robust error handling with retry mechanisms Content Deduplication**: Prevents posting duplicate content Engagement Optimization**: LinkedIn-specific formatting for maximum reach Performance Metrics Time Savings**: Eliminates 10+ hours of weekly content creation Consistency**: Maintains regular posting schedule without manual intervention Relevance**: Content always based on current trending topics Engagement**: Optimized formatting increases social media interaction Scalability**: Single workflow manages multiple platform posting Installation Notes Import JSON workflow file into n8n instance Configure all required API credentials Set up Google Sheets tracking document Test workflow execution with manual trigger Enable schedule trigger for automated operation Best Practices Monitor API usage to stay within rate limits Regularly update prompts based on content performance Review and adjust trending topic filters for your niche Maintain backup of workflow configuration Test content output before enabling automation Support & Updates Comprehensive setup documentation included Configuration troubleshooting guide provided Regular workflow updates for API changes Community support through n8n forums Tags social-media content-automation linkedin ai-generation google-trends perplexity openai marketing trend-analysis content-creation Compatibility n8n Version: 1.0+ Node Requirements: Standard n8n installation External Dependencies: API access to listed services Hosting: Compatible with cloud and self-hosted n8n instances
by Ranjan Dailata
Notice Community nodes can only be installed on self-hosted instances of n8n. Who this is for The Legal Case Research Extractor is a powerful automated workflow designed for legal tech teams, researchers, law firms, and data scientists focused on transforming unstructured legal case data into actionable, structured insights. This workflow is tailored for: Legal Researchers automating case law data mining Litigation Support Teams handling large volumes of case records LawTech Startups building AI-powered legal research assistants Compliance Analysts extracting case-specific insights AI Developers working on legal NLP, summarization, and search engines What problem is this workflow solving? Legal case data is often locked in semi-structured or raw HTML formats, scattered across jurisdiction-specific websites. Manually extracting and processing this data is tedious and inefficient. This workflow automates: Extraction of legal case data via Bright Data's powerful MCP infrastructure Parsing of HTML into clean, readable text using Google Gemini LLM Structuring and delivering the output through webhook and file storage What this workflow does Input Set the Legal Case Research URL node is responsible for setting the legal case URL for the data extraction. Bright Data MCP Data Extractor Bright Data MCP Client For Legal Case Research node is responsible for the legal case extraction via the Bright Data MCP tool - scrape_as_html Case Extractor Google Gemini based Case Extractor is responsible for producing a paginated list of cases Loop through Legal Case URLs Receives a collection of legal case links to process Each URL represents a different case from a target legal website Bright Data MCP Scraping Utilizes Bright Data’s scrape_as_html MCP mode Retrieves raw HTML content of each legal case Google Gemini LLM Extraction Transforms raw HTML into clean, structured text Performs additional information extraction if required (e.g., case summary, court, jurisdiction etc.) Webhook Notification Sends extracted legal case content to a configurable webhook URL Enables downstream processing or storage in legal databases Binary Conversion & File Persistence Converts the structured text to binary format Saves the final response to disk for archival or further processing Pre-conditions Knowledge of Model Context Protocol (MCP) is highly essential. Please read this blog post - model-context-protocol You need to have the Bright Data account and do the necessary setup as mentioned in the Setup section below. You need to have the Google Gemini API Key. Visit Google AI Studio You need to install the Bright Data MCP Server @brightdata/mcp You need to install the n8n-nodes-mcp Setup Please make sure to setup n8n locally with MCP Servers by navigating to n8n-nodes-mcp Please make sure to install the Bright Data MCP Server @brightdata/mcp on your local machine. Sign up at Bright Data. Create a Web Unlocker proxy zone called mcp_unlocker on Bright Data control panel. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Google Gemini(PaLM) Api account with the Google Gemini API key (or access through Vertex AI or proxy). In n8n, configure the credentials to connect with MCP Client (STDIO) account with the Bright Data MCP Server as shown below. Make sure to copy the Bright Data API_TOKEN within the Environments textbox above as API_TOKEN=<your-token> How to customize this workflow to your needs Target New Legal Portals Modify the legal case input URLs to scrape from different state or federal case databases Customize LLM Extraction Modify the prompt to extract specific fields: case number, plaintiff, case summary, outcome, legal precedents etc. Add a summarization step if needed Enhance Loop Handling Integrate with a Google Sheet or API to dynamically fetch case URLs Add error handling logic to skip failed cases and log them Improve Security & Compliance Redact sensitive information before sending via webhook Store processed case data in encrypted cloud storage Output Formats Save as PDF, JSON, or Markdown Enable output to cloud storage (S3, Google Drive) or legal document management systems
by Khairul Muhtadin
Tesseract - Money Mate Workflow Description Disclaimer: This template requires the n8n-nodes-tesseractjs community node, which is only available on self-hosted n8n instances. You’ll need a self-hosted n8n setup to use this workflow. Who is this for? This workflow is designed for individuals, freelancers, or small business owners who want an easy way to track expenses using Telegram. It’s ideal for anyone looking to digitize receipts—whether from photos or text messages—using free tools, without needing advanced technical skills. What problem does this workflow solve? Manually entering receipt details into a spreadsheet or app is time-consuming and prone to mistakes. This workflow automates the process by extracting information from receipt images or text messages sent via Telegram, categorizing expenses, and sending back a clear, formatted summary. It saves time, reduces errors, and makes expense tracking effortless. What this workflow does The workflow listens for messages sent to a Telegram bot, which can be either text descriptions of expenses or photos of receipts. If a photo is sent, Tesseract (an open-source text recognition tool) extracts the text. If text is sent, it’s processed directly. An AI model (LLaMA via OpenRouter) analyzes the input, categorizes it into expense types (e.g., Food & Beverages, Household, Transport), and creates a structured summary including store name, date, items, total, and category. The summary is then sent back to the user’s Telegram chat. Setup Instructions Follow these step-by-step instructions to set up the workflow. No advanced technical knowledge is required, but you’ll need a self-hosted n8n instance. Set Up a Self-Hosted n8n Instance: If you don’t have n8n installed, follow the n8n self-hosting guide to set it up. You can use platforms like Docker or a cloud provider (e.g., DigitalOcean, AWS). Ensure your n8n instance is running and accessible via a web browser. Install the Tesseract Community Node: In your n8n instance, go to Settings > Community Nodes in the sidebar. Click Install a Community Node, then enter n8n-nodes-tesseractjs in the search bar. Click Install and wait for confirmation. This node enables receipt image processing. If you encounter issues, check the n8n community nodes documentation for troubleshooting. Create a Telegram Bot: Open Telegram and search for @BotFather to start a new bot. Send /start to BotFather, then /newbot to create your bot. Follow the prompts to name your bot (e.g., “MoneyMateBot”). BotFather will provide a Bot Token (e.g., 23872837287:ExampleExampleExample). Copy this token. In n8n, go to Credentials > Add Credential, select Telegram API, and paste the token. Name the credential (e.g., “MoneyMateBot”) and save. Set Up OpenRouter for AI Processing: Sign up for a free account at OpenRouter. In your OpenRouter dashboard, generate an API Key under the API section. In n8n, go to Credentials > Add Credential, select OpenRouter API, and paste the API key. Name it (e.g., “OpenRouter Account”) and save. The free tier of OpenRouter’s LLaMA model is sufficient for this workflow. Import and Configure the Workflow: Download the workflow JSON file (provided separately or copy from the source). In n8n, go to Workflows > Import Workflow and upload the JSON file. Open the imported workflow (“Tesseract - Money Mate”). Ensure the Telegram Trigger and Send Expense Summary nodes use the Telegram credential you created. Ensure the AI Analyzer node uses the OpenRouter credential. Save the workflow. Test the Workflow: Activate the workflow by toggling the Active switch in n8n. In Telegram, find your bot (e.g., @MoneyMateBot) and send /start. Test with a sample input (see “Example Inputs” below). Check the n8n workflow execution panel to ensure data flows correctly. If errors occur, double-check credentials and node connections. Activate for Continuous Use: Once tested, keep the workflow active in n8n. Your bot will now process any text or image sent to it via Telegram. Example Inputs/Formats To help the workflow process your data accurately, use clear and structured inputs. Below are examples of valid inputs: Text Input Example: Send a message to your Telegram bot like this: Bought coffee at Starbucks, Jalan Sudirman, yesterday. Total Rp 50,000. 2 lattes, each Rp 25,000. Expected Output: hello [Your Name] Ini Rekap Belanjamu 📋 Store: Starbucks 📍 Location: Jalan Sudirman 📅 Date: 2025-05-26 🛒 Items: Latte: Rp 25,000 Latte: Rp 25,000 💸 Total: Rp 50,000 📌 Category: Food & Beverages Image Input Example: Upload a photo of a receipt to your Telegram bot. The receipt should contain: Store name (e.g., “Alfamart”) Address (e.g., “Jl. Gatot Subroto, Jakarta”) Date and time (e.g., “27/05/2025 14:00”) Items with prices (e.g., “Bread Rp 15,000”, “Milk Rp 20,000”) Total amount (e.g., “Total: Rp 35,000”) Expected Output: hello [Your Name] Ini Rekap Belanjamu 📋 Store: Alfamart 📍 Location: Jl. Gatot Subroto, Jakarta 📅 Date: 2025-05-27 14:00 🛒 Items: Bread: Rp 15,000 Milk: Rp 20,000 💸 Total: Rp 35,000 📌 Category: Household Tips for Images: Ensure the receipt is well-lit and text is readable. Avoid blurry or angled photos for better Tesseract accuracy. How to Customize This Workflow Change Expense Categories: In the **AI Categorizer node, edit the prompt to include custom categories (e.g., add “Entertainment” or “Utilities” to the list: Food & Beverages, Household, Transport). Modify Response Format: In the **Format Summary Message node, adjust the JavaScript code to change how the summary looks (e.g., add emojis, reorder fields). Save to a Database: Add a node (e.g., Google Sheets or PostgreSQL) after the **Format Summary Message node to store summaries. Support Other Languages: In the **AI Categorizer node, update the prompt to handle additional languages (e.g., Spanish, Mandarin) by specifying them in the instructions. Add Error Handling: Enhance the **Check Invalid Input node to catch more edge cases, like invalid dates. All Free, End-to-End This workflow is 100% free! It leverages: Telegram Bot API**: Free via BotFather. Tesseract**: Open-source text recognition. LLaMA via OpenRouter**: Free tier available for AI processing. Enjoy automating your expense tracking without any cost! Made by: khmuhtadin Need a custom? contact me on LinkedIn or Web
by Cheng Siong Chin
Introduction Automate peer review assignment and grading with AI-powered evaluation. Designed for educators managing collaborative assessments efficiently. How It Works Webhook receives assignments, distributes them, AI generates review rubrics, emails reviewers, collects responses, calculates scores, stores results, emails reports, updates dashboards, and posts analytics to Slack. Workflow Template Webhook → Store Assignment → Distribute → Generate Review Rubric → Notify Slack → Email Reviewers → Prepare Response → Calculate Score → Store Results → Check Status → Generate Report → Email Report → Update Dashboard → Analytics → Post to Slack → Respond to Webhook Workflow Steps Receive & Store: Webhook captures assignments, stores data. Distribute & Generate: Assigns peer reviewers, AI creates rubrics. Notify & Email: Alerts via Slack, sends review requests. Collect & Score: Gathers responses, calculates peer scores. Report & Update: Generates reports, emails results, updates dashboard. Analyze & Alert: Posts analytics to Slack, confirms completion. Setup Instructions Webhook & Storage: Configure endpoint, set up database. AI Configuration: Add OpenAI key, customize rubric prompts. Communication: Connect Gmail, Slack credentials. Dashboard: Link analytics platform, configure metrics. Prerequisites OpenAI API key Gmail account Slack workspace Database or storage system Dashboard tool Use Cases University peer review assignments Corporate training evaluations Research paper assessments Customization Multi-round review cycles Custom scoring algorithms Benefits Eliminates manual distribution Ensures consistent evaluation
by moosa
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Overview Automate your email management with this n8n workflow that fetches, summarizes, and shares critical emails from your Gmail inbox. Designed for busy professionals, this workflow runs daily to extract important emails from the past 24 hours, summarizes key details (like credentials, OTPs, deadlines, and action items), converts the summary into a PDF, and sends it to your Discord channel for quick access. Key Features Scheduled Automation**: Triggers daily at 8 PM to process emails from the last 24 hours. Gmail Integration**: Retrieves emails labeled "INBOX" and "IMPORTANT" using secure OAuth2 authentication (no hardcoded API keys). Smart Email Parsing**: Extracts essential fields (subject, sender, and plain text) while cleaning up URLs, extra lines, and formatting for clarity. AI-Powered Summarization**: Uses OpenAI's GPT-4.1-mini to create concise plain text and markdown summaries, highlighting urgent actions with "[Action Required]". PDF Conversion**: Converts the markdown summary into a professional PDF using PDF.co API. Discord Notifications**: Shares the PDF via a Discord webhook for seamless team communication. Why Use This Workflow? Save time by automating email triage and focusing on what matters. Stay organized with clear, actionable summaries delivered to Discord. Securely handle sensitive data with proper credential management. Perfect for teams, freelancers, or anyone managing high email volumes. Setup Instructions Configure Gmail OAuth2 credentials for secure access. Set up PDF.co API and Discord webhook credentials. Customize the schedule or filters as needed. Activate and let the workflow handle your daily email summaries!
by Albert Ho
How it works This workflow will trigger upon new mail in your gmail inbox and will automatically apply existing labels OR identify, create, then apply up to 5 new labels. This is currently set up to run using a OpenAI endpoint which I have set up connected to my local llama.cpp server running. After your email is done labeling, the workflow will send you a nice notification in your discord channel! Set up Instructions This workflow requires a Gmail OAuth account credential with the ability to create labels and label messages [OPTIONAL] You can create a Discord Bot Account which connects with a Bot Token which will summarize the email and show you which labels were created and added in the workflow. You might want to tweak the prompts a bit based on how detailed you want your labels. I've found after running it for a bit that the labels that get created mostly encompass everything I want. Workflow Hints The workflow works in 5 phases Phase 1:** Find existing Gmail labels as options for the LLM to choose from. Phase 2:** LLM chooses labels or defines new ones Phase 3:** Create additional Gmail labels if needed Phase 4:** Label the message with newly created and existing labels Phase 5:** Rewrite and send a message notification via Discord. Screenshot 1: Note: colors are my own Screenshot of notification sent via discord Model Tips I set this up to run completely using offline models. I had good success with Llama.cpp server with gpt-oss-20B and with a bit more tweaking could get it to work with a smaller gemma3-4b model mostly likely as well. The whole workflow runs in about 5-10 seconds on my Blackwell Nvidia GPU. Probably would be a few seconds slower on an older piece of hardware.
by Genzi
Description A comprehensive real estate chatbot automation system that handles customer inquiries, property searches, and appointment scheduling through intelligent conversation flows and email processing. How it works? This template creates an end-to-end real estate automation system that handles customer inquiries from initial contact through appointment booking. 1. Customer Entry Point Webhook receives customer messages from chat interface Link detection checks if customer shared property URLs Smart routing - if property link found, fetch details immediately; otherwise proceed to chat 2. AI Content Processing Content filter (PRIORITY) - blocks non-real estate queries upfront Information extraction - scans messages for personal details and property requirements Human handoff detection - identifies requests for live agent assistance 3. Data Collection Phase Sequential gathering: Personal info (name → phone → email) then property needs Smart validation - phone format, email structure, budget parsing No redundancy - never asks for information already provided PostgreSQL storage - saves customer data and conversation memory 4. Property Search & Matching Database query filters properties by type, location, budget, availability Image enhancement - fetches property photos from media storage Results ranking - returns top 5 matches sorted by price 5. AI Response Generation GPT-4 formatting creates engaging, professional property listings Visual enhancement - includes property images and key details Personalized tone - acknowledges customer preferences 6. Appointment Automation Gmail monitoring - checks for appointment confirmations every hour Calendar integration - creates, updates, deletes appointments automatically Smart scheduling - checks availability, suggests alternatives for conflicts Email responses - sends confirmations and follow-ups Intelligence Features Context Awareness Remembers conversation history across sessions Builds complete customer profile progressively Maintains property preferences throughout interaction Smart Extraction Recognizes property types: HDB, Condo, Apartment Parses locations and MRT preferences automatically Handles various budget formats (SGD 2,500, $2500, etc.) Identifies timeline requirements and citizenship status Professional Handoffs Detects human agent requests with keyword matching Collects complete customer context before transfer Sends structured handoff emails with all requirements Ensures smooth transition to live agents Technical Components AI Models OpenAI GPT-4 - Main conversation handling and response formatting GPT-4 Mini - Appointment processing and email management LangChain Memory - Conversation context retention Database Integration PostgreSQL - Customer data, property listings, conversation history Property search with multi-criteria filtering Media storage integration for property images Communication Channels Webhook API - Primary chat interface Gmail integration - Appointment confirmations and notifications Google Calendar - Automated scheduling and availability checking Setup Requirements Configure database - PostgreSQL with property and customer tables Set up integrations - Gmail, Google Calendar, OpenAI API Customize prompts - Adjust AI responses for your brand Test workflow - Verify end-to-end functionality Monitor performance - Track conversation success rates The system is designed to handle the complete customer journey from initial inquiry to scheduled property viewing, with intelligent automation reducing manual work while maintaining high service quality.
by Rahul Joshi
Description Automate your GoHighLevel (GHL) client onboarding process from the moment a deal is marked as “Won.” This workflow seamlessly generates client folders in Google Drive, duplicates contract and kickoff templates, schedules kickoff calls, sends branded welcome emails, creates onboarding tasks in GHL, and notifies your team in Slack. 🚀📂📧📅💬 What This Template Does Triggers automatically when an opportunity is marked as Won in GHL 🏆 Validates and formats client data to ensure clean records 🔍 Creates structured client folders in Google Drive 📂 Copies contract & kickoff deck templates with client-specific naming 📝 Sends personalized welcome email via Gmail ✉️ Schedules kickoff call in Google Calendar 📅 Creates onboarding tasks in GHL for account managers ✅ Sends Slack notifications to keep your team informed instantly 💬 Catches errors and sends alerts to a Slack error channel 🚨 Key Benefits Saves 30–45 minutes per onboarding ⏱️ Eliminates manual data entry and human errors 🧹 Guarantees consistent client experience across all deals 🤝 Automates document creation & sharing 📄 Ensures team visibility and faster response times 📲 Built-in validation and error handling for reliability 🔐 Features Webhook-based trigger from GoHighLevel ⚡ Automatic client data formatting and validation 🛠️ Google Drive folder & document automation with templates 📂 Personalized Gmail welcome email with branding ✉️ Automated kickoff call scheduling in Google Calendar 📅 Task creation in GHL for seamless follow-up 📌 Slack notifications for both success and error handling 💬 Error channel with detailed failure reports 🚨 Requirements n8n instance (cloud or self-hosted) GoHighLevel account with API access 🔑 Google Workspace (Drive, Gmail, Calendar) 📊 Slack workspace with Bot Token & channel access 💬 Pre-created contract and kickoff deck templates in Google Drive 📄 Target Audience Agencies & consultants using GoHighLevel for client management 🏢 Sales teams wanting instant onboarding after a deal closes 💰 Operations teams seeking consistent and repeatable onboarding flows ⚙️ Account managers who need structured onboarding tasks ✅ Businesses scaling client onboarding and reducing manual workload 🚀 Step-by-Step Setup Instructions Configure GHL webhook → Trigger on “Opportunity Status Changed = Won.” Connect your GHL API credentials (OAuth2 or API key). Add Google Drive OAuth2 credentials → Set parent folder ID & template IDs. Configure Gmail OAuth2 → Replace hardcoded email with client email variable. Connect Google Calendar → Select the calendar for kickoff calls. Connect Slack API → Choose channels for onboarding updates and errors. Update template IDs for contract and kickoff deck in the workflow. Import workflow into n8n, map credentials, and test once. Enable workflow → onboarding is now fully automated. ✅
by Rahul Joshi
Description This workflow is designed to evaluate newly added CVs for Diversity, Equity, and Inclusion (DEI) eligibility. It automatically ingests CVs from Google Drive, extracts key fields, analyzes them with Azure OpenAI, logs structured DEI outcomes in Google Sheets, and sends a concise DEI-focused summary email to the hiring manager. The entire flow prioritizes consistent, auditable DEI checks and controlled logic paths. What This Template Does Watches Google Drive for new CV files to trigger DEI evaluation. Downloads and extracts text/structured fields from PDF CVs. Assesses DEI eligibility using Azure OpenAI, following defined criteria and prompts. Appends DEI results (eligible/not eligible, rationale, confidence) to Google Sheets for tracking. Generates and sends a DEI-focused summary email to the hiring manager for review. Key Benefits Standardized DEI screening to support equitable hiring decisions. Centralized, structured logging in Sheets for transparency and audits. Automated DEI summaries for faster, consistent manager review. Reliable routing with true/false logic to enforce DEI evaluation steps. Features Google Drive trigger (fileCreated) for CV intake tied to DEI checks. PDF extraction mapped to fields relevant for DEI evaluation. Azure OpenAI Chat Model prompts tuned for DEI criteria and rationale. Google Sheets append with eligibility status, notes, and timestamps. Email node that delivers DEI summaries and next-step guidance. Logic branching (true/false) to control DEI evaluation and notifications. Requirements n8n instance (cloud or self-hosted). Google Drive access to the CV intake folder. Google Sheets access for DEI results logging. Azure OpenAI access and configured prompts reflecting DEI criteria. Email node credentials to send DEI summaries to managers. Step-by-Step Setup Instructions Connect Google Drive and select the CV folder for the fileCreated trigger. Configure the Download CV and Extract From PDF nodes to capture fields needed for DEI checks. Add Azure OpenAI credentials and set DEI-specific prompts (criteria, rationale, confidence). Connect Google Sheets and select the target sheet; map columns for status, rationale, and timestamps. Configure the Email to Manager node with a DEI-focused subject and template. Test with sample CVs, verify sheet entries and email content, then enable the workflow. DEI-Focused Best Practices Clarify DEI criteria and document them in your prompt and sheet schema. Avoid including sensitive PII in emails; store only necessary fields for DEI decisions. Use n8n Credentials; never hardcode API keys or private data. Maintain an audit trail (timestamps, model version, prompt version, decision rationale). Periodically review prompts and sheet schema to align with policy updates.
by Yanagi Chinatsu
Who's it for? This template is perfect for teams, communities, or anyone who wants a fun, automated way to remember and celebrate birthdays. If you love space and want to make someone's special day a little more cosmic, this is for you. It's a great way to boost morale and ensure no one's birthday is forgotten. How it works / What it does The workflow runs automatically every morning. It reads a list of birthdays from your Google Sheet and checks if any match the current date. If a match is found, it fetches the latest Earth image from NASA's EPIC API and uses an AI model to generate a unique, space-themed birthday greeting. Finally, it sends a personalized HTML email to the birthday person and posts a celebratory message in your designated Slack channel. Requirements A Google Sheet with columns for Name, Email, and Birthday (in a format like YYYY-MM-DD). Credentials for Google Sheets, Gmail, OpenAI, and Slack configured in your n8n instance. How to set up Configure Credentials: Add your API credentials for Google (for both Sheets and Gmail), OpenAI, and Slack in the respective nodes. Set Up Google Sheet: In the "Get Birthday Roster" node, enter your Google Sheet ID and the name of the sheet where the birthday list is stored. Configure Slack: In the "Post Slack Notification" node, select the channel where you want birthday announcements to be posted. Activate Workflow: Save and activate the workflow. It will now run automatically every day! How to customize the workflow Change the schedule: Modify the CRON expression in the "Every Morning at 7:00" node to change when the workflow runs. Adjust the AI prompt: Edit the system message in the "Generate Space Birthday Message" node to alter the tone, length, or style of the AI-generated greetings. Customize the email design: Modify the HTML in the "Send Birthday Email" node to change the look and feel of the birthday message.