by DataMinex
Transform property searches into personalized experiences! This powerful automation delivers dream home matches straight to clients' inboxes with professional CSV reports - all from a simple web form. ๐ What this workflow does Create a complete real estate search experience that works 24/7: โจ Smart Web Form - Beautiful property search form captures client preferences ๐ง Dynamic SQL Builder - Intelligently creates optimized queries from user input โก Lightning Database Search - Scans 1000+ properties in milliseconds ๐ Professional CSV Export - Excel-ready reports with complete property details ๐ง Automated Email Delivery - Personalized emails with property previews and attachments ๐ฏ Perfect for: Real Estate Agents** - Generate leads and impress clients with instant service Property Managers** - Automate tenant matching and recommendations Brokerages** - Provide 24/7 self-service property discovery Developers** - Showcase available properties with professional automation ๐ก Why this workflow is a game-changer > "From property search to professional report delivery in under 30 seconds!" โก Instant Results: Zero wait time for property matches ๐จ Professional Output: Beautiful emails that showcase your expertise ๐ฑ Mobile Optimized: Works flawlessly on all devices ๐ง Smart Filtering: Only searches criteria clients actually specify ๐ Infinitely Scalable: Handles unlimited searches simultaneously ๐ Real Estate Data Source Built on authentic US market data from the Github: ๐๏ธ 1000+ Real Properties across all US states ๐ฐ Actual Market Prices from legitimate listings ๐ Complete Property Details (bedrooms, bathrooms, square footage, lot size) ๐ Verified Locations with accurate cities, states, and ZIP codes ๐ข Broker Information for authentic real estate context ๐ ๏ธ Quick Setup Guide Prerequisites Checklist โ [ ] SQL Server database (MySQL/PostgreSQL also supported) [ ] Gmail account for automated emails [ ] n8n instance (cloud or self-hosted) [ ] 20 minutes setup time Step 1: Import Real Estate Data ๐ฅ ๐ Download the data ๐พ Download CSV file (1000+ properties included) ๐๏ธ Create SQL Server table with this exact schema: CREATE TABLE [REALTOR].[dbo].[realtor_usa_price] ( brokered_by BIGINT, status NVARCHAR(50), price DECIMAL(12,2), bed INT, bath DECIMAL(3,1), acre_lot DECIMAL(10,8), street BIGINT, city NVARCHAR(100), state NVARCHAR(50), zip_code INT, house_size INT, prev_sold_date NVARCHAR(50) ); ๐ Import your CSV data into this table Step 2: Configure Database Connection ๐ ๐ Set up Microsoft SQL Server credentials in n8n โ Test connection to ensure everything works ๐ฏ Workflow is pre-configured for the table structure above Step 3: Gmail Setup (The Magic Touch) ๐ง ๐ Visit Google Cloud Console ๐ Create new project (or use existing) ๐ Enable Gmail API in API Library ๐ Create OAuth2 credentials (Web Application) โ๏ธ Add your n8n callback URL to authorized redirects ๐ Configure Gmail OAuth2 credentials in n8n โจ Authorize your Google account Step 4: Launch Your Property Search Portal ๐ ๐ Import this workflow template (form is pre-configured) ๐ Copy your webhook URL from the Property Search Form node ๐ Test with a sample property search ๐จ Check email delivery with CSV attachment ๐ Go live and start impressing clients! ๐จ Customization Playground ๐ท๏ธ Personalize Your Brand // Customize email subjects in the Gmail node "๐ Exclusive Properties Curated Just for You - ${results.length} Perfect Matches!" "โจ Your Dream Home Portfolio - Handpicked by Our Experts" "๐ฏ Hot Market Alert - ${results.length} Premium Properties Inside!" ๐ง Advanced Enhancements ๐จ HTML Email Templates**: Create stunning visual emails with property images ๐ Analytics Dashboard**: Track popular searches and user engagement ๐ Smart Alerts**: Set up automated price drop notifications ๐ฑ Mobile Integration**: Connect to React Native or Flutter apps ๐ค AI Descriptions**: Add ChatGPT for compelling property descriptions ๐ Multi-Database Flexibility // Easy database switching // MySQL: Replace Microsoft SQL node โ MySQL node // PostgreSQL: Swap for PostgreSQL node // MongoDB: Use MongoDB node with JSON queries // Even CSV files: Use CSV reading nodes for smaller datasets ๐ Advanced Features & Extensions ๐ฅ Pro Tips for Power Users ๐ Bulk Processing**: Handle multiple searches simultaneously ๐พ Smart Caching**: Store popular searches for lightning-fast results ๐ Lead Scoring**: Track which properties generate most interest ๐ Follow-up Automation**: Schedule nurturing email sequences ๐ฏ Integration Possibilities ๐ข CRM Connection**: Auto-add qualified leads to your CRM ๐ Calendar Integration**: Add property viewing scheduling ๐ Price Monitoring**: Track market trends and price changes ๐ฑ Social Media**: Auto-share featured properties to social platforms ๐ฌ Chat Integration**: Connect to WhatsApp or SMS for instant alerts ๐ Expand Your Real Estate Automation ๐ Related Workflow Ideas ๐ค AI Property Valuation - Add machine learning for price predictions ๐ Market Analysis Reports - Generate comprehensive market insights ๐ฑ SMS Property Alerts - Instant text notifications for hot properties ๐ข Commercial Property Search - Adapt for office and retail spaces ๐น Investment ROI Calculator - Add financial analysis for investors ๐๏ธ Neighborhood Analytics - Include school ratings and demographics ๐ ๏ธ Technical Extensions ๐ท Image Processing: Auto-resize and optimize property photos ๐บ๏ธ Map Integration: Add interactive property location maps ๐ฑ Progressive Web App: Create mobile app experience ๐ Push Notifications: Real-time alerts for saved searches ๐ Get Started Now Import this workflow template Configure your database and Gmail Customize branding and messaging Launch your professional property search portal Watch client satisfaction soar!
by vinci-king-01
Competitor Price Monitoring Dashboard with AI and Real-time Alerts ๐ฏ Target Audience E-commerce managers and pricing analysts Retail business owners monitoring competitor pricing Marketing teams tracking market positioning Product managers analyzing competitive landscape Data analysts conducting pricing intelligence Business strategists making pricing decisions ๐ Problem Statement Manual competitor price monitoring is inefficient and often leads to missed opportunities or delayed responses to market changes. This template solves the challenge of automatically tracking competitor prices, detecting significant changes, and providing actionable insights for strategic pricing decisions. ๐ง How it Works This workflow automatically monitors competitor product prices using AI-powered web scraping, analyzes price trends, and sends real-time alerts when significant changes are detected. Key Components Scheduled Trigger - Runs the workflow at specified intervals to maintain up-to-date price data AI-Powered Scraping - Uses ScrapeGraphAI to intelligently extract pricing information from competitor websites Price Analysis Engine - Processes historical data to detect trends and anomalies Alert System - Sends notifications via Slack and email when price changes exceed thresholds Dashboard Integration - Stores all data in Google Sheets for comprehensive analysis and reporting ๐ Google Sheets Column Specifications The template creates the following columns in your Google Sheets: | Column | Data Type | Description | Example | |--------|-----------|-------------|---------| | timestamp | DateTime | When the price was recorded | "2024-01-15T10:30:00Z" | | competitor_name | String | Name of the competitor | "Amazon" | | product_name | String | Product name and model | "iPhone 15 Pro 128GB" | | current_price | Number | Current price in USD | 999.00 | | previous_price | Number | Previous recorded price | 1099.00 | | price_change | Number | Absolute price difference | -100.00 | | price_change_percent | Number | Percentage change | -9.09 | | product_url | URL | Direct link to product page | "https://amazon.com/iphone15" | | alert_triggered | Boolean | Whether alert was sent | true | | trend_direction | String | Price trend analysis | "Decreasing" | ๐ ๏ธ Setup Instructions Estimated setup time: 15-20 minutes Prerequisites n8n instance with community nodes enabled ScrapeGraphAI API account and credentials Google Sheets account with API access Slack workspace for notifications (optional) Email service for alerts (optional) Step-by-Step Configuration 1. Install Community Nodes Install required community nodes npm install n8n-nodes-scrapegraphai npm install n8n-nodes-slack 2. Configure ScrapeGraphAI Credentials Navigate to Credentials in your n8n instance Add new ScrapeGraphAI API credentials Enter your API key from ScrapeGraphAI dashboard Test the connection to ensure it's working 3. Set up Google Sheets Connection Add Google Sheets OAuth2 credentials Grant necessary permissions for spreadsheet access Create a new spreadsheet for price monitoring data Configure the sheet name (default: "Price Monitoring") 4. Configure Competitor URLs Update the websiteUrl parameters in ScrapeGraphAI nodes Add URLs for each competitor you want to monitor Customize the user prompt to extract specific pricing data Set appropriate price thresholds for alerts 5. Set up Notification Channels Configure Slack webhook or API credentials Set up email service credentials (SendGrid, SMTP, etc.) Define alert thresholds and notification preferences Test notification delivery 6. Configure Schedule Trigger Set monitoring frequency (hourly, daily, etc.) Choose appropriate time zones for your business hours Consider competitor website rate limits 7. Test and Validate Run the workflow manually to verify all connections Check Google Sheets for proper data formatting Test alert notifications with sample data ๐ Workflow Customization Options Modify Monitoring Targets Add or remove competitor websites Change product categories or specific products Adjust monitoring frequency based on market volatility Extend Price Analysis Add more sophisticated trend analysis algorithms Implement price prediction models Include competitor inventory and availability tracking Customize Alert System Set different thresholds for different product categories Create tiered alert systems (info, warning, critical) Add SMS notifications for urgent price changes Output Customization Add data visualization and reporting features Implement price history charts and graphs Create executive dashboards with key metrics ๐ Use Cases Dynamic Pricing**: Adjust your prices based on competitor movements Market Intelligence**: Understand competitor pricing strategies Promotion Planning**: Time your promotions based on competitor actions Inventory Management**: Optimize stock levels based on market conditions Customer Communication**: Proactively inform customers about price changes ๐จ Important Notes Respect competitor websites' terms of service and robots.txt Implement appropriate delays between requests to avoid rate limiting Regularly review and update your monitoring parameters Monitor API usage to manage costs effectively Keep your credentials secure and rotate them regularly Consider legal implications of automated price monitoring ๐ง Troubleshooting Common Issues: ScrapeGraphAI connection errors: Verify API key and account status Google Sheets permission errors: Check OAuth2 scope and permissions Price parsing errors: Review the Code node's JavaScript logic Rate limiting: Adjust monitoring frequency and implement delays Alert delivery failures: Check notification service credentials Support Resources: ScrapeGraphAI documentation and API reference n8n community forums for workflow assistance Google Sheets API documentation for advanced configurations Slack API documentation for notification setup
by Shayan Ali Bakhsh
About this Workflow This workflow helps you repurpose your YouTube videos across multiple social media platforms with zero manual effort. Itโs designed for creators, businesses, and marketers who want to maximize reach without spending hours re-uploading content everywhere. How It Works Trigger from YouTube The workflow checks your YouTube channel every 10 minutes via RSS feed. It compares the latest video ID with the last saved one to detect if a new video was uploaded. Tutorial: How to get YouTube Channel RSS Feed Generate Descriptions with AI Uses Gemini 2.5 Flash to automatically generate fresh, engaging descriptions for your posts. Create Images with ContentDrips ContentDrips offers multiple templates (carousel, single image, branding templates, etc.). The workflow generates a custom promotional image using your video description and thumbnail. Install node: npm install n8n-nodes-contentdrips Docs: ContentDrips Blog Tutorial Publish Across Social Platforms with SocialBu Instead of manually connecting each social media API, this workflow uses SocialBu. From a single connection, you can post to: Facebook, Instagram, TikTok, YouTube, Twitter (X), LinkedIn, Threads, Pinterest, and more. Website: SocialBu Get Real-Time Notifications via Discord After each run, the workflow sends updates to your Discord channel. Youโll know if the upload was successful, or if an error occurred (e.g., API limits). Setup guide: Discord OAuth Credentials Why Use This Workflow? Saves time by automating the entire repurposing process. Ensures consistent branding and visuals across platforms. Works around platform restrictions by leveraging SocialBuโs integrations. Keeps you updated with Discord notificationsโno guessing if something failed. Requirements YouTube channel RSS feed link ContentDrips API key, template ID, and branding setup SocialBu account with connected social media platforms Discord credentials (for webhook updates) Need Help? Message me on LinkedIn: Shayan Ali Bakhsh Happy Automation ๐
by inderjeet Bhambra
Who's it for Content creators, trainers, and educators who need to convert lengthy documents into digestible micro-learning experiences. How it works This workflow takes your source content (PDFs, articles, handbooks) and uses GPT-4 to intelligently break it into 2-3 minute learning modules. Each module includes a key concept, explanation, practical example, and knowledge check question. How to set up Configure OpenAI credentials with GPT-4 access Connect Slack workspace (optional) Set up Google Docs integration Optionally, Send content via webhook or paste directly Requirements OpenAI API key with GPT-4 access Google Docs account (for document creation) Slack workspace (optional, for notifications) How to customize the workflow Adjust module count and length in AI prompts Modify output formats (email, mobile, Slack) Change document structure and styling Add custom delivery channels Perfect for converting employee handbooks, training materials, and documentation into engaging micro-learning courses that people actually complete.
by Ibrahim Emre POLAT
How it works Automatically generates professional PDF invoices from webhook data and delivers them via email while storing backups in Google Drive. Perfect for freelancers, small businesses, and service providers who need automated billing workflows. Set up steps Configure environment variables for company information (name, address, email, phone). Set up your PDF generation API service account (PDFShift recommended). Configure SMTP email credentials for invoice delivery. Set up Google Drive OAuth2 for cloud storage. Deploy the workflow and test with sample invoice data. Key features Smart invoice number generation if not provided Automatic tax calculations with configurable rates Professional HTML templates with company branding Parallel processing for email and storage Comprehensive error handling and validation Detailed success confirmation responses Required environment variables COMPANY_NAME - Your business name COMPANY_ADDRESS - Business mailing address COMPANY_EMAIL - Billing contact email COMPANY_PHONE - Business phone number PDF_API_URL - PDF generation service endpoint PDF_API_KEY - API authentication key GDRIVE_INVOICE_FOLDER_ID - Google Drive folder ID API requirements PDF generation service (PDFShift, HTML/CSS to PDF API, or similar), SMTP email service for delivery, Google Drive API access for storage. Input format { "customerName": "John Smith", "customerEmail": "john@example.com", "items": [ {"description": "Web Design", "quantity": 1, "price": 500} ], "dueDate": "2025-02-15" }
by Yusuke
๐ง Overview Generate empathetic, professional reply drafts for customer or user messages. The workflow detects sentiment, tone, and risk level, drafts a concise response, sanitizes PII/links/emojis, and auto-escalates risky or low-confidence cases to human review. โ๏ธ How It Works Input โ Manual Test or Webhook Trigger AI Agent (Empathy) โ returns { sentiment, tone, reply, confidence, needs_handover } Post-Process & Sanitize โ removes URLs/hashtags, masks PII, caps length Risk & Handover Rules โ checks confidence threshold, risk words, and negativity Routing โ auto-send safe replies or flag to Needs Review ๐งฉ Setup Instructions (3โ5 min) Open Set Config1 and adjust: MAX_LEN (default 600) ADD_FOLLOWUP_QUESTION (true/false) FORMALITY (auto | casual | polite) EMOJI_ALLOWED (true/false), BLOCK_LINKS (true/false) RISK_WORDS (e.g., refund, lawsuit, self-harm) Connect Anthropic credential to Anthropic Chat Model (Optional) Replace Manual Trigger with Webhook Trigger for real-time use > Tip: If you need to show literal angle brackets in messages, use backticks like `<example>` (no HTML entities needed). ๐ Use Cases 1) SaaS Billing Complaints Input:** โI was billed after canceling. This is unacceptable.โ Output:** Calm, apologetic reply with refund steps; escalates if refund is in RISK_WORDS or confidence < 0.45. 2) Product Bug Reports Input:** โUpload fails on large files since yesterday.โ Output:** Acknowledges impact, requests logs, offers workaround; routes to auto-send if low risk and high confidence. 3) Delivery/Logistics Delays Input:** โMy order is late again. Should I file a complaint?โ Output:** Empathetic apology, ETA guidance, partial credit policy note; escalates if language indicates legal action. 4) Community Moderation / Abuse Input:** โSupport is uselessโyouโre all scammers.โ Output:** De-escalating, policy-aligned response; auto-flags due to negative sentiment + risk keyword match. 5) Safety / Self-harm Mentions Input:** โI feel like hurting myself if this isnโt fixed.โ Output:* *Immediate escalation**, inserts approved resources; never auto-sends. ๐จ Auto-Escalation Rules (defaults) Negative** sentiment Message matches any RISK_WORDS confidence < 0.45 Mentions of legal, harassment, or self-harm context ๐งช Notes & Best Practices ๐ No hardcoded API keys โ use n8n Credentials ๐งญ Tune thresholds and RISK_WORDS to your org policy ๐งฉ Works on self-hosted or cloud n8n โ Treat outputs as drafts; ship after human/policy review ๐ Resources GitHub (template JSON):** https://github.com/yskmtb0714/n8n-workflows/blob/main/empathy-reply-assistant.json
by Growth AI
Who's it for Marketing teams, business intelligence professionals, competitive analysts, and executives who need consistent industry monitoring with AI-powered analysis and automated team distribution via Discord. What it does This intelligent workflow automatically monitors multiple industry topics, scrapes and analyzes relevant news articles using Claude AI, and delivers professionally formatted intelligence reports to your Discord channel. The system provides weekly automated monitoring cycles with personalized bot communication and comprehensive content analysis. How it works The workflow follows a sophisticated 7-phase automation process: Scheduled Activation: Triggers weekly monitoring cycles (default: Mondays at 9 AM) Query Management: Retrieves monitoring topics from centralized Google Sheets configuration News Discovery: Executes comprehensive Google News searches using SerpAPI for each configured topic Content Extraction: Scrapes full article content from top 3 sources per topic using Firecrawl AI Analysis: Processes scraped content using Claude 4 Sonnet for intelligent synthesis and formatting Discord Optimization: Automatically segments content to comply with Discord's 2000-character message limits Automated Delivery: Posts formatted intelligence reports to Discord channel with branded "Claptrap" bot personality Requirements Google Sheets account for query management SerpAPI account for Google News access Firecrawl account for article content extraction Anthropic API access for Claude 4 Sonnet Discord bot with proper channel permissions Scheduled execution capability (cron-based trigger) How to set up Step 1: Configure Google Sheets query management Create monitoring sheet: Set up Google Sheets document with "Query" sheet Add search topics: Include industry keywords, competitor names, and relevant search terms Sheet structure: Simple column format with "Query" header containing search terms Access permissions: Ensure n8n has read access to the Google Sheets document Step 2: Configure API credentials Set up the following credentials in n8n: Google Sheets OAuth2: For accessing query configuration sheet SerpAPI: For Google News search functionality with proper rate limits Firecrawl API: For reliable article content extraction across various websites Anthropic API: For Claude 4 Sonnet access with sufficient token limits Discord Bot API: With message posting permissions in target channel Step 3: Customize scheduling settings Cron expression: Default set to "0 9 * * 1" (Mondays at 9 AM) Frequency options: Adjust for daily, weekly, or custom monitoring cycles Timezone considerations: Configure according to team's working hours Execution timing: Ensure adequate processing time for multiple topics Step 4: Configure Discord integration Set up Discord delivery settings: Guild ID: Target Discord server (currently: 919951151888236595) Channel ID: Specific monitoring channel (currently: 1334455789284364309) Bot permissions: Message posting, embed suppression capabilities Brand personality: Customize "Claptrap" bot messaging style and tone Step 5: Customize content analysis Configure AI analysis parameters: Analysis depth: Currently processes top 3 articles per topic Content format: Structured markdown format with consistent styling Language settings: Currently configured for French output (easily customizable) Quality controls: Error handling for inaccessible articles and content How to customize the workflow Query management expansion Topic categories: Organize queries by industry, competitor, or strategic focus areas Keyword optimization: Refine search terms based on result quality and relevance Dynamic queries: Implement time-based or event-triggered query modifications Multi-language support: Add international keyword variations for global monitoring Advanced content processing Article quantity: Modify from 3 to more articles per topic based on analysis needs Content filtering: Add quality scoring and relevance filtering for article selection Source preferences: Implement preferred publisher lists or source quality weighting Content enrichment: Add sentiment analysis, trend identification, or competitive positioning Discord delivery enhancements Rich formatting: Implement Discord embeds, reactions, or interactive elements Multi-channel distribution: Route different topics to specialized Discord channels Alert levels: Add priority-based messaging for urgent industry developments Archive functionality: Create searchable message threads or database storage Integration expansions Slack compatibility: Replace or supplement Discord with Slack notifications Email reports: Add formatted email distribution for executive summaries Database storage: Implement persistent storage for historical analysis and trending API endpoints: Create webhook endpoints for third-party system integration AI analysis customization Analysis templates: Create topic-specific analysis frameworks and formatting Competitive focus: Enhance competitor mention detection and analysis depth Trend identification: Implement cross-topic trend analysis and strategic insights Summary levels: Create executive summaries alongside detailed technical analysis Advanced monitoring features Intelligent content curation The system provides sophisticated content management: Relevance scoring: Automatic ranking of articles by topic relevance and publication authority Duplicate detection: Prevents redundant coverage of the same story across different sources Content quality assessment: Filters low-quality or promotional content automatically Source diversity: Ensures coverage from multiple perspectives and publication types Error handling and reliability Graceful degradation: Continues processing even if individual articles fail to scrape Retry mechanisms: Automatic retry logic for temporary API failures or network issues Content fallbacks: Uses article snippets when full content extraction fails Notification continuity: Ensures Discord delivery even with partial content processing Results interpretation Intelligence report structure Each monitoring cycle delivers: Topic-specific summaries: Individual analysis for each configured search query Source attribution: Complete citation with publication date, source, and URL Structured formatting: Consistent presentation optimized for quick scanning Professional analysis: AI-generated insights maintaining factual accuracy and business context Performance analytics Monitor system effectiveness through: Processing metrics: Track successful article extraction and analysis rates Content quality: Assess relevance and usefulness of delivered intelligence Team engagement: Monitor Discord channel activity and report utilization System reliability: Track execution success rates and error patterns Use cases Competitive intelligence Market monitoring: Track competitor announcements, product launches, and strategic moves Industry trends: Identify emerging technologies, regulatory changes, and market shifts Partnership tracking: Monitor alliance formations, acquisitions, and strategic partnerships Leadership changes: Track executive movements and organizational restructuring Strategic planning support Market research: Continuous intelligence gathering for strategic decision-making Risk assessment: Early warning system for industry disruptions and regulatory changes Opportunity identification: Spot emerging markets, technologies, and business opportunities Brand monitoring: Track industry perception and competitive positioning Team collaboration enhancement Knowledge sharing: Centralized distribution of relevant industry intelligence Discussion facilitation: Provide common information baseline for strategic discussions Decision support: Deliver timely intelligence for business planning and strategy sessions Competitive awareness: Keep teams informed about competitive landscape changes Workflow limitations Language dependency: Currently optimized for French analysis output (easily customizable) Processing capacity: Limited to 3 articles per query (configurable based on API limits) Platform specificity: Configured for Discord delivery (adaptable to other platforms) Scheduling constraints: Fixed weekly schedule (customizable via cron expressions) Content access: Dependent on article accessibility and website compatibility with Firecrawl API dependencies: Requires active subscriptions and proper rate limit management for all integrated services
by Cheng Siong Chin
Introduction Automates travel planning by aggregating flights, hotels, activities, and weather via APIs, then uses AI to generate professional itineraries delivered through Gmail and Slack. How It Works Webhook receives requests, searches APIs (Skyscanner, Booking.com, Kiwi, Viator, weather), merges data, AI builds itineraries, scores options, generates HTML emails, delivers via Gmail/Slack. Workflow Template Webhook โ Extract โ Parallel Searches (Flights/Hotels/Activities/Weather) โ Merge โ Build Itinerary โ AI Processing โ Score โ Generate HTML โ Gmail โ Slack โ Response Workflow Steps Trigger & Extract: Receives destination, dates, preferences, extracts parameters. Data Gathering: Parallel APIs fetch flights, hotels, activities, weather, merges responses. AI Processing: Analyzes data, creates itinerary, ranks recommendations. Delivery: Generates HTML email, sends via Gmail/Slack, confirms completion. Setup Instructions API Configuration: Add keys for Skyscanner, Booking.com, Kiwi, Viator, OpenWeatherMap, OpenRouter. Communication: Connect Gmail OAuth2, Slack webhook. Customization: Adjust endpoints, AI prompts, HTML template, scoring criteria. Prerequisites API keys: Skyscanner, Booking.com, Kiwi, Viator, OpenWeatherMap, OpenRouter Gmail account Slack workspace n8n instance Use Cases Corporate travel planning Vacation itinerary generation Group trip coordination Customization Add sources (Airbnb, TripAdvisor) Filter by budget preferences Add PDF generation Customize Slack format Benefits Saves 3-5 hours per trip Real-time pricing aggregation AI-powered personalization Automated multi-channel delivery
by Cristina
Automated Weekly Newsletter with AI Research, Editorial Drafting, and Approval Flow This n8n template demonstrates how to automate the full production cycle of a professional weekly newsletter. It combines AI-powered market research, editorial drafting, compliance validation, and an approval loop โ all before creating a final Gmail draft ready for distribution. Use cases are many: Wealth managers sending weekly market updates to clients Startups producing recurring research digests Teams creating automated newsletters for marketing or content distribution Good to know At time of writing, each AI call (research, editorial, QC) consumes API tokens from Perplexity and OpenAI. See provider pricing for updated info. Gmail integration requires OAuth setup with your account. You can adapt the prompts to any domain โ from finance to tech to education. How it works Schedule Trigger runs every week and sets the date window. Research LLM fetches structured JSON market data using Perplexity. Editorial LLM transforms research into a polished ~2,000 word client-ready newsletter. QC LLM validates factual accuracy and compliance risks before approval. Preview Email is sent with Approve/Revise buttons. Clicking opens a secure n8n-hosted approval form. Final Draft Creation: Once approved, the workflow generates a clean Gmail draft, ready to send. How to use Replace the demo schedule trigger with your own (weekly, daily, or event-based). Set up Gmail OAuth credentials to enable email previews and final drafts. Update branding (logo, disclaimer, signature) in the HTML builder node. Adjust prompts to your audience โ e.g., simplify tone for marketing, or keep institutional tone for financial clients. Requirements Perplexity account for research API (or your LLM of choice) OpenAI for editorial and QC steps (or your LLM of choice) Gmail account with OAuth credentials Optional: your own domain to host the approval webhook Customising this workflow This workflow can be extended beyond financial newsletters. Try: Content marketing: Automate weekly digests or trend reports Education: Generate curriculum summaries with approval loops for teachers Internal comms: Automate compliance-checked company updates
by PDF Vector
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Intelligent Document Monitoring and Alert System This workflow creates an automated monitoring system that watches for new PDF reports across multiple sources, extracts key insights using AI, and sends formatted alerts to your team via Slack or email. By combining PDF Vector's parsing capabilities with GPT-powered analysis, teams can stay informed about critical documents without manual review, ensuring important information never gets missed. Target Audience & Problem Solved This template is designed for: Finance teams** monitoring quarterly reports and regulatory filings Compliance officers** tracking policy updates and audit reports Research departments** alerting on new publications and preprints Operations teams** monitoring supplier reports and KPI documents Executive assistants** summarizing board materials and briefings It solves the problem of information overload by automatically processing incoming documents, extracting only the most relevant insights, and delivering them in digestible formats to the right people at the right time. Prerequisites n8n instance with PDF Vector node installed PDF Vector API credentials with parsing capabilities OpenAI API key for insight extraction Slack workspace admin access (for Slack alerts) SMTP credentials (for email alerts) FTP/Cloud storage access for document sources Minimum 50 API credits for continuous monitoring Step-by-Step Setup Instructions Configure Document Sources Set up FTP credentials in n8n for folder monitoring Or configure Google Drive/Dropbox integration Define the folder paths to monitor Set file naming patterns to watch (e.g., "report.pdf") Set Up API Integrations Add PDF Vector credentials in n8n Configure OpenAI credentials with appropriate model access Set up Slack app and add webhook URL Configure SMTP settings for email alerts Configure Monitoring Schedule Open the "Check Every 15 Minutes" node Adjust frequency based on your needs: // For hourly checks: "interval": 60 // For real-time monitoring (every 5 min): "interval": 5 Customize Alert Channels Slack Setup: Create dedicated channels (#reports, #alerts) Configure webhook for each channel Set up user mentions for urgent alerts Email Setup: Define recipient lists by document type Configure email templates Set up priority levels for subject lines Define Alert Rules Modify the "Extract Key Insights" prompt for your domain Set conditions for high-priority alerts Configure which metrics trigger notifications Define sentiment thresholds Implementation Details The workflow implements a comprehensive monitoring pipeline: Source Monitoring: Polls multiple sources for new PDFs Intelligent Parsing: Uses LLM-enhanced parsing for complex documents Insight Extraction: AI analyzes content for key information Priority Classification: Determines alert urgency based on content Multi-Channel Delivery: Sends formatted alerts via configured channels Audit Trail: Logs all processed documents for compliance Customization Guide Adding Custom Document Types: Extend the routing logic for specific document types: // In "Route by Document Type" node: const documentTypes = { 'invoice': /invoice|bill|payment/i, 'contract': /contract|agreement|terms/i, 'report': /report|analysis|summary/i, 'compliance': /audit|compliance|regulatory/i }; Customizing Insight Extraction: Modify the AI prompt for domain-specific analysis: // Financial documents: "Extract: 1) Revenue figures 2) YoY growth 3) Risk factors 4) Guidance changes" // Compliance documents: "Extract: 1) Policy changes 2) Deadlines 3) Required actions 4) Penalties" // Research papers: "Extract: 1) Key findings 2) Methodology 3) Implications 4) Future work" Advanced Alert Formatting: Create rich Slack messages with interactive elements: // Add buttons for quick actions: { "type": "actions", "elements": [ { "type": "button", "text": { "type": "plain_text", "text": "View Full Report" }, "url": documentUrl }, { "type": "button", "text": { "type": "plain_text", "text": "Mark as Read" }, "action_id": "mark_read" } ] } Implementing Alert Conditions: Add sophisticated filtering based on content: // Alert only if certain conditions are met: if (insights.metrics.revenue_change < -10) { priority = 'urgent'; alertChannel = '#executive-alerts'; } if (insights.findings.includes('compliance violation')) { additionalRecipients.push('legal@company.com'); } Adding Document Comparison: Track changes between document versions: // Compare with previous version: const previousDoc = await getLastVersion(documentType); const changes = compareDocuments(previousDoc, currentDoc); if (changes.significant) { alertMessage += \nโ ๏ธ Significant changes detected: ${changes.summary}; } Alert Features: Monitor multiple document sources (FTP, cloud storage, email) Extract key metrics and findings with AI Send rich, formatted notifications Track document processing history Conditional alerts based on content analysis Multi-channel alert routing Use Cases: Financial report monitoring Compliance document tracking Research publication alerts Customer report distribution Board material summarization Regulatory filing notifications Advanced Configuration Performance Optimization: Implement caching to avoid reprocessing Use batch processing for multiple documents Set up parallel processing for different sources Security Considerations: Encrypt sensitive document storage Implement access controls for different alert channels Audit log all document access
by Greypillar
How it works โข Twilio webhook detects missed/unanswered calls automatically โข Analyzes call context (time of day, business hours, weekend/weekday) โข Checks CRM for existing customer data and purchase history (optional) โข AI Chain (GPT-4o) generates personalized recovery messages for SMS and Email โข Sends instant apology with booking link via SMS and/or Email โข Logs all interactions to Airtable for tracking โข Slack alerts sales team with customer priority and context โข Waits 24 hours and checks if customer booked โข Sends automatic follow-up SMS if no booking detected โข Structured output parser ensures reliable JSON formatting Set up steps โข Time to set up: 15-20 minutes โข Add Twilio credential and configure phone number in both SMS nodes โข Get Slack channel ID for sales team alerts โข Create Airtable base with 10 columns (Caller_Number, Customer_Name, Call_Time, Priority, SMS_Sent, Email_Sent, Booking_Link, Status, Urgency_Score, Follow_Up_Date) โข Replace booking URL with your Cal.com or Calendly link โข Configure business email for SMTP sending โข Update business phone and name in follow-up SMS โข Add credentials: Twilio API, OpenAI (GPT-4o), Slack OAuth2, Airtable Token, SMTP Email โข Replace placeholder IDs in Slack and Airtable nodes โข Optional: Configure CRM API for existing customer lookup (or disable those nodes) What you'll need โข Twilio - Phone number with SMS capability โข OpenAI API - GPT-4o access for AI message generation โข Slack - Standard workspace for sales alerts โข Airtable - Free plan works for call tracking โข Email SMTP - Gmail, SendGrid, or any SMTP provider โข Cal.com or Calendly - Booking link for instant scheduling โข CRM API (optional) - HubSpot, Salesforce, or custom CRM Who this is for Service businesses, agencies, and sales teams that lose revenue from missed calls. Perfect for businesses with high call volume that need instant personalized follow-up to recover lost opportunities.
by Oneclick AI Squad
This automated n8n workflow transforms uploaded radiology images into professional, patient-friendly PDF reports. It uses AI-powered image analysis to interpret medical scans, simplify technical terms, and produce clear explanations. The reports are formatted, converted to PDF, stored in a database, and sent directly to patients via email, ensuring both accuracy and accessibility. ๐ฅ Workflow Overview: Simple Process Flow: Upload Image โ 2. AI Analysis โ 3. Generate Report โ 4. Send to Patient ๐ง How It Works: Webhook Trigger - Receives image uploads via POST request Extract Image Data - Processes patient info and image data AI Image Analysis - Uses GPT-4 Vision to analyze the radiology image Process Analysis - Structures the AI response into readable sections Generate PDF Report - Creates a beautiful HTML report Convert to PDF - Converts HTML to downloadable PDF Save to Database - Logs all reports in Google Sheets Email Patient - Sends the report via email Return Response - Confirms successful processing ๐ Key Features: AI-Powered Analysis** using GPT-4 Vision Patient-Friendly Language** (no medical jargon) Professional PDF Reports** with clear sections Email Delivery** with report attachment Database Logging** for record keeping Simple Webhook Interface** for easy integration ๐ Usage Example: Send POST request to webhook with: { "patient_name": "John Smith", "patient_id": "P12345", "scan_type": "X-Ray", "body_part": "Chest", "image_url": "https://example.com/xray.jpg", "doctor_name": "Dr. Johnson", "patient_email": "john@email.com" } โ๏ธ Required Setup: OpenAI API - For GPT-4 Vision image analysis PDF Conversion Service - HTML to PDF converter Gmail Account - For sending reports Google Sheets - For logging reports Replace YOUR_REPORTS_SHEET_ID with your actual sheet ID Want a tailored workflow for your business? Our experts can craft it quickly Contact our team