by Anwar Bouilouta
Most service businesses and freelancers track their quotes in Airtable (or something similar), but when it comes to actually sending the quote, they're still manually copying data into a document, exporting a PDF, attaching it to an email, and then going back to update the record. This workflow cuts all of that out. Who is this for Freelancers, agencies, and small service businesses who already use Airtable to track client work and want to stop manually creating quote documents. If you've ever caught yourself copying the same line items into a Google Doc for the third time in a week, this is for you. How it works The workflow starts with a webhook. You can call it from an Airtable automation (when a record hits "Ready to Send", for example) or trigger it manually from Postman or any HTTP client. Once triggered, it pulls the quote record from Airtable, grabs the client details and line items, and builds a branded HTML document with proper formatting, subtotals, tax calculation, and a grand total. That HTML then gets converted to a PDF through pdf.co's free API. From there, the PDF gets uploaded to a Google Drive folder and emailed to the client as an attachment via Gmail. The Airtable record is updated with the Drive link and marked as "Sent" so you have a clear audit trail. How to set it up Create an Airtable base with a "Quotes" table. You'll need these columns: Client Name, Client Email, Line Items (long text, one per line in the format Description | Qty | Price), Tax Rate, Notes, and Status Open the "Configure Settings" node in the workflow and fill in your business name, email address, physical address, and Google Drive folder ID Connect your Airtable, Gmail, and Google Drive accounts in n8n Sign up for a free pdf.co account and add the API key as an HTTP Header Auth credential Activate the workflow and test it by sending a POST request with {"recordId": "rec..."} to the webhook URL Requirements Airtable account (free tier works) Gmail account with OAuth connected in n8n Google Drive for PDF storage pdf.co API key (free tier gives you 100 credits/month, each conversion costs 1 credit) Customizing the workflow The HTML template lives inside the "Build HTML Quote" code node. You can change the layout, fonts, colors, and add your logo by editing the HTML string directly. The brand color is pulled from the Configure Settings node, so changing it there updates the whole document. You can also adjust the quote validity period (default is 30 days) or change the reference number format in the same code node. If you want to swap pdf.co for another HTML-to-PDF service, just update the HTTP Request node URL and payload format.
by Aslamul Fikri Alfirdausi
This n8n template demonstrates how to build O'Carla, an advanced all-in-one Discord AI assistant. It intelligently handles natural conversations, professional image generation, and visual file analysis within a single server integration. Use cases are many: Deploy a smart community manager that remembers past interactions, an on-demand artistic tool for your members, or an AI that can "read" and explain uploaded documents and images! Good to know API Costs:** Each interaction costs vary depending on the model used (Gemini vs. OpenRouter). Check your provider's dashboard for updated pricing. Infrastructure:* This workflow requires a separate Discord bot script (e.g., Node.js) to forward events to the n8n Webhook. It is recommended to host the bot using *PM2** for 24/7 uptime. How it works Webhook Trigger: Receives incoming data (text and attachments) from your Discord bot. Intent Routing: The workflow uses conditional logic to detect if the user wants an image (via keyword gambar:), a vision analysis (via attachments), or a standard chat. Multi-Model Intelligence: Gemini 2.5: Powers rapid and high-quality general chat reasoning. Llama 3.2 Vision (via OpenRouter): Specifically used to describe and analyze images or text-based files. Flux (via Pollinations): Uses a specialized AI Agent to refine prompts and generate professional-grade images. Contextual Memory: A 50-message buffer window ensures O'Carla maintains the context of your conversation based on your Discord User ID. Clean UI Output: Generated image links are automatically shortened via TinyURL to keep the Discord chat interface tidy. How to use Connect your Google Gemini and OpenRouter API keys in the respective nodes. Replace the Webhook URL in your bot script with this workflow's Production Webhook URL. Type gambar: [your prompt] in Discord to generate images. Upload an image or file to Discord to trigger the AI Vision analysis. Requirements n8n instance (Self-hosted or Cloud). Google Gemini API Key. OpenRouter API Key. Discord Bot Token and hosting environment. Customising this workflow O'Carla is highly flexible. You can change her personality by modifying the System Message in the Agent nodes, adjust the memory window length, or swap the LLM models to specialized ones like Claude 3.5 or GPT-4o.
by PDF Vector
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Intelligent Document Monitoring and Alert System This workflow creates an automated monitoring system that watches for new PDF reports across multiple sources, extracts key insights using AI, and sends formatted alerts to your team via Slack or email. By combining PDF Vector's parsing capabilities with GPT-powered analysis, teams can stay informed about critical documents without manual review, ensuring important information never gets missed. Target Audience & Problem Solved This template is designed for: Finance teams** monitoring quarterly reports and regulatory filings Compliance officers** tracking policy updates and audit reports Research departments** alerting on new publications and preprints Operations teams** monitoring supplier reports and KPI documents Executive assistants** summarizing board materials and briefings It solves the problem of information overload by automatically processing incoming documents, extracting only the most relevant insights, and delivering them in digestible formats to the right people at the right time. Prerequisites n8n instance with PDF Vector node installed PDF Vector API credentials with parsing capabilities OpenAI API key for insight extraction Slack workspace admin access (for Slack alerts) SMTP credentials (for email alerts) FTP/Cloud storage access for document sources Minimum 50 API credits for continuous monitoring Step-by-Step Setup Instructions Configure Document Sources Set up FTP credentials in n8n for folder monitoring Or configure Google Drive/Dropbox integration Define the folder paths to monitor Set file naming patterns to watch (e.g., "report.pdf") Set Up API Integrations Add PDF Vector credentials in n8n Configure OpenAI credentials with appropriate model access Set up Slack app and add webhook URL Configure SMTP settings for email alerts Configure Monitoring Schedule Open the "Check Every 15 Minutes" node Adjust frequency based on your needs: // For hourly checks: "interval": 60 // For real-time monitoring (every 5 min): "interval": 5 Customize Alert Channels Slack Setup: Create dedicated channels (#reports, #alerts) Configure webhook for each channel Set up user mentions for urgent alerts Email Setup: Define recipient lists by document type Configure email templates Set up priority levels for subject lines Define Alert Rules Modify the "Extract Key Insights" prompt for your domain Set conditions for high-priority alerts Configure which metrics trigger notifications Define sentiment thresholds Implementation Details The workflow implements a comprehensive monitoring pipeline: Source Monitoring: Polls multiple sources for new PDFs Intelligent Parsing: Uses LLM-enhanced parsing for complex documents Insight Extraction: AI analyzes content for key information Priority Classification: Determines alert urgency based on content Multi-Channel Delivery: Sends formatted alerts via configured channels Audit Trail: Logs all processed documents for compliance Customization Guide Adding Custom Document Types: Extend the routing logic for specific document types: // In "Route by Document Type" node: const documentTypes = { 'invoice': /invoice|bill|payment/i, 'contract': /contract|agreement|terms/i, 'report': /report|analysis|summary/i, 'compliance': /audit|compliance|regulatory/i }; Customizing Insight Extraction: Modify the AI prompt for domain-specific analysis: // Financial documents: "Extract: 1) Revenue figures 2) YoY growth 3) Risk factors 4) Guidance changes" // Compliance documents: "Extract: 1) Policy changes 2) Deadlines 3) Required actions 4) Penalties" // Research papers: "Extract: 1) Key findings 2) Methodology 3) Implications 4) Future work" Advanced Alert Formatting: Create rich Slack messages with interactive elements: // Add buttons for quick actions: { "type": "actions", "elements": [ { "type": "button", "text": { "type": "plain_text", "text": "View Full Report" }, "url": documentUrl }, { "type": "button", "text": { "type": "plain_text", "text": "Mark as Read" }, "action_id": "mark_read" } ] } Implementing Alert Conditions: Add sophisticated filtering based on content: // Alert only if certain conditions are met: if (insights.metrics.revenue_change < -10) { priority = 'urgent'; alertChannel = '#executive-alerts'; } if (insights.findings.includes('compliance violation')) { additionalRecipients.push('legal@company.com'); } Adding Document Comparison: Track changes between document versions: // Compare with previous version: const previousDoc = await getLastVersion(documentType); const changes = compareDocuments(previousDoc, currentDoc); if (changes.significant) { alertMessage += \n⚠️ Significant changes detected: ${changes.summary}; } Alert Features: Monitor multiple document sources (FTP, cloud storage, email) Extract key metrics and findings with AI Send rich, formatted notifications Track document processing history Conditional alerts based on content analysis Multi-channel alert routing Use Cases: Financial report monitoring Compliance document tracking Research publication alerts Customer report distribution Board material summarization Regulatory filing notifications Advanced Configuration Performance Optimization: Implement caching to avoid reprocessing Use batch processing for multiple documents Set up parallel processing for different sources Security Considerations: Encrypt sensitive document storage Implement access controls for different alert channels Audit log all document access
by Growth AI
Advanced Form Submission to CRM Automation with International Phone Support Who's it for Sales teams, marketing professionals, and business owners who need sophisticated lead management with international phone number support, automated CRM record creation, intelligent duplicate detection, and multi-channel team notifications. What it does This advanced workflow automatically processes form submissions from your website and creates a complete, intelligent CRM structure in Pipedrive. It transforms raw form data into organized sales records including companies, contacts, deals, and relevant notes while handling international phone number formatting and providing real-time team notifications via Discord and WhatsApp messaging. How it works The workflow follows an intelligent automation process with four distinct scenarios: Form Trigger: Captures form submissions from your website (Webflow in this example) Advanced Phone Processing: Automatically detects and formats international phone numbers with proper country codes for 20+ countries including France, Belgium, Switzerland, Germany, Spain, Italy, Morocco, Algeria, Tunisia, and more Intelligent CRM Logic: Uses a sophisticated 4-scenario approach: Scenario A: Existing Organization + Existing Person - Links records and creates new deal Scenario B: Existing Organization + New Person - Creates person, links to organization, creates deal Scenario C: New Organization + Existing Person - Creates organization, links person, creates deal Scenario D: New Organization + New Person - Creates complete new structure from scratch Enhanced Data Management: Adds lead source tracking, custom properties, and conditional data enhancement Multi-Channel Communication: Sends formatted alerts to Discord and personalized WhatsApp messages to leads Requirements Webflow account (or any platform that supports webhook triggers) Pipedrive CRM account with proper API credentials Team notification service: Discord, Slack, Microsoft Teams, email service, or any webhook-compatible notification tool WhatsApp Business API access for lead messaging International phone number handling capability How to set up Step 1: Configure your form trigger Default setup: The template uses Webflow Form Trigger with site ID configuration Alternative platforms: Replace with webhook trigger for other platforms (WordPress, custom websites, etc.) Webhook configuration: Set up your website's form to send data to the n8n webhook URL Form fields: Ensure your form captures the necessary fields: Prénom (First Name) Nom (Last Name) Entreprise (Company) Mail professionnel (Professional Email) Téléphone pro (Professional Phone) URL du site internet (Website URL) Message Step 2: Configure API credentials Set up the following credentials in n8n: Webflow OAuth2: For form trigger authentication (or webhook authentication for other platforms) Pipedrive API: For CRM record creation and management - ensure proper permissions for organizations, persons, deals, and notes Discord Bot API: For team notifications with guild and channel access WhatsApp Business API: For automated lead messaging with phone number ID configuration Step 3: Customize international phone formatting The "international dialing code" node automatically handles: European countries: France (+33), Belgium (+32), Switzerland (+41), Germany (+49), Spain (+34), Italy (+39), Portugal (+351) North African countries: Morocco (+212), Algeria (+213), Tunisia (+216) Global coverage: US/Canada (+1), UK (+44), and many Asian countries Fallback handling: Defaults to French formatting for unrecognized patterns Error management: Uses +330000000000 as fallback for invalid numbers Step 4: Configure Pipedrive settings Adjust Pipedrive-specific settings in deal creation nodes: Deal pipeline stage: Currently set to default stage (customize for your pipeline) Deal ownership: Configure owner_id for appropriate team member assignment Currency settings: Adjust currency code for your business region Custom properties: Lead source automatically set to "Growth AI" (customize as needed) Step 5: Set up team notifications Configure your preferred notification system: Discord (default): Set guild ID: 1377297267014504520, channel ID: 1380469490139009106 Alternative platforms: Replace Discord node with Slack, Teams, email, or custom webhook Message formatting: Customize notification content and structure Multi-channel setup: Add multiple notification nodes for different channels Step 6: Configure WhatsApp messaging Set up automated lead engagement: Phone number ID: Configure WhatsApp Business API phone number (currently: 752773604591912) Message personalization: Uses prospect's first name and customizable content International compatibility: Works with formatted international phone numbers Message templates: Customize welcome messages and follow-up content How to customize the workflow Form platform integration Webflow: Use the existing Webflow trigger with site ID configuration WordPress: Replace with webhook trigger and configure Contact Form 7, Gravity Forms, or WPForms Custom websites: Set up webhook trigger with your form's POST endpoint Landing page builders: Configure webhook integration (Unbounce, Leadpages, Instapage, etc.) Form field mapping: Adjust the "Data refinement" node for your specific form structure Advanced CRM customization Pipeline management: Configure different stage IDs for various lead sources Lead scoring: Add conditional logic for deal values based on form responses Custom fields: Map additional form fields to Pipedrive custom properties Multiple pipelines: Route different form types to different sales pipelines Ownership rules: Implement round-robin or territory-based assignment logic International phone number expansion The phone formatting system supports extensive customization: Additional countries: Add new country patterns to the JavaScript code Regional preferences: Modify default formatting rules for specific regions Validation rules: Implement stricter phone number validation Carrier detection: Add mobile vs. landline detection logic Notification enhancements Multi-platform notifications: Send to Discord, Slack, Teams, and email simultaneously Conditional notifications: Route different lead types to different channels Rich formatting: Add embeds, attachments, or rich text formatting Escalation rules: Implement priority-based notification routing Integration expansion: Connect to internal tools or third-party notification services Data validation and enrichment Email validation: Add email verification steps before CRM creation Company enrichment: Integrate with data enrichment services (Clearbit, ZoomInfo, Apollo) Duplicate detection: Enhanced logic to check for existing contacts across multiple fields Lead qualification: Implement sophisticated scoring based on form responses and external data Data cleaning: Add standardization for company names, job titles, and other fields Advanced conditional logic features Intelligent scenario routing The workflow uses sophisticated logic to determine the correct processing path: Organization detection: Exact matching search for existing companies Person identification: Full name matching within relevant organization contexts Relationship preservation: Maintains proper links between organizations, persons, and deals Data consistency: Ensures no duplicate records while preserving historical relationships Smart data handling Enhanced conditional processing includes: Phone number intelligence: Automatic international formatting with country detection Message processing: Creates deal notes only when message field contains meaningful content URL handling: Adds website URLs as separate notes when provided Empty field management: Gracefully handles incomplete form submissions Custom property management: Adds lead source tracking and other metadata Error handling and resilience Graceful failures: Workflow continues even if individual steps fail Data validation: Comprehensive checks for required fields before processing Notification reliability: Ensures team is notified even if some CRM operations fail Logging capabilities: Detailed error tracking for troubleshooting Rollback mechanisms: Ability to handle partial failures without data corruption Results interpretation CRM structure created For each form submission, the workflow creates: Organization record: Complete company information with proper formatting Person record: Contact information linked to correct organization with phone formatting Deal record: Sales opportunity with appropriate stage, owner, and metadata Enhanced notes: Separate notes for messages and website URLs when provided Proper relationships: Full linking between organization, person, and deal records Custom tracking: Lead source attribution and other custom properties Team notifications and engagement Comprehensive communication includes: Discord notifications: Formatted team alerts with complete prospect information WhatsApp engagement: Personalized messages to leads with international number support Immediate alerts: Real-time notifications for instant follow-up capability Formatted display: Clean, organized presentation of all prospect data Multi-channel flexibility: Easy adaptation to any notification platform Advanced use cases International lead generation Global forms: Handle submissions from multiple countries with proper phone formatting Multi-language support: Process forms in different languages with consistent data structure Regional routing: Route leads to appropriate regional sales teams based on phone country codes Currency handling: Automatic currency assignment based on detected country Sophisticated lead management Lead scoring: Advanced qualification based on company size, industry, and message content Progressive profiling: Build complete prospect profiles over multiple interactions Engagement tracking: Monitor response rates and optimize messaging Attribution analysis: Track lead sources and optimize marketing spend Enterprise integration Custom CRM fields: Map to complex Pipedrive custom field structures Multiple pipelines: Route leads to different sales processes based on criteria Team assignment: Intelligent routing based on territory, expertise, or workload Compliance handling: Ensure data processing meets regional privacy requirements Workflow architecture details Processing phases Form capture and data extraction: Webflow trigger processes submitted data International phone formatting: Advanced JavaScript processing for global numbers Organization discovery: Intelligent search and creation logic Person management: Sophisticated duplicate detection and relationship management Deal creation: Context-aware opportunity generation with proper associations Enhanced communication: Multi-channel notifications and lead engagement Performance characteristics Processing time: Typically completes within 10-15 seconds for complex scenarios Reliability: Built-in error handling ensures high success rates Scalability: Handles high-volume form submissions without performance degradation Flexibility: Easy customization for different business requirements and CRM configurations Limitations and considerations Platform dependencies: Currently optimized for Webflow and Pipedrive but adaptable Phone number coverage: Supports 20+ countries but may need expansion for specific regions CRM limitations: Requires proper Pipedrive API permissions and rate limit considerations Form structure: Field mapping requires customization for different form designs Language considerations: Currently configured for French field names but easily adaptable Notification dependencies: Requires proper configuration of Discord and WhatsApp APIs for full functionality
by Kirill Khatkevich
This workflow transforms your Meta Ads creatives into a rich dataset of actionable insights. It's designed for data-driven marketers, performance agencies, and analysts who want to move beyond basic metrics and understand the specific visual and textual elements that drive ad performance. By automatically analyzing every video and image with Google's powerful AI (Video Intelligence and Vision APIs), it systematically deconstructs your creatives into labeled data, ready for correlation with campaign results. Use Case You know some ads perform better than others, but do you know why? Is it the presence of a person, a specific object, the on-screen text, or the spoken words in a video? Answering these questions manually is nearly impossible at scale. This workflow automates the deep analysis process, allowing you to: Automate Creative Analysis:** Stop guessing and start making data-backed decisions about your creative strategy. Uncover Hidden Performance Drivers:** Identify which objects, themes, text, or spoken phrases correlate with higher engagement and conversions. Build a Structured Creative Database:** Create a detailed, searchable log of every element within your ads for long-term analysis and trend-spotting. Save Countless Hours:** Eliminate the tedious manual process of watching, tagging, and logging creative assets. How it Works The workflow is triggered on a schedule and follows a clear, structured path: 1. Configuration & Ad Ingestion: The workflow begins on a schedule (e.g., weekly on Monday at 10 AM). It starts by fetching all active ads from a specific Meta Ads Campaign, which you define in the Set Campaign ID node. 2. Intelligent Branching (Video vs. Image): An IF node inspects each creative to determine its type. Video creatives** are routed to the Google Video Intelligence API pipeline. Image creatives** are routed to the Google Vision API pipeline. 3. The Video Analysis Pipeline: For each video, the workflow gets a direct source URL, downloads the file, and converts it to a Base64 string. It then initiates an asynchronous analysis job in the Google Video Intelligence API, requesting LABEL_DETECTION, SPEECH_TRANSCRIPTION, and TEXT_DETECTION. A loop with a wait timer periodically checks the job status until the analysis is complete. Finally, a Code node parses the complex JSON response, structuring the annotations (like detected objects with timestamps or full speech transcripts) into clean rows. 4. The Image Analysis Pipeline: For each image, the file is downloaded, converted to Base64, and sent to the Google Vision API. It requests a wide range of features, including label, text, logo, and object detection. A Code node parses the response and formats the annotations into a standardized structure. 5. Data Logging & Robust Error Handling: All successfully analyzed data from both pipelines is appended to a primary Google Sheet. The workflow is built to be resilient. If an error occurs (e.g., a video fails to be processed by the API, or an image URL is missing), a detailed error report is logged to a separate errors sheet in your Google Sheet, ensuring no data is lost and problems are easy to track. Setup Instructions To use this template, you need to configure a few key nodes. 1. Credentials: Connect your Meta Ads account. Connect your Google account. This account needs access to Google Sheets and must have the Google Cloud Vision API and Google Cloud Video Intelligence API enabled in your GCP project. 2. The Set Campaign ID Node: This is the primary configuration step. Open this Set node and replace the placeholder value with the ID of the Meta Ads campaign you want to analyze. 3. Google Sheets Nodes: You need to configure two Google Sheets nodes: Add Segments data:** Select your spreadsheet and the specific sheet where you want to save the successful analysis results. Ensure your sheet has the following headers: campaign_id, ad_id, creative_id, video_id, file_name, image_url, source, annotation_type, label_or_text, category, full_transcript, confidence, start_time_s, end_time_s, language_code, processed_at_utc. Add errors:** Select your spreadsheet and the sheet you want to use for logging errors (e.g., a sheet named "errors"). Ensure this sheet has headers like: error_type, error_message, campaign_id, ad_id, creative_id, file_name, processed_at_utc. 4. Activate the Workflow: Set your desired frequency in the Run Weekly on Monday at 10 AM (Schedule Trigger) node. Save and activate the workflow. Further Ideas & Customization This workflow provides the "what" inside your creatives. The next step is to connect it to performance. Build a Performance Analysis Workflow:** Create a second workflow that reads this Google Sheet, fetches performance data (spend, clicks, conversions) for each ad_id from the Meta Ads API, and merges the two datasets. This will allow you to see which labels correlate with the best performance. Create Dashboards:** Use the structured data in your Google Sheet as a source for a Looker Studio or Tableau dashboard to visualize creative trends. Incorporate Generative AI:** Add a final step that sends the combined performance and annotation data to an LLM (like in the example you provided) to automatically generate qualitative summaries and recommendations for each creative. Add Notifications:** Use the Slack or Email nodes to send a summary after each run, reporting how many creatives were analyzed and if any errors occurred.
by Jimmy Gay
🔧 AI-Powered Auto-Maintenance System for n8n Transform your n8n instance management with this advanced automation system featuring artificial intelligence-driven workflow selection. This template provides comprehensive maintenance operations with smart filtering capabilities. ✨ Key Features 🤖 Artificial Intelligence Engine Multi-criteria scoring system for intelligent workflow selection Semantic analysis for business-critical pattern recognition Automated decision-making with configurable thresholds 🎯 Core Maintenance Operations Security Audits**: Automated vulnerability scanning with Google Sheets reporting Smart Pause/Resume**: Intelligent workflow suspension during maintenance windows AI Backup Creation**: Selective duplication of high-value workflows Intelligent Export**: Comprehensive system backups with metadata 🔐 Enterprise Security Token-based authentication with request validation Protected workflow safeguards (never modifies critical systems) Comprehensive error handling and logging ⚡ Automation & Scheduling Configurable maintenance schedules (daily, weekly, monthly) Webhook-driven operations for external integration Real-time monitoring and statistics 🎯 Perfect For DevOps Teams**: Streamline n8n maintenance operations Enterprise Users**: Manage large-scale workflow environments System Administrators**: Automated security and backup management Advanced Users**: Leverage AI for intelligent workflow management 🚀 Quick Setup Import the template Configure 4 credentials (n8n API, Google Sheets, Google Drive, Webhook Auth) Set your security token and Google Sheet ID Activate and enjoy automated maintenance! 🧠 AI Intelligence Highlights The system evaluates workflows using 6+ criteria including activity status, complexity, priority tags, business criticality, and recent updates. Workflows are automatically scored and selected based on intelligent thresholds. Selection Logic: Duplicate threshold: ≥3 points (smart backup selection) Export threshold: ≥5 points (comprehensive backup) System workflows always protected 📊 Includes 25+ configured nodes with emoji naming 4 detailed markdown documentation cards Pre-configured schedules and examples Comprehensive error handling Statistical reporting and monitoring Perfect for organizations looking to implement intelligent, automated n8n maintenance with minimal manual intervention.
by Cheng Siong Chin
Introduction Automate price monitoring for e-commerce competitors—ideal for retailers, analysts, and pricing teams. Scrapes competitor sites, extracts pricing/stock data via AI, detects changes, and sends instant alerts for dynamic pricing strategies. How It Works Scrapes competitor URLs via Firecrawl and Apify, extracts data with AI, detects price/stock changes, logs to Google Sheets, and sends Telegram alerts. Workflow Template Trigger → Scrape URL → AI Extract → Parse → Merge Historical → Detect Changes → Update Sheets + Send Telegram Alert Workflow Steps Trigger & Scrape → Manual/scheduled trigger → Firecrawl + Apify fetch competitor data AI Processing → Claude extracts product details → Parses and structures data Change Detection → Reads historical prices → Merges with current data → Identifies updates Output → Logs alerts to Sheets → Updates historical data → Sends Telegram notification Setup Instructions 1. Firecrawl API Get key from dashboard → Add to n8n 2. Apify API Get key from console → Add to n8n → Configure actors 3. AI Model (Claude/OpenAI) Get API key → Add to n8n 4. Google Sheets OAuth2 Create OAuth2 in Google Cloud Console → Authorize in n8n → Enable API 5. Telegram Bot Create via BotFather → Get token & chat ID → Add to n8n 6. Spreadsheet Setup Create Sheet with required columns → Copy ID → Paste in workflow Prerequisites Self-hosted n8n, Firecrawl account, Apify account, Claude/OpenAI API key, Google account (Sheets OAuth2),Telegram bot Customization Add more URLs, adjust scraping intervals, change detection thresholds, switch to Slack/email alerts, integrate databases Benefits Saves 2+ hours daily, real-time tracking, automated alerts, historical analysis, multi-source scraping
by Growth AI
Intelligent chatbot with custom knowledge base Who's it for Businesses, developers, and organizations who need a customizable AI chatbot for internal documentation access, customer support, e-commerce assistance, or any use case requiring intelligent conversation with access to specific knowledge bases. What it does This workflow creates a fully customizable AI chatbot that can be deployed on any platform supporting webhook triggers (websites, Slack, Teams, etc.). The chatbot accesses a personalized knowledge base stored in Supabase and can perform advanced actions like sending emails, scheduling appointments, or updating databases beyond simple conversation. How it works The workflow combines several powerful components: Webhook Trigger: Accepts messages from any platform that supports webhooks AI Agent: Processes user queries with customizable personality and instructions Vector Database: Searches relevant information from your Supabase knowledge base Memory System: Maintains conversation history for context and traceability Action Tools: Performs additional tasks like email sending or calendar booking Technical architecture Chat trigger connects directly to AI Agent Language model, memory, and vector store all connect as tools/components to the AI Agent Embeddings connect specifically to the Supabase Vector Store for similarity search Requirements Supabase account and project AI model API key (any LLM provider of your choice) OpenAI API key (for embeddings - this is covered in Cole Medin's tutorial) n8n built-in PostgreSQL access (for conversation memory) Platform-specific webhook configuration (optional) How to set up Step 1: Configure your trigger The template uses n8n's default chat trigger For external platforms: Replace with webhook trigger and configure your platform's webhook URL Supported platforms: Any service with webhook capabilities (websites, Slack, Teams, Discord, etc.) Step 2: Set up your knowledge base For creating and managing your vector database, follow this comprehensive guide: Watch Cole Medin's tutorial on document vectorization This video shows how to build a complete knowledge base on Supabase The tutorial covers document processing, embedding creation, and database optimization Important: The video explains the OpenAI embeddings configuration required for vector search Step 3: Configure the AI agent Define your prompt: Customize the agent's personality and role Example: "You are the virtual assistant for example.com. Help users by answering their questions about our products and services." Select your language model: Choose any AI provider you prefer (OpenAI, Anthropic, Google, etc.) Set behavior parameters: Define response style, tone, and limitations Step 4: Connect Supabase Vector Store Add the "Supabase Vector Store" tool to your agent Configure your Supabase project credentials Mode: Set to "retrieve-as-tool" for automatic agent integration Tool Description: Customize description (default: "Database") to describe your knowledge base Table configuration: Specify the table containing your knowledge base (example shows "growth_ai_documents") Ensure your table name matches your actual knowledge base structure Multiple tables: You can connect several tables for organized data structure The agent will automatically decide when to search the knowledge base based on user queries Step 5: Set up conversation memory (recommended) Use "Postgres Chat Memory" with n8n's built-in PostgreSQL credentials Configure table name: Choose a name for your chat history table (will be auto-created) Context Window Length: Set to 20 messages by default (adjustable based on your needs) Benefits: Conversation traceability and analytics Context retention across messages Unique conversation IDs for user sessions Stored in n8n's database, not Supabase How to customize the workflow Basic conversation features Response style: Modify prompts to change personality and tone Knowledge scope: Update Supabase tables to expand or focus the knowledge base Language support: Configure for multiple languages Response length: Set limits for concise or detailed answers Memory retention: Adjust context window length for longer or shorter conversation memory Advanced action capabilities The chatbot can be extended with additional tools for: Email automation: Send support emails when users request assistance Calendar integration: Book appointments directly in Google Calendar Database updates: Modify Airtable or other databases based on user interactions API integrations: Connect to external services and systems File handling: Process and analyze uploaded documents Platform-specific deployments Website integration Replace chat trigger with webhook trigger Configure your website's chat widget to send messages to the n8n webhook URL Handle response formatting for your specific chat interface Slack/Teams deployment Set up webhook trigger with Slack/Teams webhook URL Configure response formatting for platform-specific message structures Add platform-specific features (mentions, channels, etc.) E-commerce integration Connect to product databases Add order tracking capabilities Integrate with payment systems Configure support ticket creation Results interpretation Conversation management Chat history: All conversations stored in n8n's PostgreSQL database with unique IDs Context tracking: Agent maintains conversation flow and references previous messages Analytics potential: Historical data available for analysis and improvement Knowledge retrieval Semantic search: Vector database returns most relevant information based on meaning, not just keywords Automatic decision: Agent automatically determines when to search the knowledge base Source tracking: Ability to trace answers back to source documents Accuracy improvement: Continuously refine knowledge base based on user queries Use cases Internal applications Developer documentation: Quick access to technical guides and APIs HR support: Employee handbook and policy questions IT helpdesk: Troubleshooting guides and system information Training assistant: Learning materials and procedure guidance External customer service E-commerce support: Product information and order assistance Technical support: User manuals and troubleshooting Sales assistance: Product recommendations and pricing FAQ automation: Common questions and instant responses Specialized implementations Lead qualification: Gather customer information and schedule sales calls Appointment booking: Healthcare, consulting, or service appointments Order processing: Take orders and update inventory systems Multi-language support: Global customer service with language detection Workflow limitations Knowledge base dependency: Quality depends on source documentation and embedding setup Memory storage: Requires active n8n PostgreSQL connection for conversation history Platform restrictions: Some platforms may have webhook limitations Response time: Vector search may add slight delay to responses Token limits: Large context windows may increase API costs Embedding costs: OpenAI embeddings required for vector search functionality
by Growth AI
Who's it for Marketing teams, business intelligence professionals, competitive analysts, and executives who need consistent industry monitoring with AI-powered analysis and automated team distribution via Discord. What it does This intelligent workflow automatically monitors multiple industry topics, scrapes and analyzes relevant news articles using Claude AI, and delivers professionally formatted intelligence reports to your Discord channel. The system provides weekly automated monitoring cycles with personalized bot communication and comprehensive content analysis. How it works The workflow follows a sophisticated 7-phase automation process: Scheduled Activation: Triggers weekly monitoring cycles (default: Mondays at 9 AM) Query Management: Retrieves monitoring topics from centralized Google Sheets configuration News Discovery: Executes comprehensive Google News searches using SerpAPI for each configured topic Content Extraction: Scrapes full article content from top 3 sources per topic using Firecrawl AI Analysis: Processes scraped content using Claude 4 Sonnet for intelligent synthesis and formatting Discord Optimization: Automatically segments content to comply with Discord's 2000-character message limits Automated Delivery: Posts formatted intelligence reports to Discord channel with branded "Claptrap" bot personality Requirements Google Sheets account for query management SerpAPI account for Google News access Firecrawl account for article content extraction Anthropic API access for Claude 4 Sonnet Discord bot with proper channel permissions Scheduled execution capability (cron-based trigger) How to set up Step 1: Configure Google Sheets query management Create monitoring sheet: Set up Google Sheets document with "Query" sheet Add search topics: Include industry keywords, competitor names, and relevant search terms Sheet structure: Simple column format with "Query" header containing search terms Access permissions: Ensure n8n has read access to the Google Sheets document Step 2: Configure API credentials Set up the following credentials in n8n: Google Sheets OAuth2: For accessing query configuration sheet SerpAPI: For Google News search functionality with proper rate limits Firecrawl API: For reliable article content extraction across various websites Anthropic API: For Claude 4 Sonnet access with sufficient token limits Discord Bot API: With message posting permissions in target channel Step 3: Customize scheduling settings Cron expression: Default set to "0 9 * * 1" (Mondays at 9 AM) Frequency options: Adjust for daily, weekly, or custom monitoring cycles Timezone considerations: Configure according to team's working hours Execution timing: Ensure adequate processing time for multiple topics Step 4: Configure Discord integration Set up Discord delivery settings: Guild ID: Target Discord server (currently: 919951151888236595) Channel ID: Specific monitoring channel (currently: 1334455789284364309) Bot permissions: Message posting, embed suppression capabilities Brand personality: Customize "Claptrap" bot messaging style and tone Step 5: Customize content analysis Configure AI analysis parameters: Analysis depth: Currently processes top 3 articles per topic Content format: Structured markdown format with consistent styling Language settings: Currently configured for French output (easily customizable) Quality controls: Error handling for inaccessible articles and content How to customize the workflow Query management expansion Topic categories: Organize queries by industry, competitor, or strategic focus areas Keyword optimization: Refine search terms based on result quality and relevance Dynamic queries: Implement time-based or event-triggered query modifications Multi-language support: Add international keyword variations for global monitoring Advanced content processing Article quantity: Modify from 3 to more articles per topic based on analysis needs Content filtering: Add quality scoring and relevance filtering for article selection Source preferences: Implement preferred publisher lists or source quality weighting Content enrichment: Add sentiment analysis, trend identification, or competitive positioning Discord delivery enhancements Rich formatting: Implement Discord embeds, reactions, or interactive elements Multi-channel distribution: Route different topics to specialized Discord channels Alert levels: Add priority-based messaging for urgent industry developments Archive functionality: Create searchable message threads or database storage Integration expansions Slack compatibility: Replace or supplement Discord with Slack notifications Email reports: Add formatted email distribution for executive summaries Database storage: Implement persistent storage for historical analysis and trending API endpoints: Create webhook endpoints for third-party system integration AI analysis customization Analysis templates: Create topic-specific analysis frameworks and formatting Competitive focus: Enhance competitor mention detection and analysis depth Trend identification: Implement cross-topic trend analysis and strategic insights Summary levels: Create executive summaries alongside detailed technical analysis Advanced monitoring features Intelligent content curation The system provides sophisticated content management: Relevance scoring: Automatic ranking of articles by topic relevance and publication authority Duplicate detection: Prevents redundant coverage of the same story across different sources Content quality assessment: Filters low-quality or promotional content automatically Source diversity: Ensures coverage from multiple perspectives and publication types Error handling and reliability Graceful degradation: Continues processing even if individual articles fail to scrape Retry mechanisms: Automatic retry logic for temporary API failures or network issues Content fallbacks: Uses article snippets when full content extraction fails Notification continuity: Ensures Discord delivery even with partial content processing Results interpretation Intelligence report structure Each monitoring cycle delivers: Topic-specific summaries: Individual analysis for each configured search query Source attribution: Complete citation with publication date, source, and URL Structured formatting: Consistent presentation optimized for quick scanning Professional analysis: AI-generated insights maintaining factual accuracy and business context Performance analytics Monitor system effectiveness through: Processing metrics: Track successful article extraction and analysis rates Content quality: Assess relevance and usefulness of delivered intelligence Team engagement: Monitor Discord channel activity and report utilization System reliability: Track execution success rates and error patterns Use cases Competitive intelligence Market monitoring: Track competitor announcements, product launches, and strategic moves Industry trends: Identify emerging technologies, regulatory changes, and market shifts Partnership tracking: Monitor alliance formations, acquisitions, and strategic partnerships Leadership changes: Track executive movements and organizational restructuring Strategic planning support Market research: Continuous intelligence gathering for strategic decision-making Risk assessment: Early warning system for industry disruptions and regulatory changes Opportunity identification: Spot emerging markets, technologies, and business opportunities Brand monitoring: Track industry perception and competitive positioning Team collaboration enhancement Knowledge sharing: Centralized distribution of relevant industry intelligence Discussion facilitation: Provide common information baseline for strategic discussions Decision support: Deliver timely intelligence for business planning and strategy sessions Competitive awareness: Keep teams informed about competitive landscape changes Workflow limitations Language dependency: Currently optimized for French analysis output (easily customizable) Processing capacity: Limited to 3 articles per query (configurable based on API limits) Platform specificity: Configured for Discord delivery (adaptable to other platforms) Scheduling constraints: Fixed weekly schedule (customizable via cron expressions) Content access: Dependent on article accessibility and website compatibility with Firecrawl API dependencies: Requires active subscriptions and proper rate limit management for all integrated services
by Jitesh Dugar
👤 Who’s it for This workflow is designed for employees who need to submit expense claims for business trips. It automates the process of extracting data from receipts/invoices, logging it to a Google Sheet, and notifying the finance team via email. Ideal users: Employees submitting business trip expense claims HR or Admins reviewing travel-related reimbursements Finance teams responsible for processing claims ⚙️ How it works / What it does Employee submits a form with trip information (name, department, purpose, dates) and uploads one or more receipts/invoices (PDF). Uploaded files are saved to Google Drive for record-keeping. Each PDF is passed to a DocClaim Assistant agent, which uses GPT-4o and a structured parser to extract structured invoice data. The data is transformed and formatted into a standard JSON structure. Two parallel paths are followed: Invoice records are appended to a Google Sheet for centralized tracking. A detailed HTML email summarizing the trip and expenses is generated and sent to the finance department for claim processing. 🛠 How to set up Create a form to capture: Employee Name Department Trip Purpose From Date / To Date Receipt/Invoice File Upload (multiple PDFs) Configure file upload node to store files in a specific Google Drive folder. Set up DocClaim Agent using: GPT-4o or any LLM with document analysis capability Output parser for standardizing extracted receipt data (e.g., vendor, total, tax, date) Transform extracted data into a structured claim record (Code Node). Path 1: Save records to a Google Sheet (one row per expense). Path 2: Format the employee + claim data into a dynamic HTML email Use Send Email node to notify the finance department (e.g., finance@yourcompany.com) ✅ Requirements Jotform account with expense form setup Sign up for free here n8n running with access to: Google Drive API (for file uploads) Google Sheets API (for logging expenses) Email node (SMTP or Gmail for sending) GPT-4o or equivalent LLM with document parsing ability PDF invoices with clear formatting Shared Google Sheet for claim tracking Optional: Shared inbox for finance team 🧩 How to customize the workflow Add approval steps**: route the email to a manager before finance Attach original PDFs**: include uploaded files in the email as attachments Localize for other languages**: adapt form labels, email content, or parser prompts Sync to ERP or accounting system**: replace Google Sheet with QuickBooks, Xero, etc. Set limits/validation**: enforce max claim per trip or required fields before submission Auto-tag expenses**: add categories (e.g., travel, accommodation) for better reporting
by Oneclick AI Squad
Analyzes influencer profiles and scores authenticity before brand partnership approval. Detects fake followers, bot accounts, and suspicious engagement patterns using AI-powered behavioral analysis. 🎯 How It Works Simple 7-Node Workflow: Input → Submit influencer username and platform (Instagram/Twitter/TikTok) Fetch → Retrieve complete profile data and engagement metrics Analyze → Examine follower patterns, ratios, growth velocity, engagement AI Check → Deep behavioral analysis with Claude AI Report → Generate comprehensive fraud assessment Notify → Send detailed email report to partnership team Log → Save to database for tracking 📊 Detection Capabilities Follower Authenticity**: Analyzes follower-to-following ratio (red flag if < 0.5) Engagement Quality**: Calculates engagement rate (industry avg: 1-5%) Growth Patterns**: Detects suspicious rapid follower spikes Content Consistency**: Evaluates posting frequency and regularity Profile Completeness**: Checks verification, bio, activity AI Behavioral Analysis**: Deep pattern recognition for sophisticated fraud ⚙️ Setup Instructions 1. Configure API Access Social Platform APIs: Instagram**: Get Graph API access token from Meta for Developers Twitter**: OAuth 2.0 credentials from Twitter Developer Portal TikTok**: Business API credentials (optional) AI Analysis: Anthropic Claude API**: Get key from console.anthropic.com Used for advanced behavioral fraud detection 2. Setup Notifications Configure SMTP in "Send Report" node Update recipient email (partnerships@company.com) Customize HTML template if needed 3. Database (Optional) Create PostgreSQL table (schema below) Add database credentials to final node Skip if you don't need historical tracking Database Schema CREATE TABLE partnerships.influencer_fraud_reports ( id SERIAL PRIMARY KEY, report_id VARCHAR(255) UNIQUE, username VARCHAR(255), platform VARCHAR(50), profile_url TEXT, followers BIGINT, following BIGINT, posts INTEGER, verified BOOLEAN, authenticity_score INTEGER, risk_level VARCHAR(50), final_decision TEXT, partnership_recommendation VARCHAR(100), ai_verdict VARCHAR(50), ai_confidence VARCHAR(20), red_flags JSONB, fake_follower_estimate VARCHAR(20), detailed_analysis JSONB, created_at TIMESTAMP ); 🚀 How to Use Webhook Endpoint: POST /webhook/influencer-fraud-check Request Body: { "username": "influencer_handle", "platform": "instagram" // or "twitter", "tiktok" } Example: curl -X POST https://your-n8n.com/webhook/influencer-fraud-check \ -H "Content-Type: application/json" \ -d '{"username":"example_user","platform":"instagram"}' 📈 Scoring System Overall Authenticity Score (0-100): 80-100**: LOW RISK → Approved for partnership 60-79**: MEDIUM RISK → Requires manual review 40-59**: HIGH RISK → Caution advised 0-39**: CRITICAL RISK → Rejected Weighted Components: Follower Quality (25%) Engagement Quality (35%) Content Consistency (15%) Growth Pattern (15%) Profile Completeness (10%) Final Score = 70% Automated + 30% AI Analysis 🚩 Red Flags Detected Following-to-follower ratio > 2:1 Engagement rate < 0.5% Rapid growth (>50K followers/month) Large following with <10 posts No verification with >100K followers Bot-like comment patterns Suspicious audience demographics 💰 Cost Estimate Instagram/Twitter API**: Free tier usually sufficient Claude AI**: ~$0.10-0.20 per analysis Estimated**: $5-10/month for 50 checks 💡 Best Practices Always verify HIGH and MEDIUM risk profiles manually Cross-reference with other influencer databases Request media kit and past campaign results Trial campaigns before large commitments Monitor performance metrics post-partnership Update detection thresholds based on your findings 🎯 What You Get Detailed Report Includes: Overall authenticity score (0-100) Risk level classification Partnership recommendation (APPROVE/REVIEW/REJECT) Engagement quality analysis Fake follower percentage estimate AI behavioral insights Specific red flags and concerns Next steps and recommendations
by Growth AI
AI-powered alt text generation from Google Sheets to WordPress media Who's it for WordPress site owners, content managers, and accessibility advocates who need to efficiently add alt text descriptions to multiple images for better SEO and web accessibility compliance. What it does This workflow automates the process of generating and updating alt text for WordPress media files using AI analysis. It reads image URLs from a Google Sheet, analyzes each image with Claude AI to generate accessibility-compliant descriptions, updates the sheet with the generated alt text, and automatically applies the descriptions to the corresponding WordPress media files. The workflow includes error handling to skip unsupported media formats and continue processing. How it works Input: Provide a Google Sheets URL containing image URLs and WordPress media IDs Authentication: Retrieves WordPress credentials from a separate sheet and generates Base64 authentication Processing: Loops through each image URL in the sheet AI Analysis: Claude AI analyzes each image and generates concise, accessible alt text (max 125 characters) Error Handling: Automatically skips unsupported media formats and continues with the next item Update Sheet: Writes the generated alt text back to the Google Sheet WordPress Update: Updates the WordPress media library with the new alt text via REST API Requirements Google Sheets with image URLs and WordPress media IDs WordPress site with Application Passwords enabled Claude AI (Anthropic) API credentials WordPress admin credentials stored in Google Sheets Export Media URLs WordPress plugin for generating the media list How to set up Step 1: Export your WordPress media URLs Install the "Export Media URLs" plugin on your WordPress site Go to the plugin settings and check both ID and URL columns for export (these are mandatory for the workflow) Export your media list to get the required data Step 2: Configure WordPress Application Passwords Go to WordPress Admin → Users → Your Profile Scroll down to "Application Passwords" section Enter application name (e.g., "n8n API") Click "Add New Application Password" Copy the generated password immediately (it won't be shown again) Step 3: Set up Google Sheets Duplicate this Google Sheets template to get the correct structure. The template includes two sheets: Sheet 1: "Export media" - Paste your exported media data with columns: ID (WordPress media ID) URL (image URL) Alt text (will be populated by the workflow) Sheet 2: "Infos client" - Add your WordPress credentials: Admin Name: Your WordPress username KEY: The application password you generated Domaine: Your site URL without https:// (format: "example.com") Step 4: Configure API credentials Add your Anthropic API credentials to the Claude node Connect your Google Sheets account to the Google Sheets nodes How to customize Language: The Claude prompt is in French - modify it in the "Analyze image" node for other languages Alt text length: Adjust the 125-character limit in the Claude prompt Batch processing: Change the batch size in the Split in Batches node Error handling: The workflow automatically handles unsupported formats, but you can modify the error handling logic Authentication: Customize for different WordPress authentication methods This workflow is perfect for managing accessibility compliance across large WordPress media libraries while maintaining consistent, AI-generated descriptions. It's built to be resilient and will continue processing even when encountering unsupported media formats.