by Veena Pandian
Who is this for? SEO managers, content marketers, bloggers, and growth teams who want to automatically catch declining content performance before it's too late — without manually checking Google Search Console every week. What this workflow does This workflow runs weekly to compare your recent Google Search Console performance against a historical baseline. It identifies pages experiencing traffic decay at three severity levels, sends detailed reports via Slack and email, logs all data to a tracking sheet, and auto-generates prioritized fix tasks for your most critical pages. How it works Weekly trigger fires every Monday at 8 AM. Fetches two GSC date ranges in parallel — the last 7 days (recent) and the previous 28 days (baseline, normalized to weekly averages). Compares per-page metrics including clicks, impressions, average position, and CTR. Classifies each page into one of five signals: CRITICAL_DECAY — clicks dropped 50%+ or position fell 5+ spots with 30%+ click loss DECAYING — clicks dropped 30%+ or position fell 3+ spots EARLY_DECAY — clicks dropped 15%+ or position fell 1.5+ spots STABLE — no significant change GROWING — clicks increased 20%+ Logs all results to a Decay Log Google Sheet tab for historical trending. Builds a weekly report with summary counts, estimated clicks lost, and per-page breakdowns. Sends the report to Slack and email simultaneously. Auto-generates fix tasks for critical pages with specific recommendations (backlink audit, content refresh, CTR optimization, or technical investigation) and logs them to a Fix Tasks sheet tab. Setup steps Set environment variables in your n8n instance: GSC_SITE_URL — your verified site URL (e.g., https://yoursite.com) DECAY_SHEET_URL — URL of your Google Sheet for logging Create a Google Sheet with two tabs: Decay Log with headers: date, page_path, signal, clicks_now, clicks_before, click_change_pct, position_now, position_before, position_change, impressions_now, impression_change_pct, ctr_now Fix Tasks with headers: created, priority, page_path, page_url, signal, click_change_pct, position_change, recommended_action Connect Google Search Console OAuth2 credentials (your site must be verified in GSC). Connect Google Sheets OAuth2 credentials. Connect Slack OAuth2 credentials and configure your alert channel. Configure email (SMTP) credentials and update the recipient email address in the "Email Weekly Report" node. Activate the workflow. Requirements n8n instance (self-hosted or cloud) Google Search Console property with verified ownership Google Cloud project with Search Console API and Sheets API enabled Slack workspace with a bot configured SMTP email credentials (or swap for Gmail node) How to customize Decay thresholds** — Adjust the percentage and position-change cutoffs in the "Compare Periods and Detect Decay" code node to match your sensitivity needs. Schedule** — Change from weekly to daily or bi-weekly in the trigger node. Baseline period** — Modify the 28-day comparison window to 14 or 90 days. Row limit** — Increase the rowLimit in GSC API calls beyond 500 if you have a large site. Fix task logic** — Enhance the remediation recommendations with AI-powered content analysis or integrate with project management tools (Notion, Asana, Trello). Notifications** — Add Telegram, Discord, or Microsoft Teams alongside or instead of Slack.
by Dean Pike
Client Form → Draft → Approve → Sign → Deliver, fully automated This workflow automates the entire agreement lifecycle from client form submission to signed document delivery. It generates personalized agreements from templates, manages internal approvals, orchestrates e-signatures via Signwell, and delivers fully executed documents with complete audit trails in n8n Data Tables. Good to know Handles client data collection via JotForm with custom field mapping Automatically populates Google Doc templates with client-specific details Internal approval workflow with email-based confirmation Signwell integration for embedded e-signatures - test mode enabled by default - disable for legally binding documents Complete lifecycle tracking in n8n Data Tables (draft → approval → sent → signed) Auto-cleanup: removes documents from Signwell after completion to save storage Who's it for Service businesses, consultants, agencies, and freelancers who send agreements to clients regularly. Perfect for anyone wanting to avoid other costly e-signature platforms with limited API and automation capabilities. Signwell has an affordable entry level tier with generous API limits. If you're looking to eliminate manual document preparation, have an approval workflow, and track signatures while maintaining professional client communication, then this solution is a good fit. How it works Phase 1: Draft Creation JotForm trigger captures client submission (company name, address, contact details, position) Standardizes form data and duplicates Google Doc template with custom filename Replaces template variables with client information (company name, address, full name, position, dates) Creates clean document URL and logs initial record to Data Tables Emails internal team with draft review link and client details Phase 2: Approval & Preparation Gmail monitors inbox for "Approved" reply email Fetches agreement record from Data Tables and marks as approved Downloads Google Doc as PDF and uploads to Drive folder Grants temporary public sharing access (required for Signwell file import) Creates Signwell document with embedded signature fields and signing URL Emails client with personalized signing link Revokes public sharing access for security and updates Data Tables with Signwell details Phase 3: Signature & Delivery Gmail monitors for Signwell completion notification Extracts signed document download link from notification email Downloads fully executed PDF from Signwell Uploads to "Final Versions" folder in Google Drive Updates Data Tables with completion status and final document URLs Sends confirmation email to client with signed PDF attached Deletes document from Signwell to free up storage Requirements JotForm account (free tier works) Gmail account with OAuth2 access Google Drive account (OAuth2) Google Docs account (OAuth2) with a draft Agreement template Signwell account with API key n8n Data Tables (built-in, no external service needed) Google Drive folders: "Services Agreements - Drafts" and "Services Agreements - Final Versions" How to set up Add credentials: JotForm API, Gmail OAuth2, Google Drive OAuth2, Google Docs OAuth2, Signwell API key Create JotForm: Build form with fields: Company Name, Company Address (address field), Full Name (name field), Your Position/Job Title, Email In "JotForm Trigger" node: select your form Create Google Doc template: Add variables {{clientCompanyName}}, {{clientFullName}}, {{clientNamePosition}}, {{clientCompanyAddress}}, {{agreementDate1}}, {{agreementDate2}} In "Copy and Rename File" node: select your template document and update folder ID to your "Drafts" folder Create Data Table: Name it "Services Agreements" with columns: documentFileName, clientEmail, clientFullName, clientNamePosition, clientCompanyName, clientCompanyAddress, documentUrl, approvalStatus, sentDocumentPdfUrl, sentDate, signwellUrl, signwellDocID, docSigned, finalExecutedDocGDrive, finalExecutedDocSignwellUrl In "Insert Row" and all "Get/Update Row" nodes: select your Data Table Create Gmail labels: "_AGREEMENTS" with 2 nested (sublabels) Agreement-Approvals" and "Agreement-Completed" for filtering In "Check for Email Approval" node: select your approval label and update internal email address In "Check Email for Completed Notification" node: select your completed label In "Create Document in Signwell" node: update API key and adjust signature field coordinates for your document Set Signwell to live mode: Change "test_mode": true to "test_mode": false when ready for production Activate workflow Customizing this workflow Change template variables: Edit "Update New File" node to add/remove fields (e.g., pricing, terms, scope of work) Modify approval email: Edit "Share Email Draft" node to change recipient, subject line, or message format Adjust Signwell fields: Edit "Create Document in Signwell" node to change signature/date field positions (x, y coordinates) to match your agreement template, and add any other fields you'd like Add approval deadline: Add Wait node with timeout after "Share Email Draft" to auto-remind for pending approvals Multi-signer support: Modify "Create Document in Signwell" recipients array to add multiple signers (e.g., both parties) Change storage folders: Update folder IDs in "Upload PDF File" and "Upload Completed Doc" nodes Add Slack notifications: Add Slack nodes after key milestones (draft created, approved, signed) Custom client messaging: Edit "Send Prepared Agreement to Client" and "Send Client Completed Agreement PDF" nodes for personalized communication Add reminder logic: Insert Wait + Send Email nodes between signing and completion to remind client if not signed within X days Quick Troubleshooting JotForm not triggering: Verify webhook is active in JotForm settings and form ID matches "JotForm Trigger" node Template variables not replacing: Check variable names in template doc exactly match {{variableName}} format (case-sensitive) Wrong internal email for approval: Update email address in "Share Email Draft" node to your own email Approval email not detected: Confirm Gmail label "Agreement-Approvals" exists and reply contains exact word "Approved" Signwell document creation fails: Verify PDF has public sharing enabled before API call AND Signwell API key is valid in "Create Document in Signwell" node Signature fields in wrong position: Adjust x/y coordinates in "Create Document in Signwell" node (test in Signwell UI first to find correct pixel positions) Completed document not downloading: Check Signwell completion email format - Code node extracts link via regex pattern Data Tables errors: Ensure documentFileName exactly matches between "Insert Row" and "Get/Update Row" operations Client emails not sending: Re-authorize Gmail OAuth2 credentials and verify sender name/address in Gmail nodes Drive folder not found: Update folder IDs in "Copy and Rename File", "Upload PDF File", and "Upload Completed Doc" nodes to your own folder IDs Signwell deletion fails: Verify signwellDocID was correctly stored in Data Tables before deletion (check "Update Row - Additional Doc Details" output) 401/403 API errors: Re-authorize all OAuth2 credentials (Gmail, Google Drive, Google Docs) Test mode documents: Change "test_mode": true to "test_mode": false in "Create Document in Signwell" node for production signatures Sample Outputs Agreement Drafts and Final folders/files in Google Drive File References Agreement Template (sample) Final Agreement Signed (sample)
by Seb
Stripe invoicing automation that is connected to your CRM, in this example, it is ClickUp. At the end of the flow, once your lead has been sent an invoice, you (or your team) will be sent an email notifying you of the newly sent invoice with all relevant details. How it works: • Monitors ClickUp task status → triggers workflow when status changes to send invoice. • Fetches task details from ClickUp, including customer name, email, and project cost. • Creates a Stripe customer using the fetched information. • Generates a Stripe invoice via HTTP request, including description, footer, and due date (calculated in Unix timestamp). • Adds invoice items automatically with correct amounts (converted to cents for Stripe). • Sends the invoice to the customer automatically (manual or auto-charge option). • Sends notification emails to team members with a link to the ClickUp task. Works with other CRMs like Monday or HubSpot, not just ClickUp. Test mode is available in Stripe to validate the workflow without sending real invoices. Setup steps: • You will need to connect your ClickUp Account • Connect your Stripe Account via HTTP Request (Shown in YouTube Video Linked Below) • You will need to connect your email account to N8N (Gmail, Outlook etc) for sending the emails to your team and the client Important Have your Stripe account and PUT IT IN TEST/DEVELOPER MODE when testing and developing the automation. Alternatively, set up an entirely separate account from your main Stripe account. This is only up until the point where you want to send the invoice, as you cannot send an invoice when your Stripe account is in test/developer mode For a complete rundown on how to set this up watch my YouTube tutorial linked below See full video tutorial here: https://youtu.be/vthK5I8x33k?si=W0Nreu403pDs-ud3 My LinkedIn: https://www.linkedin.com/in/seb-gardner-5b439a260/
by Cheng Siong Chin
How It Works Automates monthly revenue aggregation from multiple sources with intelligent tax forecasting using GPT-4 structured analysis. Fetches revenue data from up to three distinct sources, consolidates datasets into unified records, applies OpenAI GPT-4 model for predictive tax obligation forecasting with context awareness. System generates formatted reports with structured forecast outputs and automatically sends comprehensive tax projections to agents via Gmail, storing results in Google Sheets for audit trails. Designed for tax professionals, accounting firms, and finance teams requiring accurate predictive tax planning, cash flow forecasting, and proactive compliance strategy without manual calculations. Setup Steps Configure OpenAI API key for GPT-4 model access Connect three revenue data sources with appropriate credentials Map data aggregation logic for multi-source consolidation Define structured output schema for forecast results Set up Gmail for automated agent notification Configure Google Sheets destination Prerequisites OpenAI API key with GPT-4 access, Gmail account, Google Sheets, three revenue data source credentials Use Cases Monthly tax liability projections, quarterly estimated tax planning Customization Adjust forecast model parameters, add additional revenue sources, modify email templates Benefits Eliminates manual tax calculations, enables proactive tax planning, improves cash flow forecasting accuracy
by Evervise
Transform database design from weeks to minutes with this intelligent multi-agent system. Perfect for agencies, consultancies, and SaaS companies offering database architecture as a lead magnet or service. 🤖 4 Specialized AI Agents: Agent 1 (Architect):** Designs complete schema with tables, relationships, indexes Agent 2 (Reviewer):** Validates design for performance, security, scalability Agent 3 (Optimizer):** Adds advanced features and scores the design (0-100) Agent 4 (SQL Generator):** Creates production-ready migration scripts 🔄 Smart Quality Loop: Automatically retries up to 3 times if score falls below B grade, feeding previous feedback to improve the design iteratively. ✨ What You Get: Complete database schema (JSON) Comprehensive score card with letter grade Review feedback with severity levels (Critical/High/Medium/Low) Production-ready SQL migration script Optional auto-execution in PostgreSQL/MySQL Iteration count and optimization recommendations 💼 Perfect For: Digital agencies offering database design services SaaS companies needing rapid prototyping Consultancies creating lead magnets Developers modernizing legacy systems Startups validating data models 🎯 Use as Lead Magnet: Offer free database blueprints to capture leads, then upsell implementation, custom automations, and ongoing optimization services. ⚙️ Technical Highlights: Optimized temperature settings per agent (0.1-0.5) Claude Sonnet 4.5 for maximum quality Structured JSON output for easy integration Error handling and graceful degradation Execution time: 60-90 seconds average Cost: ~$0.15-0.30 per run Use Cases Agency Lead Magnet Capture leads by offering free database architecture reviews and blueprints Rapid Prototyping Quickly generate database schemas for MVP development and validation Legacy System Modernization Help companies redesign outdated database structures with modern best practices Technical Consulting Provide instant database assessments and recommendations to clients Educational Tool Teach database design principles through AI-generated examples and feedback Pre-Sales Tool Demonstrate technical expertise to prospects before engagement Key Features ✅ Multi-agent AI collaboration with specialized roles ✅ Automatic quality control and iterative improvement (max 3 retries) ✅ Support for PostgreSQL, MySQL, MSSQL, MariaDB ✅ Production-ready SQL script generation ✅ Comprehensive scoring system (Schema/Performance/Scalability/Security) ✅ Optional automatic SQL execution ✅ Detailed feedback with actionable recommendations ✅ Customizable form fields for different industries ✅ Error handling and graceful failures ✅ Complete audit trail of all agent decisions Setup Instructions PREREQUISITES: Anthropic API key (Claude Sonnet 4.5 access) PostgreSQL/MySQL database (optional, for auto-execution) n8n version 1.0+ with LangChain nodes CONFIGURATION STEPS: Import the workflow JSON into your n8n instance Configure Anthropic API credentials: Add your Anthropic API key in n8n credentials Connect all 4 AI model nodes to your credential (Optional) Configure database connection: In "Execute SQL in PostgreSQL" node, add your database credentials Use a TEST/SANDBOX database, never production Or disable this node if you prefer manual execution Customize the form (optional): Edit form fields in "On form submission" node Add industry-specific questions Adjust required fields based on your needs Test the workflow: Use the form URL to submit a test request Check execution time and quality Verify all agents are responding correctly Customize agent prompts (optional): Adjust system messages for industry-specific requirements Modify scoring criteria in Agent 3 Add custom validation rules in Agent 2 Deploy: Share the form URL as your lead magnet Embed in website or landing pages Set up email notifications for submissions COST CONSIDERATIONS: Each execution costs ~$0.15-0.30 in API calls Failed attempts (retries) increase cost Consider rate limiting for public forms Requirements REQUIRED: Anthropic API Key (Claude access) n8n version 1.0+ LangChain nodes enabled OPTIONAL: PostgreSQL/MySQL database connection (for auto-execution) Email service (for result delivery) CRM integration (for lead capture) Tags #ai-agents #database-design #postgresql #mysql #lead-generation #automation #langchain #claude #schema-design #multi-agent #consulting-tool #saas-tool #development #code-generation #sql-generator 📖 Website: https://evervise.ai/ ✨ Support: mark.marin@evervise.com N8N Link
by yuta tokumitsu
Automate intelligent customer support responses with AI and Slack How it works Receive request via webhook with customer question Analyze sentiment and detect urgency using JavaScript Send urgent alerts to Slack for critical cases Search knowledge base and fetch conversation history from PostgreSQL Generate AI response with context-aware prompts Route intelligently: Auto-respond via email OR escalate to Slack Log everything to Google Sheets and PostgreSQL for analytics Setup steps Slack webhooks: Replace YOUR_URGENT_WEBHOOK and YOUR_ESCALATION_WEBHOOK with your webhook URLs Google Sheets: Replace YOUR_SPREADSHEET_ID with your spreadsheet ID and authenticate Email: Configure SMTP/Gmail credentials in the email node PostgreSQL (optional): Create support_conversations table or disable DB nodes Production: Replace mock AI nodes with OpenAI/Anthropic API nodes Key features Multi-language support (Japanese & English) Sentiment analysis with urgency detection Smart escalation routing Real-time Slack notifications Comprehensive analytics logging
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Keep your SEO performance on track with this automated SEO Watchlist Monitor! This workflow combines AI-powered strategy analysis with real-time search ranking checks to track keyword positions, identify content gaps, and alert you to critical ranking drops. Perfect for marketing teams ensuring search visibility and competitive intelligence across platforms. 🚀🔍 What This Template Does 1️⃣ Triggers daily SEO intelligence checks to monitor keyword performance. 2️⃣ Configures target keywords, competitor domains, and geographic focus. 3️⃣ Validates SEO configuration to ensure proper setup. 4️⃣ Uses AI to analyze keyword competitiveness and strategic opportunities. 5️⃣ Checks real-time search rankings using Google Search scraper. 6️⃣ Detects critical ranking drops below position 10. 7️⃣ Saves SEO intelligence to Google Sheets for tracking. 8️⃣ Sends email alerts for urgent ranking issues. 9️⃣ Provides daily Slack summaries of SEO performance. Key Benefits ✅ Monitors keyword rankings and competitor movements daily ✅ Identifies content gaps and strategic opportunities with AI analysis ✅ Alerts instantly to critical ranking drops for quick action ✅ Centralizes SEO intelligence in Google Sheets for team visibility ✅ Combines AI insights with real-time search data for comprehensive monitoring Features Daily automated schedule for continuous monitoring AI-powered SEO strategy analysis and competitive intelligence Real-time search ranking checks using Decodo scraper Critical alert system for ranking drops Google Sheets integration for data centralization Slack and Gmail notifications for team awareness Configuration validation and error logging Structured data parsing for consistent reporting Requirements OpenAI API credentials for AI analysis Decodo API credentials for search scraping Google Sheets OAuth2 credentials with edit access Gmail OAuth2 credentials for email alerts Slack Bot Token with chat:write permission Environment variables for configuration settings Target Audience SEO and digital marketing teams 🎯 Content strategy and growth teams 📈 Competitive intelligence professionals 🔍 Marketing operations teams 🚀 Agency account managers managing multiple clients 💼 Step-by-Step Setup Instructions 1️⃣ Connect OpenAI credentials for AI analysis capabilities 2️⃣ Set up Decodo API credentials for search scraping functionality 3️⃣ Configure Google Sheets with required headers (Keyword, Rank, description, etc.) 4️⃣ Add Gmail and Slack credentials for alerting and notifications 5️⃣ Set your target keywords, competitors, and geographic focus in the configuration node 6️⃣ Configure the cron schedule (hourly) for daily monitoring frequency 7️⃣ Run once manually to verify all integrations and data flow 8️⃣ Activate for ongoing SEO performance tracking and alerting ✅ Pro Tip: Use coupon code "YARON" to get 23K requests for testing (in Decodo)
by Praneel S
⚠️ Disclaimer: This workflow uses WhatsApp, Google Calendar, and Gmail nodes that must be configured manually. Who’s it for This workflow is built for professionals, teams, and automation enthusiasts who want to manage their Google Calendar and Gmail directly from WhatsApp, powered by an AI assistant using OpenAI GPT or Google Gemini. It enables users to chat naturally through WhatsApp to schedule meetings, send emails, and check events — all without opening Gmail or Google Calendar. How it works The WhatsApp Trigger node captures incoming messages from users. The AI Agent (powered by Gemini or GPT) interprets user queries and determines the best tool to use. The Simple Memory node keeps context between messages using the user’s phone number. The Google Calendar nodes handle: Listing, creating, and updating events. Checking your availability before scheduling. The Gmail nodes handle: Sending emails. Reading and summarizing recent messages. The Date & Time node converts natural language like “next Monday at 3 PM” into proper ISO time format. The assistant responds via Send WhatsApp Response, sending clear confirmations and replies. Features Manage Gmail and Calendar entirely via WhatsApp. AI-powered understanding of natural language commands. Integrated with Google Meet for automatic conferencing links. Short-term memory for context retention. Fully modular – swap Gemini with OpenAI GPT or any LLM. Setup Steps Configure WhatsApp Cloud API via Meta for Developers. Set up Google Calendar and Gmail OAuth2 credentials. Add your Google API keys and calendar email. Connect your OpenAI or Gemini model credentials. Activate and test the workflow with messages like: “Schedule a meeting tomorrow at 5 PM.” “Check my latest emails.” “Send an email to alex@example.com about our project.” Requirements n8n instance (self-hosted or cloud) WhatsApp Business API (Meta Developer Account) Google Workspace or Gmail account OpenAI API key or Google Gemini API key Properly configured webhooks for WhatsApp Trigger Example Prompts “What’s on my calendar this week?” “Email John to confirm our meeting.” “When am I free tomorrow afternoon?” Customization Replace Gemini with OpenAI GPT in the AI Agent node. Adjust memory length for longer or shorter conversations. Add Slack or Teams notification nodes. Modify the prompt personality or response tone. Credits Created by Praneel For detailed setup help, visit praneel.tech/contact
by Habeeb Mohammed
Who's it for This workflow transforms hours of manual video editing into an automated AI-powered pipeline. Perfect for anyone looking to repurpose long-form content into viral short-form clips. Ideal users include: Content Creators** - YouTubers producing long-form videos who want to maximize reach by automatically generating TikTok, Reels, and Shorts from their content Social Media Managers** - Agencies and freelancers handling multiple clients who need to scale clip production without hiring additional editors Podcasters** - Audio and video podcast hosts wanting to create promotional clips highlighting the best moments from each episode Video Editors** - Professional editors looking to automate repetitive clipping tasks and focus on creative decisions rather than technical execution Marketing Teams** - B2B and B2C teams extracting key moments from webinars, product demos, tutorials, and educational content for social campaigns Whether you're a solo creator or managing content at scale, this workflow saves 5-10 hours per video while maintaining professional quality output. How it works This workflow combines AI analysis with professional video editing tools to automatically identify and produce viral-ready clips from any YouTube video. The process flows through three main stages: Stage 1: Download and Analysis Submit a YouTube URL through the built-in form trigger yt-dlp simultaneously downloads the video in highest quality and extracts subtitles or auto-generated transcripts The transcript is intelligently chunked into 150-segment batches for optimal AI processing Each batch is analyzed by Gemini AI using specialized prompts that evaluate viral potential based on hooks, pacing, emotional peaks, and engagement triggers AI identifies 3-5 high-quality moments per batch and assigns virality scores to each potential clip Stage 2: Clip Selection and Extraction All AI-identified clips are merged and sorted by their virality scores The top 10 candidates are automatically selected for processing FFmpeg extracts each clip segment from the original video at precise timestamps Clips are processed sequentially to prevent system overload Stage 3: Professional Editing Pipeline Each clip enters a multi-stage editing subworkflow with automated operations: Smart 9:16 cropping that intelligently frames the subject for vertical platforms Precise trimming to remove dead air and optimize pacing Dynamic subtitle generation with sizing calculated based on video resolution Professional subtitle styling including bold text, high-contrast colors, strategic positioning, and text wrapping Subtitles are burned directly into the video as permanent overlays Final Delivery: The workflow processes clips with configurable wait times to match your system's capabilities. When all clips complete processing, you receive an email notification and find your social-ready clips in the /data/clips/ directory, ready for upload to any platform. Requirements ⚠️ Self-hosted n8n only - This workflow requires command-line access and cannot run on n8n Cloud due to its dependency on system-level tools. System dependencies you must install: FFmpeg** - Industry-standard video processing tool for trimming, cropping, and subtitle burning. Install on your n8n host system following this comprehensive guide. Most Linux systems can install via package manager: apt-get install ffmpeg or yum install ffmpeg. yt-dlp** - Advanced YouTube downloader that handles video and subtitle extraction. Follow official installation instructions. Recommended: pip install yt-dlp or direct binary download. FFprobe** - Usually included with FFmpeg, used for detecting video dimensions for dynamic subtitle sizing. Credentials needed: Google Gemini API account** - Powers the AI analysis for clip identification and editing instructions. Get your free API key with generous free tier limits. Gmail OAuth2 credentials** - Enables email notifications when clips are ready. Set up through n8n's credential system. Storage requirements: Ensure /data/clips/ directory exists with write permissions Plan for 2-3x the original video size in temporary storage during processing Final clips typically use 10-30% of original video size How to set up Step 1: Install system dependencies SSH into your n8n host and install required tools. For Ubuntu/Debian systems, run: apt-get update apt-get install ffmpeg pip install yt-dlp Verify installations by running ffmpeg -version and yt-dlp --version. Step 2: Configure directory structure Create the clips output directory with proper permissions: mkdir -p /data/clips chmod 755 /data/clips Step 3: Import the workflow Download the workflow JSON and import it into your n8n instance. You'll see several sticky notes color-coded by stage: yellow for description, blue for download/analysis, pink for editing operations, and green for clipping. Step 4: Set up credentials Navigate to the "viral clips identification" node and add your Google Gemini API credentials. The workflow uses the gemini-2.5-flash model for optimal speed and quality balance. Then configure Gmail OAuth2 in the "Send a message" node following n8n's authentication wizard. Step 5: Update email notification Open the "Send a message" node and replace habeebmohammedfaiz@gmail.com with your email address. Step 6: Create the editing subworkflow The workflow references a separate subworkflow for the editing pipeline. Create a new workflow in n8n, copy all nodes from the "EDITING" section (between the Execute Workflow Trigger and the final output), and save it. Note the workflow ID from the URL. Step 7: Link the subworkflow In the main workflow, open the "Call subworkflow" node and update the workflow ID to match your newly created editing workflow. Step 8: Test with a short video Start with a 5-10 minute YouTube video for your first test. Use the manual trigger or form submission. Monitor the execution to ensure all nodes complete successfully and clips appear in /data/clips/. Step 9: Adjust performance settings Based on your system's performance during the test, modify the Wait node durations. Systems with 8GB+ RAM and modern CPUs can reduce wait times to 30 seconds. Limited systems should keep 60-second waits or increase them. How to customize the workflow Adjust clip quantity and quality thresholds Open the "filter out top clips according to score" node. The code currently uses .slice(0, 10) to select the top 10 clips. Change this number based on your needs: use .slice(0, 5) for only the best clips, or .slice(0, 20) for more options. You can also add score filtering by adding results.filter(c => c.score > 0.7) before the slice operation to only include clips with virality scores above 70%. Customize subtitle appearance Navigate to the "calculate relative subtitle size" node. The JavaScript code defines several styling variables you can modify: fontSize - Currently calculated dynamically, but you can hardcode it: const fontSize = 48; fontName - Change from Arial to any system font: const fontName = 'Impact'; primaryColor - Modify text color using BGR hex format: '&H00FF00&' for green, '&HFF0000&' for red borderColor - Adjust outline color for better contrast outlineWidth - Increase from 1 to 2 or 3 for thicker borders marginV - Control vertical position (higher values move text up from bottom) Modify AI analysis prompts In the "viral clips identification" node, edit the Gemini prompt to target specific content types. For educational content, add "Focus on key teaching moments and actionable tips." For entertainment, emphasize "Identify funny moments, reactions, and unexpected events." For podcast clips, specify "Extract controversial opinions, storytelling segments, and quotable statements." Change aspect ratios The workflow defaults to 9:16 for vertical video. To create horizontal clips for YouTube or other platforms, open the "Analyze the actual whole video" node and change the aspect ratio in the JSON schema from "aspect_ratio": "9:16" to "aspect_ratio": "16:9". The AI will automatically adjust cropping coordinates accordingly. Enable audio normalization By default, audio normalization is disabled for faster processing. To enable it, open the "extract all actionable operations" node, find the audio_normalize task object, and change enabled: false to enabled: true. This ensures consistent volume levels across all clips but adds processing time. Add custom editing operations The editing pipeline is modular. You can add new operations like: Color grading by inserting FFmpeg color filters Logo overlays by adding watermark commands Intro/outro sequences by concatenating video files Background music by mixing audio tracks Add these as new task objects in the "extract all actionable operations" node following the existing pattern. Customize notification content Open the "Send a message" node to modify the email subject, body text, or add clip details. You can include clip metadata like timestamps, scores, and descriptions using expressions like {{ $json.hook }} or {{ $json.score }}. Integrate with cloud storage Add nodes after clip generation to automatically upload finished clips to Google Drive, Dropbox, AWS S3, or any n8n-supported storage service. Use the Loop Over Items1 output to access completed clip file paths. Schedule automated processing Replace the Form Trigger with a Schedule Trigger to automatically process videos from a spreadsheet or RSS feed. Combine with Google Sheets integration to maintain a queue of videos to process overnight.
by Pixcels Themes
Who’s it for This template is for clinics, hospitals, care teams, and telemedicine providers who need a structured, automated system for post-surgery follow-up. It helps reduce manual workload while ensuring every patient gets timely check-ins and appropriate triage. What it does / How it works This workflow automates daily recovery monitoring using Google Sheets and Telegram. It sends scheduled check-in messages to all patients within their follow-up window. When a patient replies, the message is: Captured by Telegram Trigger Cleaned and structured Summarized by an AI agent Classified into low, moderate, or high intensity Based on the intensity level: Low:** Sends a supportive, non-urgent response Moderate:** Sends guidance + schedules a follow-up event in Google Calendar High:** Sends an alert email to the doctor via Gmail All logic runs automatically. Requirements Google Sheets OAuth2 credentials Gmail OAuth2 credentials Google Calendar OAuth2 credentials Telegram Bot credentials Gemini API credentials A Google Sheet with patient name, surgery type, follow-up duration, and doctor email How to set up Connect all required credentials inside n8n. Replace the Google Sheet ID with your own patient sheet. Adjust column mappings if your sheet structure differs. Test by sending a Telegram message to your bot. Enable the Schedule Trigger to begin automated daily follow-ups. How to customize the workflow Modify AI prompts inside the AI Agent nodes Adjust triage logic for intensity levels Change follow-up intervals in the Schedule Trigger Add additional notification channels (SMS, Slack, CRM logging)
by Oneclick AI Squad
This n8n workflow automates the monitoring, health assessment, and self-healing of AWS EC2 instances in production environments. It runs periodic checks, identifies unhealthy instances based on status and metrics, restarts them automatically, and notifies teams via multi-channel alerts while logging data for auditing and reporting. Key Features Triggers health checks every 5 minutes to proactively monitor EC2 fleet status. Fetches and loops through all production EC2 instances for individualized analysis. Evaluates instance health using AWS metrics and custom thresholds to detect issues like high CPU or stopped states. Performs automatic restarts on unhealthy instances to minimize downtime. Sends instant WhatsApp notifications for urgent alerts, detailed email reports for team review, and logs metrics to Google Sheets for long-term tracking. Includes sticky notes for quick reference on configuration, self-healing logic, and alert setup. Workflow Process The Schedule Trigger node runs the workflow every 5 minutes, ensuring frequent health monitoring without overwhelming AWS APIs. The Get EC2 Instances node fetches all production-tagged EC2 instances from AWS, filtering by environment (e.g., tag: Environment=Production). The Loop Over Instances node iterates through each fetched instance individually, allowing parallel processing for scalability. The Check Instance Status node retrieves detailed health metrics for the current instance via AWS API (e.g., status checks, CPU utilization, and state). The Health Status Check node evaluates the instance's status against predefined thresholds (e.g., failed system checks or high load); if healthy, it skips to logging. The Analyze Health Data node assesses metrics in depth to determine action (e.g., restart if CPU > 90% for 5+ minutes) and prepares alert payloads. The Restart Instance node automatically initiates a reboot on unhealthy instances using AWS EC2 API, with optional dry-run mode for testing. The WhatsApp Notification node (part of Multi-Channel Alerts) sends instant alerts via Twilio WhatsApp API, including instance ID, issue summary, and restart status. The Email Report node generates and sends a detailed HTML report to the team via SMTP, summarizing checked instances, actions taken, and metrics trends. The Google Sheets Logging node appends health data, timestamps, and outcomes to a specified spreadsheet for historical analysis and dashboards. The Sticky Notes nodes provide inline documentation: one for AWS credential setup, one explaining self-healing thresholds, and one for alert channel configurations. Setup Instructions Import the workflow into n8n and activate the Schedule Trigger with a 5-minute cron expression (e.g., */5 * * * *). Configure AWS credentials in the Get EC2 Instances, Check Instance Status, and Restart Instance nodes using IAM roles with EC2 read/restart permissions. Set up Twilio credentials in the WhatsApp Notification node, including your Twilio SID, auth token, and WhatsApp-enabled phone numbers for sender/receiver. Add SMTP credentials (e.g., Gmail or AWS SES) in the Email Report node, and update sender/receiver email addresses in the node parameters. Link Google Sheets in the Google Sheets Logging node by providing the spreadsheet ID, sheet name, and OAuth credentials for write access. Customize health thresholds in Health Status Check and Analyze Health Data (e.g., via expressions for CPU/memory limits). Test the workflow by manually executing it on a small set of instances and verifying alerts/logging before enabling production scheduling. Review sticky notes within n8n for quick tips, and monitor executions in the dashboard to fine-tune intervals or error handling. Prerequisites AWS account with EC2 access and IAM user/role for DescribeInstances, DescribeInstanceStatus, and RebootInstances actions. Twilio account with WhatsApp sandbox or approved number for notifications. SMTP email service (e.g., Gmail, Outlook) with app-specific passwords enabled. Google Workspace or personal Google account for Sheets integration. n8n instance with AWS, Twilio, SMTP, and Google Sheets nodes installed (cloud or self-hosted). Production EC2 instances tagged consistently (e.g., Environment=Production) for filtering. Modification Options Adjust the Schedule Trigger interval to hourly for less frequent checks or integrate with AWS CloudWatch Events for dynamic triggering. Expand Analyze Health Data to include advanced metrics (e.g., disk I/O via CloudWatch) or ML-based anomaly detection. Add more alert channels in Multi-Channel Alerts, such as Slack webhooks or PagerDuty integrations, by duplicating the WhatsApp/Email branches. Enhance Google Sheets Logging with charts or conditional formatting via Google Apps Script for visual dashboards. Implement approval gates in Restart Instance (e.g., via email confirmation) to prevent auto-restarts in sensitive environments. Explore More AI Workflows: Get in touch with us for custom n8n automation!
by Emilio Loewenstein
Turn your sales and onboarding calls into actionable insights — automatically! This workflow connects Fireflies.ai with OpenAI to analyze and grade your call transcripts. The results, along with your lead’s details, are logged directly into a Google Sheet for easy tracking. Plus, you’ll get an instant Slack or Gmail notification with the evaluation so you can take quick action. 🚀 What It Does Triggers on new Fireflies.ai transcripts** Uses AI to evaluate and grade your calls** Logs lead + scoring data into Google Sheets** Sends instant updates via Slack or Gmail** 💡 Why It’s Valuable Save hours of manual call reviews Keep a consistent, unbiased scoring system Centralize data for reporting and coaching Act faster with real-time notifications Perfect for sales, onboarding, or customer success teams who want to improve call quality at scale while saving time. 🛠️ Setup Instructions Connect Fireflies.ai – Enable transcript export from your Fireflies.ai account. Integrate with OpenAI – Use the provided API key to analyze and score transcripts automatically. Configure Google Sheets – Create a sheet with columns for: Lead Name Company Call Date Transcript Link AI Score Notes/Insights Enable Notifications – Connect Slack or Gmail to receive instant alerts with evaluation details. Test & Launch – Run a sample call to ensure transcripts flow correctly into the sheet and notifications are triggered. 🔄 Detailed Workflow A new call transcript is generated in Fireflies.ai. The transcript is sent to OpenAI, where the call is evaluated and scored based on quality, engagement, and outcomes. The results + lead data are logged automatically into Google Sheets for centralized tracking. A Slack or Gmail notification instantly alerts your team with the score and key insights, so you can take immediate action. 📊 Google Sheets Your Google Sheet should include the following columns: Lead Name** Email/Contact** Company Name** Call Date & Time** Transcript URL** AI Evaluation Score** Summary/Next Steps** This structure ensures clarity, easy reporting, and consistent data across all calls. ⚠️ Community Node Disclaimer This workflow is created with community nodes and integrations. Please review security and API key management best practices before deploying in production. 🖼️ Workflow Template