by Evervise
Transform database design from weeks to minutes with this intelligent multi-agent system. Perfect for agencies, consultancies, and SaaS companies offering database architecture as a lead magnet or service. 🤖 4 Specialized AI Agents: Agent 1 (Architect):** Designs complete schema with tables, relationships, indexes Agent 2 (Reviewer):** Validates design for performance, security, scalability Agent 3 (Optimizer):** Adds advanced features and scores the design (0-100) Agent 4 (SQL Generator):** Creates production-ready migration scripts 🔄 Smart Quality Loop: Automatically retries up to 3 times if score falls below B grade, feeding previous feedback to improve the design iteratively. ✨ What You Get: Complete database schema (JSON) Comprehensive score card with letter grade Review feedback with severity levels (Critical/High/Medium/Low) Production-ready SQL migration script Optional auto-execution in PostgreSQL/MySQL Iteration count and optimization recommendations 💼 Perfect For: Digital agencies offering database design services SaaS companies needing rapid prototyping Consultancies creating lead magnets Developers modernizing legacy systems Startups validating data models 🎯 Use as Lead Magnet: Offer free database blueprints to capture leads, then upsell implementation, custom automations, and ongoing optimization services. ⚙️ Technical Highlights: Optimized temperature settings per agent (0.1-0.5) Claude Sonnet 4.5 for maximum quality Structured JSON output for easy integration Error handling and graceful degradation Execution time: 60-90 seconds average Cost: ~$0.15-0.30 per run Use Cases Agency Lead Magnet Capture leads by offering free database architecture reviews and blueprints Rapid Prototyping Quickly generate database schemas for MVP development and validation Legacy System Modernization Help companies redesign outdated database structures with modern best practices Technical Consulting Provide instant database assessments and recommendations to clients Educational Tool Teach database design principles through AI-generated examples and feedback Pre-Sales Tool Demonstrate technical expertise to prospects before engagement Key Features ✅ Multi-agent AI collaboration with specialized roles ✅ Automatic quality control and iterative improvement (max 3 retries) ✅ Support for PostgreSQL, MySQL, MSSQL, MariaDB ✅ Production-ready SQL script generation ✅ Comprehensive scoring system (Schema/Performance/Scalability/Security) ✅ Optional automatic SQL execution ✅ Detailed feedback with actionable recommendations ✅ Customizable form fields for different industries ✅ Error handling and graceful failures ✅ Complete audit trail of all agent decisions Setup Instructions PREREQUISITES: Anthropic API key (Claude Sonnet 4.5 access) PostgreSQL/MySQL database (optional, for auto-execution) n8n version 1.0+ with LangChain nodes CONFIGURATION STEPS: Import the workflow JSON into your n8n instance Configure Anthropic API credentials: Add your Anthropic API key in n8n credentials Connect all 4 AI model nodes to your credential (Optional) Configure database connection: In "Execute SQL in PostgreSQL" node, add your database credentials Use a TEST/SANDBOX database, never production Or disable this node if you prefer manual execution Customize the form (optional): Edit form fields in "On form submission" node Add industry-specific questions Adjust required fields based on your needs Test the workflow: Use the form URL to submit a test request Check execution time and quality Verify all agents are responding correctly Customize agent prompts (optional): Adjust system messages for industry-specific requirements Modify scoring criteria in Agent 3 Add custom validation rules in Agent 2 Deploy: Share the form URL as your lead magnet Embed in website or landing pages Set up email notifications for submissions COST CONSIDERATIONS: Each execution costs ~$0.15-0.30 in API calls Failed attempts (retries) increase cost Consider rate limiting for public forms Requirements REQUIRED: Anthropic API Key (Claude access) n8n version 1.0+ LangChain nodes enabled OPTIONAL: PostgreSQL/MySQL database connection (for auto-execution) Email service (for result delivery) CRM integration (for lead capture) Tags #ai-agents #database-design #postgresql #mysql #lead-generation #automation #langchain #claude #schema-design #multi-agent #consulting-tool #saas-tool #development #code-generation #sql-generator 📖 Website: https://evervise.ai/ ✨ Support: mark.marin@evervise.com N8N Link
by Praneel S
⚠️ Disclaimer: This workflow uses WhatsApp, Google Calendar, and Gmail nodes that must be configured manually. Who’s it for This workflow is built for professionals, teams, and automation enthusiasts who want to manage their Google Calendar and Gmail directly from WhatsApp, powered by an AI assistant using OpenAI GPT or Google Gemini. It enables users to chat naturally through WhatsApp to schedule meetings, send emails, and check events — all without opening Gmail or Google Calendar. How it works The WhatsApp Trigger node captures incoming messages from users. The AI Agent (powered by Gemini or GPT) interprets user queries and determines the best tool to use. The Simple Memory node keeps context between messages using the user’s phone number. The Google Calendar nodes handle: Listing, creating, and updating events. Checking your availability before scheduling. The Gmail nodes handle: Sending emails. Reading and summarizing recent messages. The Date & Time node converts natural language like “next Monday at 3 PM” into proper ISO time format. The assistant responds via Send WhatsApp Response, sending clear confirmations and replies. Features Manage Gmail and Calendar entirely via WhatsApp. AI-powered understanding of natural language commands. Integrated with Google Meet for automatic conferencing links. Short-term memory for context retention. Fully modular – swap Gemini with OpenAI GPT or any LLM. Setup Steps Configure WhatsApp Cloud API via Meta for Developers. Set up Google Calendar and Gmail OAuth2 credentials. Add your Google API keys and calendar email. Connect your OpenAI or Gemini model credentials. Activate and test the workflow with messages like: “Schedule a meeting tomorrow at 5 PM.” “Check my latest emails.” “Send an email to alex@example.com about our project.” Requirements n8n instance (self-hosted or cloud) WhatsApp Business API (Meta Developer Account) Google Workspace or Gmail account OpenAI API key or Google Gemini API key Properly configured webhooks for WhatsApp Trigger Example Prompts “What’s on my calendar this week?” “Email John to confirm our meeting.” “When am I free tomorrow afternoon?” Customization Replace Gemini with OpenAI GPT in the AI Agent node. Adjust memory length for longer or shorter conversations. Add Slack or Teams notification nodes. Modify the prompt personality or response tone. Credits Created by Praneel For detailed setup help, visit praneel.tech/contact
by V3 Code Studio
🚀 Never miss a new lead again — get instant email alerts and stay ahead of every opportunity! This workflow automatically notifies your team the moment a new lead is created in your CRM or form submission. It keeps your sales, marketing, and support teams aligned — so no lead goes unnoticed, and every customer feels heard right away. ✨ How it works ✅ Capture: A webhook receives new lead data from your CRM or online form. 🧩 Clean: The workflow filters and formats the data for clear presentation. 💌 Compose: Generates a beautiful HTML email with your branding and lead details. 📨 Send: Instantly emails your team or the lead using Gmail or SMTP. ⚡ React Fast: Your team gets notified in seconds — no manual checks needed! ⚙️ Set up steps 🔗 Add a POST webhook in your CRM or app that points to your n8n webhook URL. 🏢 Update your company info (logo, name, website) in the configuration node. 📧 Connect Gmail OAuth2 or your SMTP credentials to send branded emails instantly. Compatible with Odoo, HubSpot, Zoho CRM, Salesforce, Pipedrive, Typeform, and any system that supports outgoing webhooks.
by Habeeb Mohammed
Who's it for This workflow transforms hours of manual video editing into an automated AI-powered pipeline. Perfect for anyone looking to repurpose long-form content into viral short-form clips. Ideal users include: Content Creators** - YouTubers producing long-form videos who want to maximize reach by automatically generating TikTok, Reels, and Shorts from their content Social Media Managers** - Agencies and freelancers handling multiple clients who need to scale clip production without hiring additional editors Podcasters** - Audio and video podcast hosts wanting to create promotional clips highlighting the best moments from each episode Video Editors** - Professional editors looking to automate repetitive clipping tasks and focus on creative decisions rather than technical execution Marketing Teams** - B2B and B2C teams extracting key moments from webinars, product demos, tutorials, and educational content for social campaigns Whether you're a solo creator or managing content at scale, this workflow saves 5-10 hours per video while maintaining professional quality output. How it works This workflow combines AI analysis with professional video editing tools to automatically identify and produce viral-ready clips from any YouTube video. The process flows through three main stages: Stage 1: Download and Analysis Submit a YouTube URL through the built-in form trigger yt-dlp simultaneously downloads the video in highest quality and extracts subtitles or auto-generated transcripts The transcript is intelligently chunked into 150-segment batches for optimal AI processing Each batch is analyzed by Gemini AI using specialized prompts that evaluate viral potential based on hooks, pacing, emotional peaks, and engagement triggers AI identifies 3-5 high-quality moments per batch and assigns virality scores to each potential clip Stage 2: Clip Selection and Extraction All AI-identified clips are merged and sorted by their virality scores The top 10 candidates are automatically selected for processing FFmpeg extracts each clip segment from the original video at precise timestamps Clips are processed sequentially to prevent system overload Stage 3: Professional Editing Pipeline Each clip enters a multi-stage editing subworkflow with automated operations: Smart 9:16 cropping that intelligently frames the subject for vertical platforms Precise trimming to remove dead air and optimize pacing Dynamic subtitle generation with sizing calculated based on video resolution Professional subtitle styling including bold text, high-contrast colors, strategic positioning, and text wrapping Subtitles are burned directly into the video as permanent overlays Final Delivery: The workflow processes clips with configurable wait times to match your system's capabilities. When all clips complete processing, you receive an email notification and find your social-ready clips in the /data/clips/ directory, ready for upload to any platform. Requirements ⚠️ Self-hosted n8n only - This workflow requires command-line access and cannot run on n8n Cloud due to its dependency on system-level tools. System dependencies you must install: FFmpeg** - Industry-standard video processing tool for trimming, cropping, and subtitle burning. Install on your n8n host system following this comprehensive guide. Most Linux systems can install via package manager: apt-get install ffmpeg or yum install ffmpeg. yt-dlp** - Advanced YouTube downloader that handles video and subtitle extraction. Follow official installation instructions. Recommended: pip install yt-dlp or direct binary download. FFprobe** - Usually included with FFmpeg, used for detecting video dimensions for dynamic subtitle sizing. Credentials needed: Google Gemini API account** - Powers the AI analysis for clip identification and editing instructions. Get your free API key with generous free tier limits. Gmail OAuth2 credentials** - Enables email notifications when clips are ready. Set up through n8n's credential system. Storage requirements: Ensure /data/clips/ directory exists with write permissions Plan for 2-3x the original video size in temporary storage during processing Final clips typically use 10-30% of original video size How to set up Step 1: Install system dependencies SSH into your n8n host and install required tools. For Ubuntu/Debian systems, run: apt-get update apt-get install ffmpeg pip install yt-dlp Verify installations by running ffmpeg -version and yt-dlp --version. Step 2: Configure directory structure Create the clips output directory with proper permissions: mkdir -p /data/clips chmod 755 /data/clips Step 3: Import the workflow Download the workflow JSON and import it into your n8n instance. You'll see several sticky notes color-coded by stage: yellow for description, blue for download/analysis, pink for editing operations, and green for clipping. Step 4: Set up credentials Navigate to the "viral clips identification" node and add your Google Gemini API credentials. The workflow uses the gemini-2.5-flash model for optimal speed and quality balance. Then configure Gmail OAuth2 in the "Send a message" node following n8n's authentication wizard. Step 5: Update email notification Open the "Send a message" node and replace habeebmohammedfaiz@gmail.com with your email address. Step 6: Create the editing subworkflow The workflow references a separate subworkflow for the editing pipeline. Create a new workflow in n8n, copy all nodes from the "EDITING" section (between the Execute Workflow Trigger and the final output), and save it. Note the workflow ID from the URL. Step 7: Link the subworkflow In the main workflow, open the "Call subworkflow" node and update the workflow ID to match your newly created editing workflow. Step 8: Test with a short video Start with a 5-10 minute YouTube video for your first test. Use the manual trigger or form submission. Monitor the execution to ensure all nodes complete successfully and clips appear in /data/clips/. Step 9: Adjust performance settings Based on your system's performance during the test, modify the Wait node durations. Systems with 8GB+ RAM and modern CPUs can reduce wait times to 30 seconds. Limited systems should keep 60-second waits or increase them. How to customize the workflow Adjust clip quantity and quality thresholds Open the "filter out top clips according to score" node. The code currently uses .slice(0, 10) to select the top 10 clips. Change this number based on your needs: use .slice(0, 5) for only the best clips, or .slice(0, 20) for more options. You can also add score filtering by adding results.filter(c => c.score > 0.7) before the slice operation to only include clips with virality scores above 70%. Customize subtitle appearance Navigate to the "calculate relative subtitle size" node. The JavaScript code defines several styling variables you can modify: fontSize - Currently calculated dynamically, but you can hardcode it: const fontSize = 48; fontName - Change from Arial to any system font: const fontName = 'Impact'; primaryColor - Modify text color using BGR hex format: '&H00FF00&' for green, '&HFF0000&' for red borderColor - Adjust outline color for better contrast outlineWidth - Increase from 1 to 2 or 3 for thicker borders marginV - Control vertical position (higher values move text up from bottom) Modify AI analysis prompts In the "viral clips identification" node, edit the Gemini prompt to target specific content types. For educational content, add "Focus on key teaching moments and actionable tips." For entertainment, emphasize "Identify funny moments, reactions, and unexpected events." For podcast clips, specify "Extract controversial opinions, storytelling segments, and quotable statements." Change aspect ratios The workflow defaults to 9:16 for vertical video. To create horizontal clips for YouTube or other platforms, open the "Analyze the actual whole video" node and change the aspect ratio in the JSON schema from "aspect_ratio": "9:16" to "aspect_ratio": "16:9". The AI will automatically adjust cropping coordinates accordingly. Enable audio normalization By default, audio normalization is disabled for faster processing. To enable it, open the "extract all actionable operations" node, find the audio_normalize task object, and change enabled: false to enabled: true. This ensures consistent volume levels across all clips but adds processing time. Add custom editing operations The editing pipeline is modular. You can add new operations like: Color grading by inserting FFmpeg color filters Logo overlays by adding watermark commands Intro/outro sequences by concatenating video files Background music by mixing audio tracks Add these as new task objects in the "extract all actionable operations" node following the existing pattern. Customize notification content Open the "Send a message" node to modify the email subject, body text, or add clip details. You can include clip metadata like timestamps, scores, and descriptions using expressions like {{ $json.hook }} or {{ $json.score }}. Integrate with cloud storage Add nodes after clip generation to automatically upload finished clips to Google Drive, Dropbox, AWS S3, or any n8n-supported storage service. Use the Loop Over Items1 output to access completed clip file paths. Schedule automated processing Replace the Form Trigger with a Schedule Trigger to automatically process videos from a spreadsheet or RSS feed. Combine with Google Sheets integration to maintain a queue of videos to process overnight.
by Jainik Sheth
What is this? This workflow is a Lead Qualification Voice Agent automation template built in n8n. It automates the process of qualifying leads by initiating outbound calls, analyzing call transcripts, updating lead data, and sending follow-up emails. The workflow integrates Google Sheets, OpenAI, and Twilio (via ElevenLabs API) to streamline lead management for sales or customer success teams. 1. Initiate Call Purpose: This step starts the lead qualification process by detecting new leads and preparing to contact them. How it works: The workflow is triggered when a new lead is added to your data source (such as a new row in Google Sheets, using the “Google Sheets Trigger” node). The trigger node captures essential lead information (e.g., name, phone number, service of interest) from the new entry. The workflow then prepares a personalized call script or opener, often using an AI node (like OpenAI) to generate a friendly and relevant introduction based on the lead’s details. An outbound call is initiated to the lead’s phone number using a telephony integration node (such as HTTP Request to ElevenLabs or Twilio). Note: Ensure your trigger node is correctly set up to monitor the right data source and columns. Configure the telephony node with valid API credentials and the correct phone number field. Optionally, customize the AI prompt to tailor the call opener to your business context. 2. Fetch Client Data Purpose: This step retrieves detailed information about the lead from your data source (such as a CRM, database, or Google Sheets). How it works: The workflow uses a data retrieval node (such as Google Sheets, HTTP Request, or a database node) to fetch client details based on a unique identifier (like phone number or email) captured in the trigger step. The node is configured to query the data source for all relevant fields, such as the lead’s name, company, previous interactions, and any custom attributes important for your qualification process. The fetched data is then made available to subsequent nodes, allowing for dynamic personalization of the call script and follow-up actions. Note: Ensure your data source is up-to-date and contains all necessary fields for lead qualification. 3. Outbound Call Processing Purpose: To manage the workflow after the outbound call is initiated, including handling call results, analyzing transcripts, and updating lead status. How it works: After the call is placed (using a telephony integration node), the workflow waits for the call to complete. This can be achieved using a webhook node (if your provider supports callbacks) or a polling node to check call status. Once the call is finished, the workflow retrieves the call outcome (e.g., answered, voicemail, failed) and, if available, the call transcript. The transcript can be sent to an AI node (like OpenAI) for analysis—such as extracting intent, sentiment, or key information discussed during the call. Based on the analysis, the workflow updates the lead’s status in your data source (e.g., marking as “Qualified,” “Needs Follow-up,” or “Unreachable”). Optionally, the workflow can trigger additional actions, such as sending a follow-up email, notifying a sales rep, or logging the interaction in your CRM. Note: Use webhook nodes for real-time call result processing if your telephony provider supports it. Log all call results and analysis for future reference and reporting. Prerequisities OpenAI Google Sheets Twilio Elevenlabs: Import the twilio number into Elevenlabs Set up an agent in Elevenlabs Set up prompt into Elevenlabs Set up webhooks for fetch client initiation data and post call data Customization Change Data Source: Swap the Google Sheets node for another data source (e.g., Airtable, MySQL) as needed. Modify Call Script: Edit the OpenAI prompt to generate different call openers or scripts based on your use case. Adjust Call Logic: Modify the outbound call node to use a different telephony provider or to handle call scheduling. Notification Options: Integrate with email, chat, or notification services to alert your team of qualified leads.
by Emilio Loewenstein
Turn your sales and onboarding calls into actionable insights — automatically! This workflow connects Fireflies.ai with OpenAI to analyze and grade your call transcripts. The results, along with your lead’s details, are logged directly into a Google Sheet for easy tracking. Plus, you’ll get an instant Slack or Gmail notification with the evaluation so you can take quick action. 🚀 What It Does Triggers on new Fireflies.ai transcripts** Uses AI to evaluate and grade your calls** Logs lead + scoring data into Google Sheets** Sends instant updates via Slack or Gmail** 💡 Why It’s Valuable Save hours of manual call reviews Keep a consistent, unbiased scoring system Centralize data for reporting and coaching Act faster with real-time notifications Perfect for sales, onboarding, or customer success teams who want to improve call quality at scale while saving time. 🛠️ Setup Instructions Connect Fireflies.ai – Enable transcript export from your Fireflies.ai account. Integrate with OpenAI – Use the provided API key to analyze and score transcripts automatically. Configure Google Sheets – Create a sheet with columns for: Lead Name Company Call Date Transcript Link AI Score Notes/Insights Enable Notifications – Connect Slack or Gmail to receive instant alerts with evaluation details. Test & Launch – Run a sample call to ensure transcripts flow correctly into the sheet and notifications are triggered. 🔄 Detailed Workflow A new call transcript is generated in Fireflies.ai. The transcript is sent to OpenAI, where the call is evaluated and scored based on quality, engagement, and outcomes. The results + lead data are logged automatically into Google Sheets for centralized tracking. A Slack or Gmail notification instantly alerts your team with the score and key insights, so you can take immediate action. 📊 Google Sheets Your Google Sheet should include the following columns: Lead Name** Email/Contact** Company Name** Call Date & Time** Transcript URL** AI Evaluation Score** Summary/Next Steps** This structure ensures clarity, easy reporting, and consistent data across all calls. ⚠️ Community Node Disclaimer This workflow is created with community nodes and integrations. Please review security and API key management best practices before deploying in production. 🖼️ Workflow Template
by Oneclick AI Squad
Transform your attendance management with this enterprise-grade automated workflow featuring AI-powered analytics, multi-dimensional insights, and intelligent alerting. Running hourly, it integrates multiple data sources (attendance logs + employee master data), performs sophisticated statistical analysis, detects anomalies, generates department-specific insights, and delivers beautiful HTML reports via email and Slack. Get real-time visibility into attendance patterns, punctuality trends, and actionable alerts for HR, management, and department heads. 📊🎯✨ Good to Know Intelligent Scheduling**: Runs hourly but only sends management alerts when critical thresholds are breached Multi-Source Integration**: Combines attendance logs with employee master data for enriched context Smart Analytics**: Calculates attendance rate, punctuality rate, absenteeism rate, and department-wise metrics Conditional Routing**: Management emails are triggered only for high-priority alerts (no email fatigue!) Production-Ready**: Includes error handling, data validation, and comprehensive logging Scalable Design**: Handles multiple departments, shifts, and employee types efficiently How It Works Automated Trigger – Runs hourly to monitor attendance with zero manual effort. Dual Data Ingestion – Fetches attendance and employee master data, then merges them for enriched analytics. Advanced Analytics Engine – Analyzes attendance, calculates key metrics, detects anomalies, and generates alerts. Smart Conditional Routing – Validates data, prioritizes alerts, and routes notifications via email, Slack, and database. Rich Email Reports – Sends visually formatted reports with metrics, alerts, and detailed employee breakdowns. Slack Block Kit Integration – Delivers structured, real-time attendance alerts with visual indicators to team channels. Data Persistence & Logging – Logs daily summaries, maintains audit trails, and prepares data for trend analysis dashboards. How to Use Basic Setup Import the Workflow: Copy JSON → n8n → Import from File Configure Credentials: Add Google Sheets, SMTP, and Slack credentials Update Spreadsheet IDs: Replace all placeholder sheet IDs with your actual Google Sheet IDs Set Email Addresses: Update sender and recipient email addresses Configure Slack Channel: Replace channel ID with your team's attendance channel Test: Execute workflow manually to verify connections Activate: Turn on the Schedule Trigger for hourly execution Advanced Configuration Adjust Alert Thresholds**: Modify the 10% (late) and 15% (absent) thresholds in Analytics Engine Customize Email Design**: Edit HTML in Format Email node for brand alignment Add More Channels**: Extend with Microsoft Teams, Discord, or SMS notifications Enhance Analytics**: Add shift-specific analysis, contractor vs. full-time breakdowns Integrate BI Tools**: Enable Power BI push dataset for real-time dashboards Requirements Google Sheets Setup Sheet 1: AttendanceLogs | Column | Type | Description | Example | |--------|------|-------------|---------| | EmployeeID | Text | Unique employee identifier | EMP001 | | EmployeeName | Text | Full name | John Doe | | Date | Date | Attendance date (YYYY-MM-DD) | 2025-01-15 | | Status | Text | Present/Absent/Late/Leave/WFH/Half-Day | Present | | CheckInTime | Time | Arrival time (HH:MM) | 09:15 | | CheckOutTime | Time | Departure time (HH:MM) | 18:00 | | Notes | Text | Optional comments | Doctor appointment | Sheet 2: Employees (Master Data) | Column | Type | Description | Example | |--------|------|-------------|---------| | EmployeeID | Text | Unique identifier (matches AttendanceLogs) | EMP001 | | EmployeeName | Text | Full name | John Doe | | Department | Text | Department name | Engineering | | Manager | Text | Reporting manager name | Jane Smith | | Shift | Text | Day/Night/Evening | Day | | Email | Email | Work email address | john.doe@company.com | | ContractType | Text | Full-Time/Part-Time/Contract | Full-Time | Sheet 3: DailySummary (Auto-populated by workflow) | Column | Type | Description | |--------|------|-------------| | Date | Date | Report date | | Hour | Number | Hour of day (0-23) | | Present | Number | Present count | | Absent | Number | Absent count | | Late | Number | Late count | | AttendanceRate | Number | Attendance percentage | Credentials Needed Google Sheets OAuth2 API Enable Google Sheets API in Google Cloud Console Create service account credentials Share all 3 sheets with service account email SMTP Email Account Gmail: Enable "App Passwords" in Google Account settings Or use company SMTP server details Requires: Host, Port (587), Username, Password Slack Bot Token Create Slack App at api.slack.com/apps Enable permissions: chat:write, chat:write.public Install app to workspace Copy Bot User OAuth Token (starts with xoxb-) Invite bot to target channel Placeholder Values to Replace | Placeholder | Where to Find | Example | |-------------|---------------|---------| | YOUR_ATTENDANCE_SPREADSHEET_ID | Google Sheets URL | 1BxiMVs0XRA5nFMdKvBdBZjgmUUqptlbs74OgvE2upms | | YOUR_EMPLOYEE_SPREADSHEET_ID | Google Sheets URL | 1BxiMVs0XRA5nFMdKvBdBZjgmUUqptlbs74OgvE2upms | | YOUR_SUMMARY_SPREADSHEET_ID | Google Sheets URL | 1BxiMVs0XRA5nFMdKvBdBZjgmUUqptlbs74OgvE2upms | | C12345678 | Slack → Right-click channel → Copy link → Extract ID | C05XXXXXXXX | | hr@company.com | Your HR email | hr@yourcompany.com | | management@company.com | Management distribution list | leadership@yourcompany.com | Customizing This Workflow Modify Alert Thresholds In the Analytics Engine node, find these lines: const lateThreshold = metrics.totalEmployees * 0.1; // Change 0.1 to 0.15 for 15% const absentThreshold = metrics.totalEmployees * 0.15; // Change to 0.20 for 20% Add New Status Types In the Analytics Engine node, add to the switch statement: case 'Remote': statusCount.remote++; break; Customize Email Template In the Format Email node, modify the emailHtml variable: Change gradient colors in the header div Add company logo: `` Adjust grid columns: grid-template-columns: repeat(5, 1fr); Add SMS Notifications Add Twilio or Vonage node after "Critical Alerts" Send to on-call manager for high-severity alerts Use message: 🚨 ${data.absent} employees absent today. Review required. Multi-Location Support Modify Analytics Engine to group by Location field: const locationMetrics = {}; todayRecords.forEach(record => { const location = employee.location; if (!locationMetrics[location]) { locationMetrics[location] = { present: 0, absent: 0, late: 0 }; } // ... aggregate by location }); Shift-Specific Analysis Add logic to compare check-in times against shift schedules: const shiftTimes = { 'Day': '09:00', 'Night': '21:00', 'Evening': '14:00' }; const expectedTime = shiftTimes[employee.shift]; Integration with HR Systems Add HTTP Request node to push data to: Workday**: POST attendance records via REST API BambooHR**: Update employee time-off balances ADP**: Sync payroll-related attendance data SAP SuccessFactors**: Update time management module Advanced Analytics Add Python or R nodes for: Predictive Analytics**: ML models to predict absence patterns Sentiment Analysis**: Correlate attendance with employee surveys Seasonality Detection**: Identify patterns (Monday blues, pre-holiday trends) Correlation Analysis**: Link attendance with project deadlines, weather, events Dashboard Integration Export data to visualization platforms: Power BI**: Use HTTP Request node to push to streaming dataset Tableau**: Write to database (PostgreSQL, MySQL) via Execute Query node Grafana**: Store in InfluxDB time-series database Custom Dashboard**: Build React app that reads from DailySummary sheet Troubleshooting Common Issues "Could not find node" Ensure all data sources are fetched before Analytics Engine runs Check node names match exactly in $('Node Name').all() syntax "No records found" Verify Date column format is YYYY-MM-DD in Google Sheets Check timezone settings in Schedule Trigger vs. Sheet locale Confirm EmployeeID values match between AttendanceLogs and Employees sheets "Email not sending" Verify SMTP credentials are correct Check if "Critical Alerts?" condition is being met Test email node independently with static data "Slack message failed" Ensure bot is invited to the target channel Verify channel ID format (starts with C, not # symbol) Check bot has chat:write permission scope Performance Optimization Large Datasets (>1000 employees)**: Add pagination in Google Sheets fetch Multiple Departments**: Run parallel branches for each department Historical Analysis**: Schedule separate weekly/monthly summary workflows Reduce API Calls**: Cache employee master data in workflow static data Security Best Practices ✅ Use service account credentials (not personal OAuth) ✅ Restrict sheet access to service account only ✅ Store credentials in n8n credential store (encrypted) ✅ Enable 2FA on all integrated accounts ✅ Audit logs regularly for unauthorized access ✅ Use HTTPS for all external API calls Discover more workflows – Get in touch with us
by Jitesh Dugar
Transform new hire onboarding from 3-4 hours of manual document compilation to 3 minutes of automated generation - creates personalized, role-specific document packages including welcome letters, benefits guides, IT setup instructions, and required forms, all branded and delivered with complete tracking. What This Workflow Does Revolutionizes employee onboarding with intelligent document generation, role-based customization, and automated delivery workflows: Webhook-Triggered Generation** - Automatically creates complete onboarding packages when new hires accept offers or from HR system triggers Smart Data Validation** - Verifies employee information, validates email addresses, generates employee IDs, and enriches data with company defaults Role-Based Customization** - Automatically detects job requirements and customizes documents for technical roles, management positions, or sales functions Department-Specific Details** - Populates office floor, dress code, parking assignments, and team information based on department Welcome Letter Generation** - Creates personalized welcome letters with start date details, first-day instructions, manager information, and what to bring Comprehensive Benefits Guide** - Generates detailed enrollment guides covering health insurance, dental, vision, 401(k), PTO, disability, and life insurance options IT Setup Instructions** - Produces role-specific IT guides with equipment lists, software access, network configuration, and security requirements Required Forms Package** - Creates emergency contact forms and direct deposit authorization with signature fields and document checklists Parallel Document Generation** - Simultaneously generates multiple documents for faster processing and efficiency Batch PDF Conversion** - Converts all HTML documents to professional, print-ready PDFs in one workflow execution Organized Drive Storage** - Creates employee-specific folders and archives all documents with systematic naming conventions Document Aggregation** - Collects all generated PDFs and prepares them as email attachments for delivery Automated Email Delivery** - Sends complete onboarding package to new hire with action items and first-day instructions HR System Integration** - Logs document generation, tracks completion status, manages signature requirements, and maintains audit trails Slack Team Notifications** - Alerts HR team when onboarding packages are successfully delivered with employee details Benefits Eligibility Logic** - Automatically determines benefits eligibility based on employment type and calculates enrollment start dates Signature Tracking** - Monitors which forms require signatures and tracks completion deadlines Key Features Intelligent Role Detection**: Automatically identifies technical roles requiring IT equipment, management positions needing leadership training, and sales roles requiring CRM access Equipment Allocation Logic**: Generates different equipment packages based on role (MacBook Pro for developers vs standard laptop for other roles) Employment Type Handling**: Differentiates between full-time, part-time, and contractor status affecting benefits eligibility and documentation Manager Information Auto-Population**: Pulls reporting structure, manager contact details, and department leadership information Benefits Start Date Calculation**: Automatically computes benefits eligibility dates (typically 30 days after start date) with formatted display Office Location Mapping**: Maps departments to specific floors, dress codes, and parking assignments for seamless first-day experience Dynamic Form Generation**: Creates fillable forms with proper spacing, signature lines, and checkbox fields for manual completion Multi-Document Packaging**: Generates 4+ separate documents covering welcome, benefits, IT setup, and compliance requirements Professional HTML Templates**: Beautifully designed documents with company branding, color-coded sections, and modern layouts Document Versioning**: Includes employee ID, generation timestamp, and unique document pack IDs for version control Email Action Items**: Summarizes required actions with deadlines, what to bring on first day, and pre-start preparation checklist Emergency Contact Management**: Collects primary and secondary emergency contacts with full contact information requirements Direct Deposit Authorization**: Provides bank account forms supporting primary and secondary accounts with percentage or fixed amount splits IT Security Compliance**: Documents mandatory security requirements including MFA setup, VPN configuration, and password policies Benefits Options Breakdown**: Details multiple plan options (PPO, HMO, HDHP) with premium costs and coverage comparisons Folder Organization System**: Creates hierarchical folder structure organizing documents by employee ID and full name Perfect For HR Departments** - Streamline new hire paperwork and reduce manual document preparation time Growing Companies** - Scale onboarding processes without proportionally increasing HR headcount Remote-First Organizations** - Deliver complete onboarding packages to distributed employees electronically Compliance-Focused Industries** - Maintain audit trails and ensure all required documentation is generated and tracked Companies with Complex Benefits** - Clearly communicate multiple benefit options with enrollment guidance IT-Heavy Organizations** - Provide detailed technical setup instructions for equipment and system access Multi-Department Enterprises** - Customize onboarding based on department, role, and location requirements Regulated Industries** - Ensure consistent documentation and signature tracking for compliance requirements What You Will Need Required Integrations HTML to PDF API** - PDF conversion service for professional document generation (approximately 1-5 cents per document) Gmail or SMTP** - Email delivery service for sending onboarding packages to new hires Google Drive** - Cloud storage for document archival and HR record-keeping Optional Integrations Slack Webhook** - HR team notifications when onboarding packages are delivered HR Management System** - HRIS integration for automatic logging and status tracking (BambooHR, Workday, ADP) DocuSign/HelloSign** - E-signature integration for digital form completion and signature collection Benefits Administration** - Connect to benefits platforms for enrollment link generation Applicant Tracking System** - Trigger workflow when candidates accept offers (Greenhouse, Lever, Jobvite) Employee Directory** - Sync employee data with directory systems (Okta, Azure AD) Quick Start Import Template - Copy JSON workflow and import into your n8n instance Configure PDF Service - Add HTML to PDF API credentials in the Convert to PDF node Setup Gmail - Connect Gmail OAuth2 credentials and update sender email address Connect Google Drive - Add Google Drive OAuth2 credentials and configure base folder path Customize Company Info - Edit validation node with your company name, address, website, and contact details Update HR Contact - Modify HR department name, email, and phone number in enrichment logic Configure Department Info - Adjust department mappings for office floors, dress codes, and parking assignments Customize Benefits - Edit benefits guide with your actual plan options, premiums, and coverage details Set IT Equipment - Modify equipment lists based on actual hardware provided to different roles Update Role Detection - Adjust role-based logic to match your job titles and equipment requirements Configure Email Template - Customize welcome email message with company-specific instructions Add Slack Webhook - Configure Slack notification URL for HR team alerts Test Complete Workflow - Submit sample employee data to verify all documents generate correctly Setup HR System Integration - Replace logging code with actual API calls to your HR platform Customization Options Additional Documents** - Add company policy handbooks, confidentiality agreements, or handbook acknowledgments Multi-Language Support** - Generate documents in multiple languages for international employees Custom Branding** - Add company logos, color schemes, and custom styling to all document templates Conditional Sections** - Show/hide document sections based on employment type, location, or department Variable Pay Structures** - Include salary information, bonus structure, or commission plans in welcome letters Onboarding Schedules** - Generate detailed first-week schedules with meetings, training, and orientation sessions Team Introductions** - Include team member photos, bios, and contact information in welcome packages Location-Specific Content** - Different documents for headquarters vs remote vs international employees Probationary Period Info** - Add probation terms, review schedules, and performance expectations Company Culture Content** - Include mission, values, culture guide, and employee testimonials Video Embeddings** - Add QR codes or links to welcome videos from CEO or department heads Interactive Checklists** - Generate pre-boarding checklists with tasks to complete before start date Equipment Order Forms** - Include forms for employees to select laptop preferences or accessories Background Check Status** - Conditional content based on background check completion Referral Program Info** - Include employee referral program details and bonus structure Expected Results 95% time savings** - Reduce document preparation from 3-4 hours to 3 minutes per employee 100% consistency** - Eliminate errors from manual document creation and ensure brand compliance Same-day delivery** - New hires receive complete packages within minutes of offer acceptance Zero document loss** - Systematic archival prevents missing paperwork or compliance gaps Improved new hire experience** - Professional, organized packages create positive first impression Faster time-to-productivity** - Clear instructions and preparation reduce first-day confusion Reduced HR workload** - Automation frees HR team for strategic onboarding activities Better compliance** - Consistent documentation and tracking meets regulatory requirements Scalable onboarding** - Handle 10x more new hires without additional HR staff Complete audit trail** - Timestamp and track every document generation for compliance reviews Pro Tips Test with Multiple Roles** - Verify role detection logic works correctly for all job titles in your organization Validate Email Delivery** - Ensure onboarding emails don't trigger spam filters with test sends Set Realistic Deadlines** - Give new hires adequate time to review and complete forms before start date Include Document Checklist** - Help new hires track which forms require signatures or return Provide IT Support Contact** - Make sure IT help desk info is accurate and responsive Update Benefits Annually** - Review and refresh benefits content during open enrollment periods Personalize Welcome Messages** - Include hiring manager or team-specific welcome notes when possible Archive Systematically** - Maintain consistent folder structure for easy retrieval and compliance audits Track Form Completion** - Follow up with new hires who haven't returned required documents Gather Feedback** - Survey new hires about onboarding package clarity and usefulness Keep Templates Current** - Regularly review and update document templates with latest policies Add Video Walkthroughs** - Link to video tours of office, parking, and first-day procedures Include FAQ Document** - Answer common new hire questions proactively in package Customize for Remote Workers** - Create alternate documents for fully remote employees Coordinate with Managers** - Notify hiring managers when their new reports receive onboarding packages Business Impact Metrics Track these key metrics to measure workflow success: Document Generation Time** - Average minutes from trigger to package delivery (target: under 5 minutes) HR Productivity Gain** - Hours saved per month on document preparation (typical: 15-20 hours monthly) New Hire Satisfaction** - Survey rating on onboarding package quality and clarity (target: 4.5/5) Form Completion Rate** - Percentage of required forms returned by start date (target: 95%+) Documentation Errors** - Reduction in incorrect or missing information on documents (target: 100% accuracy) Time-to-Productivity** - Days until new hire reaches full productivity (expect 20-30% faster) Compliance Adherence** - Percentage of complete onboarding files meeting regulatory standards (target: 100%) Package Delivery Speed** - Hours between offer acceptance and package receipt (target: same day) First-Day Preparedness** - Percentage of new hires arriving with completed forms (target: 90%+) HR Scalability** - Number of new hires HR can onboard simultaneously without quality loss Template Compatibility Compatible with n8n version 1.0 and above Works with n8n Cloud and Self-Hosted instances Requires HTML to PDF API service subscription No coding required for basic setup Fully customizable document templates and content Supports unlimited employees and departments Integrates with major HRIS platforms via API Handles all employment types and role variations Scalable to process multiple onboarding packages simultaneously Ready to transform your employee onboarding process? Import this template and start generating professional, role-specific onboarding packages that delight new hires, ensure compliance, and free your HR team to focus on creating exceptional employee experiences instead of shuffling paperwork!
by Matheus Pedrosa
Workflow Overview This workflow provides a complete, automated post-purchase solution triggered by a successful payment webhook from Abacate Pay. (For international users, think of Abacate Pay as 'the Brazilian Stripe' – a popular and robust payment gateway in Brazil). When a successful payment is registered, this workflow instantly sends a beautiful, responsive HTML confirmation email to the customer and posts a detailed notification to a designated Slack channel. The key feature is its ability to dynamically reward first-time buyers. The workflow checks the customer's purchase history, and if it's their first order, it automatically generates a 10% discount coupon for their next purchase and includes it in the welcome email. Key Features: Webhook Trigger:** Securely listens for successful payment events from Abacate Pay. First-Time Buyer Detection:** Queries past orders to identify new customers. Automatic Coupon Generation:** Creates a unique, single-use discount coupon for new buyers. Dynamic Content:** The email and Slack messages automatically include the coupon details, but only for first-time buyers. Responsive HTML Email:** Sends a professional and mobile-friendly order confirmation. Rich Slack Notifications:** Uses Block Kit for well-formatted, actionable alerts for your team. Centralized Configuration:** A single Configs node makes it easy to manage static variables like company name and tokens. Setup Instructions: Webhook Node: Copy the webhook URL and paste it into your Abacate Pay dashboard for the "payment successful" event (e.g., billing.paid). Configs Node: Open this Set node and fill in your company's information (companyName, companySite, companyEmail) and the token used to validate the webhook requests. HTTP Request Nodes (GetOrders, CreateCustomCoupon): In both nodes, go to the "Authentication" or "Header" section and replace the placeholder Bearer Token with your actual Abacate Pay API key. Send Email Node: Select your email credentials from the dropdown or configure a new account (e.g., SMTP, Gmail). Slack Node: Select your Slack credentials and choose the channel where you want to receive notifications. Activate the Workflow: Save and activate the workflow. It's now ready to process incoming orders! Required Credentials: Abacate Pay API Credentials Email Credentials (e.g., SMTP, SendGrid, Gmail) Slack API Credentials
by Guillaume Duvernay
This template introduces a revolutionary approach to automated web research. Instead of a rigid workflow that can only find one type of information, this system uses a "thinker" and "doer" AI architecture. It dynamically interprets your plain-English research request, designs a custom spreadsheet (CSV) with the perfect columns for your goal, and then deploys a web-scraping AI to fill it out. It's like having an expert research assistant who not only finds the data you need but also builds the perfect container for it on the fly. Whether you're looking for sales leads, competitor data, or market trends, this workflow adapts to your request and delivers a perfectly structured, ready-to-use dataset every time. Who is this for? Sales & marketing teams:** Generate targeted lead lists, compile competitor analysis, or gather market intelligence with a simple text prompt. Researchers & analysts:** Quickly gather and structure data from the web for any topic without needing to write custom scrapers. Entrepreneurs & business owners:** Perform rapid market research to validate ideas, find suppliers, or identify opportunities. Anyone who needs structured data:** Transform unstructured, natural language requests into clean, organized spreadsheets. What problem does this solve? Eliminates rigid, single-purpose workflows:** This workflow isn't hardcoded to find just one thing. It dynamically adapts its entire research plan and data structure based on your request. Automates the entire research process:** It handles everything from understanding the goal and planning the research to executing the web search and structuring the final data. Bridges the gap between questions and data:** It translates your high-level goal (e.g., "I need sales leads") into a concrete, structured spreadsheet with all the necessary columns (Company Name, Website, Key Contacts, etc.). Optimizes for cost and efficiency:* It intelligently uses a combination of deep-dive and standard web searches from *Linkup.so** to gather high-quality initial results and then enrich them cost-effectively. How it works (The "Thinker & Doer" Method) The process is cleverly split into two main phases: The "Thinker" (AI Planner): You submit a research request via the built-in form (e.g., "Find 50 US-based fashion companies for a sales outreach campaign"). The first AI node acts as the "thinker." It analyzes your request and determines the optimal structure for your final spreadsheet. It dynamically generates a plan, which includes a discoveryQuery to find the initial list, an enrichmentQuery to get details for each item, and the JSON schemas that define the exact columns for your CSV. The "Doer" (AI Researcher): The rest of the workflow is the "doer," which executes the plan. Discovery: It uses a powerful "deep search" with Linkup.so to execute the discoveryQuery and find the initial list of items (e.g., the 50 fashion companies). Enrichment: It then loops through each item in the list. For each one, it performs a fast and cost-effective "standard search" with Linkup to execute the enrichmentQuery, filling in all the detailed columns defined by the "thinker." Final Output: The workflow consolidates all the enriched data and converts it into a final CSV file, ready for download or further processing. Setup Connect your AI provider: In the OpenAI Chat Model node, add your AI provider's credentials. Connect your Linkup account: In the two Linkup (HTTP Request) nodes, add your Linkup API key (free account at linkup.so). We recommend creating a "Generic Credential" of type "Bearer Token" for this. Linkup offers €5 of free credits monthly, which is enough for 1k standard searches or 100 deep queries. Activate the workflow: Toggle the workflow to "Active." You can now use the form to submit your first research request! Taking it further Add a custom dashboard:** Replace the form trigger and final CSV output with a more polished user experience. For example, build a simple web app where users can submit requests and download their completed research files. Make it company-aware:** Modify the "thinker" AI's prompt to include context about your company. This will allow it to generate research plans that are automatically tailored to finding leads or data relevant to your specific products and services. Add an AI summary layer:** After the CSV is generated, add a final AI node to read the entire file and produce a high-level summary, such as "Here are the top 5 leads to contact first and why," turning the raw data into an instant, actionable report.
by Olaf Titel
Setup & Instructions — fluidX: Create Session, Analyze & Notify Goal: This workflow demonstrates the full fluidX THE EYE integration — starting a live session, inviting both the customer (via SMS) and the service agent (via email), and then accessing the media (photos and videos) created during the session. Captured images are automatically analyzed with AI, uploaded to an external storage (such as Google Drive), and a media summary for the session is generated at the end. The agent receives an email with a link to join the live session. The customer receives an SMS with a link to start sharing their camera. Once both are connected, the agent can view the live feed, and the system automatically stores uploaded images and videos in Google Drive. When the session ends, the workflow collects all media and creates a complete AI-powered session summary (stored and updated in Google Drive). Below is an example screenshot from the customer’s phone: Prerequisites Developer account:* https://live.fluidx.digital (activate the *TEST plan**, €0) API docs (Swagger):** fluidX.digital API 🔐 Required Credentials 1️⃣ fluidX API key (HTTP Header Auth) • Credential name in n8n: fluidx API key • Header name: x-api-key • Header value: YOUR_API_KEY 2️⃣ SMTP account (for outbound email) • Credential name in n8n: SMTP account • Configure host, port, username, and password according to your provider • Enable TLS/SSL as required 3️⃣ Google Drive account • Used to store photos, videos, and automatically update the session summary files. 4️⃣ OpenAI API (for AI analysis & summary) •Used in the Analyze Images (AI) and Generate Summary parts of the workflow. • Credential type: OpenAI • Credential name (suggested): OpenAI account • API Key: your OpenAI API key • Model: e.g. gpt-4.1, gpt-4o, or similar (choose in the OpenAI node settings) ⚙️ Configuration (in the “Set Config” node) BASE_URL: https://live.fluidx.digital company / project / billingcode / sku: adjust as needed emailAgent: set before running (empty in template) phoneNumberUser: set before running (empty in template) Flow Overview Form Trigger → Create Session → Set Session Vars → Send SMS (User) → Send Email (Agent) → Monitor Media → Analyze Images (AI) → Upload Files to Google Drive → Generate Summary → Update Summary File The workflow starts automatically when a Form submission is received. Users enter the customer’s phone number and agent’s email, and the system creates a new fluidX THE EYE session. As media is uploaded during the session, the workflow automatically retrieves, stores, analyzes, and summarizes it — providing a complete end-to-end automation example for remote inspection, support, or field-service use cases. Notes Do not store real personal data inside the template. Manage API keys and secrets via n8n Credentials or environment variables. Log out of https://live.fluidx.digital in the agent’s browser before testing, to ensure a clean invite flow and session creation.
by JJ Tham
Generate AI Voiceovers from Scripts and Upload to Google Drive This is the final piece of the AI content factory. This workflow takes your text-based video scripts and automatically generates high-quality audio voiceovers for each one, turning your text into ready-to-use audio assets for your video ads. Go from a spreadsheet of text to a folder of audio files, completely on autopilot. ⚠️ CRITICAL REQUIREMENTS (Read First!) This is an advanced, self-hosted workflow that requires specific local setup: Self-Hosted n8n Only:** This workflow uses the Execute Command and Read/Write Files nodes, which requires you to run your own instance of n8n. It will not work on n8n Cloud. FFmpeg Installation:** You must have FFmpeg installed on the same machine where your n8n instance is running. This is used to convert the audio files to a standard format. What it does This is Part 3 of the AI marketing series. It connects to the Google Sheet where you generated your video scripts (in Part 2). For each script that hasn't been processed, it: Uses the Google Gemini Text-to-Speech (TTS) API to generate a voiceover. Saves the audio file to your local computer. Uses FFmpeg to convert the raw audio into a standard .wav file. Uploads the final .wav file to your Google Drive. Updates the original Google Sheet with a link to the audio file in Drive and marks the script as complete. How to set up IMPORTANT: This workflow is Part 3 of a series and requires the output from Part 2 ("Generate AI Video Ad Scripts"). If you need Part 1 or Part 2 of this workflow series, you can find them for free on my n8n Creator Profile. Connect to Your Scripts Sheet: In the "Getting Video Scripts" node, connect your Google Sheets account and provide the URL to the sheet containing your generated video scripts from Part 2. Configure AI Voice Generation (HTTP Request): In the "HTTP Request To Generate Voice" node, go to the Query Parameters and replace INSERT YOUR API KEY HERE with your Google Gemini API key. In the JSON Body, you can customize the voice prompt (e.g., change <INSERT YOUR DESIRED ACCENT HERE>). Set Your Local File Path: In the first "Read/Write Files from Disk" node, update the File Name field to a valid directory on your local machine where n8n has permission to write files. Replace /Users/INSERT_YOUR_LOCAL_STORAGE_HERE/. Connect Google Drive: In the "Uploading Wav File" node, connect your Google Drive account and choose the folder where your audio files will be saved. Update Your Tracking Sheet: In the final "Uploading Google Drive Link..." node, ensure it's connected to the same Google Sheet from Step 1. This node will update your sheet with the results. Name and Description for Submission Form Here are the name and description, updated with the new information, ready for you to copy and paste. Name: Generate AI Voiceovers from Scripts and Upload to Google Drive Description: Welcome to the final piece of the AI content factory! 🔊 This advanced workflow takes the video ad scripts you've generated and automatically creates high-quality audio voiceovers for each one, completing your journey from strategy to ready-to-use media assets. ⚠️ This is an advanced workflow for self-hosted n8n instances only and requires FFmpeg to be installed locally. ⚙️ How it works This workflow is Part 3 of a series. It reads your video scripts from a Google Sheet, then for each script it: Generates a voiceover using the Google Gemini TTS API. Saves the audio file to your local machine. Converts the file to a standard .wav format using FFmpeg. Uploads the final audio file to Google Drive. Updates your Google Sheet with a link to the new audio file. 👥 Who’s it for? Video Creators & Marketers: Mass-produce voiceovers for video ads, tutorials, or social media content without hiring voice actors. Automation Power Users: A powerful example of how n8n can bridge cloud APIs with local machine commands. Agencies: Drastically speed up the production of audio assets for client campaigns. 🛠️ How to set up This workflow requires specific local setup due to its advanced nature. IMPORTANT: This is Part 3 of a series. To find Part 1 ("Generate a Strategic Plan") and Part 2 ("Generate Video Scripts"), please visit my n8n Creator Profile where they are available for free. Setup involves connecting to your scripts sheet, configuring the AI voice API, setting a local file path for n8n to write to, and connecting your Google Drive.