by n8nwizard
📌 Overview This advanced multi-phase n8n workflow automates the complete research, analysis, and ideation pipeline for a YouTube strategist. It scrapes competitor channels, analyzes top-performing titles and thumbnails, identifies niche trends, gathers audience sentiment from comments, and produces data-driven content ideas—automatically writing them into a structured Google Sheets dashboard. This system is ideal for: YouTube creators Agencies Content strategists Automation engineers Anyone who wants to generate YouTube content ideas backed by real data 🧠 High‑Level Architecture The workflow is split into 5 phases, each building on the previous: Phase 1 — Niche Outliers (Input: User Form) Scrapes 3 high‑quality channels from your niche, extracts their outlier videos, and analyzes why they work. Phase 2 — Broad Niche Insights (Weekly) Scrapes the top trending content in your broad niche (e.g., "AI", "fitness", "personal finance") and logs weekly insights. Phase 3 — Niche Insights (Daily) Scrapes the top videos in your specific micro‑niche daily to keep track of content momentum. Phase 4 — Comment Analysis Analyzes real comments from your channel to understand what your audience likes, dislikes, and wants more of. Phase 5 — Content Ideation Generates 3 highly‑optimized title + thumbnail concepts using all prior insights. Everything is automatically logged into a Google Sheets dashboard. 🧩 Phase-by-Phase Breakdown ⭐ Phase 1 — Niche Outliers (Form Trigger) User enters 3 YouTube channel URLs in a form. Workflow scrapes each channel using Apify YouTube Scraper. Filters for top-performing videos. Extracts: title, views, likes, thumbnail, URL. AI analyzes: Power words in titles (OpenRouter/GPT 4.1-mini) Thumbnail attention hooks (OpenAI Vision) All insights are appended into the “Niche Outliers” sheet. Purpose: Understand what the best creators in your niche are doing. 🌐 Phase 2 — Broad Niche Insights (Weekly — Sundays @ 5 AM) Workflow scrapes the top videos for a broad niche (e.g., “artificial intelligence”). Analyzes: Title structure Power words Thumbnail cues Writes weekly insights to “Broad Niche Weekly” sheet. Purpose: Stay informed about macro‑level trends. 🎯 Phase 3 — Niche Insights (Daily @ 6 AM) Scrapes the top videos in your specific micro‑niche (e.g., “n8n automations”). Runs title + thumbnail analysis. Appends daily results to “Niche Daily”. Results feed directly into Phase 5. Purpose: Track daily momentum and trending formats. 💬 Phase 4 — Comment Analysis (Channel Feedback) Scrapes your channel’s latest 5 videos. Extracts up to 30 comments from each. Aggregates comments. AI identifies: What viewers love What viewers dislike What viewers are asking for Stores patterns in “Comment Analysis” sheet. Purpose: Understand real audience demand. 💡 Phase 5 — Content Ideation Using AI Using insights from all previous phases: Top titles Power words Thumbnail patterns Daily niche trends Audience comment analysis Channel positioning The Creative Agent produces: 3 optimized video titles 3 matching thumbnail concepts These are appended to the “Ideation” sheet. A Slack notification is sent when ideation is ready. Purpose: Fully automated content idea generation. 🗂️ Outputs in Google Sheets The workflow populates these tabs: 📌 Niche Outliers (top competitor videos) 📌 Broad Niche Weekly (weekly trend analysis) 📌 Niche Daily (daily trend analysis) 📌 Comment Analysis (audience sentiment) 📌 Ideation (final titles + thumbnails) 🔧 What This Workflow Automates ✔ Competitor analysis ✔ Thumbnail + title breakdowns ✔ Daily niche tracking ✔ Weekly niche tracking ✔ Viewer sentiment analysis ✔ Fully AI‑generated content ideas ✔ Automatic data logging to Google Sheets ✔ Slack notifications This is essentially a 24/7 AI YouTube strategist. ⚙️ Setup Requirements Apify API Key** (used in 5 scraper nodes) OpenRouter API Key** (for GPT 4.1-mini intelligence) OpenAI API Key** (for thumbnail image analysis) Google Sheets OAuth2 Credential** Make a copy of the provided sheet template Fill Set nodes: Phase II: Broad niche (e.g., "AI") Phase III: Micro niche (e.g., "n8n automations") Phase IV: Your Channel URL Phase V: Your Channel Description 🧪 Testing Guide Test the Form Trigger with 3 competitor channel URLs. Test both Schedule Triggers (weekly + daily) manually. Verify Sheets are receiving rows. Run the full pipeline end‑to‑end. Confirm Slack notification. Everything should chain together smoothly. 🎉 Final Result By the end of this workflow, you have a: 🧠 Data‑driven YouTube strategy system that: studies your niche finds outliers understands your audience detects trends generates smart content ideas every day
by Emilio Loewenstein
This workflow automates customer email support by combining Gmail, AI classification, and a knowledge base to provide instant, accurate, and friendly responses. It’s designed for businesses that want to improve customer satisfaction while reducing manual workload. 🚀 How it Works Gmail Trigger The workflow listens for new incoming Gmail messages. Text Classification Each email is classified using AI as either Customer Support or Other. If it’s Other, the workflow stops. If it’s Customer Support, the email continues to the AI agent. AI Agent with Knowledge Base The AI agent: Reads the customer’s message. Searches the Pinecone Knowledge Base for FAQs and policies. Generates a helpful, polite, and detailed reply using an OpenRouter model. Signs off as Mrs. Helpful from Tech Haven Solutions. Reply to Gmail The drafted email is automatically sent back to the customer. 💡 Value ✅ Save Time – No more manual triaging and drafting of replies. ✅ Consistency – Every answer is based on your own FAQ/policies. ✅ Customer Satisfaction – Fast, friendly, and accurate responses 24/7. ✅ Scalable – Handle higher email volume without scaling headcount. 🔑 Credentials Needed To use this workflow, connect the following accounts: Gmail OAuth2** → for receiving and sending emails. OpenRouter API** → for text classification and AI-generated replies. OpenAI API** → for embeddings (to connect FAQs with AI). Pinecone API** → for storing and retrieving knowledge base content. 🛠 Example Use Case A customer writes: > “Hi, I placed an order last week but haven’t received a shipping confirmation yet. Can you check the status?” The workflow will: Detect it’s a support-related email. Retrieve order policy information from the knowledge base. Generate a friendly response explaining order tracking steps. Automatically reply to the customer in Gmail. ⚡️ Setup Instructions Import this workflow into your n8n instance. Connect your Gmail, OpenRouter, OpenAI, and Pinecone credentials. Populate your Pinecone knowledge base with FAQs and policies. Activate the workflow. From now on, all support-related emails will be automatically answered by your AI-powered support agent.
by Hardikkumar
This workflow is the AI analysis and alerting engine for a complete social media monitoring system. It's designed to work with data scraped from X (formerly Twitter) using a tool like the Apify Tweet Scraper, which logs the data into a Google Sheet. The workflow then automatically analyzes new tweets with Google Gemini and sends tailored alerts to Slack. How it works This workflow automates the analysis and reporting part of your social media monitoring: tweet Hunting:** It finds tweets for the query entered in the set node and passes the data to the google sheets Fetches New Tweets:** It gets all new rows from your Google Sheet that haven't been processed yet (it looks for "Notmarked" in the 'action taken' column). Prepares for AI:** It combines the data from all new tweets into a single, clean prompt for the AI to analyze. AI Analysis with Gemini:* It sends the compiled data to Google Gemini, asking for a full summary report *and a separate, machine-readable JSON list of any urgent items. Splits the Response:** The workflow intelligently separates the AI's text summary from the JSON data for urgent alerts. Sends Notifications:** The high-level summary is sent to a general Slack channel (e.g., #brand-alerts). Each urgent item is sent as a separate, detailed alert to a high-priority Slack channel (e.g., #urgent). Set up steps It should take about 5-10 minutes to get this workflow running. Prerequisite - Data Source: Ensure you have a Google Sheet being populated with tweet data. For a complete automation, you can set up a new google sheet with the same structure for saving the tweets data and run the Tweet Scraper on a schedule. Configure Credentials: Make sure you have credentials set up in your n8n instance for Google Sheets, Google Gemini (PaLM) API, and Slack. Google Sheets Node ("Get row(s) in sheet"): Select your Google Sheet containing the tweet data. Choose the specific sheet name from the dropdown. Ensure your sheet has a column named action taken so the filter works correctly. Google Gemini Chat Model Node: Select your Google Gemini credential from the dropdown. Slack Nodes ("Send a message" & "Send a message1"): In the first Slack node, choose the channel for the summary report. In the second Slack node, choose the channel for urgent alerts. Save and Activate: Once configured, save your workflow and turn it on!
by Oneclick AI Squad
This automated workflow monitors your website's keyword rankings daily and sends instant alerts to your team when significant ranking drops occur. It fetches current ranking positions, compares them with historical data, and triggers notifications through Slack and email when keywords drop beyond your defined threshold. Good to know The workflow uses SERP API for accurate ranking data; API costs apply based on your usage volume Ranking checks are performed daily to avoid overwhelming search engines with requests The system tracks ranking changes over time and maintains historical data for trend analysis Slack integration requires workspace permissions and proper bot configuration False positives may occur due to personalized search results or data center variations How it works Daily SEO Check Trigger** initiates the workflow on a scheduled basis Get Keywords Database** retrieves your keyword list and current ranking data Filter Active Keywords Only** processes only keywords marked as active for monitoring Fetch Google Rankings via SERP API** gets current ranking positions for each keyword Wait For Response** Wait for gets current ranking positions Parse Rankings & Detect Changes** compares new rankings with historical data and identifies significant drops Filter Significant Ranking Drops** isolates keywords that dropped beyond your threshold (e.g., 5+ positions) Send Slack Ranking Alert** notifies your team channel about ranking drops Send Email Ranking Alert** sends detailed email reports to stakeholders Update Rankings in Google Sheet** saves new ranking data for historical tracking Generate SEO Monitoring Summary** creates a comprehensive report of all ranking changes How to use Import the workflow into n8n and configure your SERP API credentials Set up your Google Sheet with the required keyword database structure Configure Slack webhook URL and email SMTP settings Define your ranking drop threshold (recommended: 5+ position drops) Test the workflow with a small keyword set before full deployment Schedule the workflow to run daily during off-peak hours Requirements SERP API account** with sufficient credits for daily keyword checks Google Sheets access** for keyword database and ranking storage Slack workspace** with webhook permissions for team notifications Email service** (SMTP or API) for stakeholder alerts Keywords database** properly formatted in Google Sheets Database/Sheet Columns Required Google Sheet: "Keywords Database" Create a Google Sheet with the following columns: | Column Name | Description | Example | |-------------|-------------|---------| | keyword | Target keyword to monitor | "best seo tools" | | domain | Your website domain | "yourwebsite.com" | | current_rank | Latest ranking position | 5 | | previous_rank | Previous day's ranking | 3 | | status | Monitoring status | "active" | | target_url | Expected ranking URL | "/best-seo-tools-guide" | | search_volume | Monthly search volume | 1200 | | difficulty | Keyword difficulty score | 65 | | date_added | When keyword was added | "2025-01-15" | | last_checked | Last monitoring date | "2025-07-30" | | drop_threshold | Custom drop alert threshold | 5 | | category | Keyword grouping | "Product Pages" | Customising this workflow Modify ranking thresholds** in the "Filter Significant Ranking Drops" node to adjust sensitivity (e.g., 3+ positions vs 10+ positions) Add competitor monitoring** by duplicating the SERP API node and tracking competitor rankings for the same keywords Customize alert messages** in Slack and email nodes to include your brand voice and specific stakeholder information Extend to multiple search engines** by adding Bing or Yahoo ranking checks alongside Google Implement ranking improvement alerts** to celebrate when keywords move up significantly Add mobile vs desktop tracking** by configuring separate SERP API calls for different device types
by Sankalp Dev
Automated Marketing Analytics Report with AI Agent How it works Transform your marketing data into actionable insights with this intelligent automation workflow. The system combines scheduled triggers with AI-powered analysis to deliver comprehensive marketing reports directly to your inbox. Key Features: Scheduled automated reporting (daily, weekly, or monthly) AI-powered data analysis using advanced language models Multi-platform marketing data integration via GoMarble MCP Intelligent report generation with actionable recommendations Direct email delivery of formatted reports Set up steps Prerequisites: GoMarble MCP account and API access Gmail account for report delivery n8n instance (cloud or self-hosted) Configuration Time: ~15-20 minutes Step-by-step setup: Connect GoMarble MCP to n8n Follow the integration guide: GoMarble n8n Setup Configure your marketing platform credentials (Google Ads, Facebook Ads, Analytics) Configure the Schedule Trigger Set your preferred reporting frequency Choose optimal timing for data availability Customize the Report Prompt Define specific metrics and KPIs to track Set analysis parameters and report format preferences Set up AI Agent Configuration Choose between Anthropic Claude or OpenAI models Configure the GoMarble MCP tools for your marketing platforms Configure Gmail Integration Set recipient email addresses Customize email template and subject line Advanced Configuration: Add conditional logic for performance thresholds Include custom data visualization requests Set up alert triggers for significant metric changes What you'll get Automated Intelligence:** Regular marketing performance analysis without manual effort Cross-Platform Insights:** Unified view of Google Ads, Facebook Ads, and Analytics data AI-Powered Recommendations:** Strategic insights and optimization suggestions Professional Reports:** Well-formatted, executive-ready marketing summaries Scalable Solution:** Easy to extend with additional marketing platforms or custom metrics Perfect for marketing teams, agencies, and business owners who want to stay on top of their marketing performance with minimal manual work.
by Guillaume Duvernay
Never worry about losing your n8n workflows again. This template provides a powerful, automated backup system that gives you the peace of mind of version control without the complexity of Git. On a schedule you define, it intelligently scans your n8n instance for new workflow versions and saves them as downloadable snapshots in a clean and organized Airtable base. But it’s more than just a backup. This workflow uses AI to automatically generate a concise summary of what each workflow does and even documents the changes between versions. The result is a fully searchable, self-documenting library of all your automations, making it the perfect "single source of truth" for your team or personal projects. Who is this for? Self-hosted n8n users:** This is an essential insurance policy to protect your critical automations from server issues or data loss. n8n developers & freelancers:** Maintain a complete version history for client projects, allowing you to easily review changes and restore previous versions. Teams using n8n:** Create a central, browseable, and documented repository of all team workflows, making collaboration and handovers seamless. Any n8n user who values their work:** Protect your time and effort with an easy-to-use, "set it and forget it" backup solution. What problem does this solve? Prevents catastrophic data loss:** Provides a simple, automated way to back up your most critical assets—your workflows. Creates "no-code" version control:** Offers the benefits of version history (like Git) but in a user-friendly Airtable interface, allowing you to browse and download any previous snapshot. Automates documentation:** Who has time to document every change? The AI summary and changelog features mean you always have up-to-date documentation, even if you forget to write it yourself. Improves workflow discovery:** Your Airtable base becomes a searchable and browseable library of all your workflows and their purposes, complete with AI-generated summaries. How it works Scheduled check: On a recurring schedule (e.g., daily), the workflow fetches a list of all workflows from your n8n instance. Detect new versions: It compares the current version ID of each workflow with the snapshot IDs already saved in your Airtable base. It only proceeds with new, unsaved versions. Generate AI documentation: For each new snapshot, the workflow performs two smart actions: AI Changelog: It compares the new workflow JSON with the previously saved version and uses AI to generate a one-sentence summary of what’s changed. AI Summary: It periodically re-analyzes the entire workflow to generate a fresh, high-level summary of its purpose, ensuring the main description stays up-to-date. Store in Airtable: It saves everything neatly in the provided two-table Airtable base: A Workflows table holds the main record and the AI summary. A linked Snapshots table stores the version-specific details, the AI changelog, and the actual .json backup file as an attachment. Setup Duplicate the Airtable base: Before you start, click here to duplicate the Airtable Base template into your own Airtable account. Configure the workflow: Connect your n8n API credentials to the n8n nodes. Connect your Airtable credentials and map the nodes to the base you just duplicated. Connect your AI provider credentials to the OpenAI Chat Model nodes. Important: In the Store workflow file into Airtable (HTTP Request) node, you must replace <AIRTABLE-BASE-ID> in the URL with your own base ID (it starts with app...). Set your schedule: Configure the Schedule Trigger to your desired frequency (daily is a good start). Activate the workflow. Your automated, AI-powered backup system is now live! Taking it further Add notifications:* Add a *Slack* or *Email** node at the end of the workflow to send a summary of which workflows were backed up during each run. Use different storage:* While designed for Airtable, you could adapt the logic to store the JSON files in *Google Drive* or *Dropbox* and the metadata in *Google Sheets* or *Notion**. Optimize AI costs:* The *Check workflow status** (Code) node is set to regenerate the main AI summary for the first few snapshots and then every 5th snapshot. You can edit the code in this node to change this frequency and manage your token consumption.
by Miftah Rahmat
Automate Water Bill Calculations with Telegram, Gemini AI, and Google Sheets This workflow automates the calculation of monthly water bills. Residents can send a photo of their water meter along with their name via Telegram. The workflow uses Gemini AI to extract the meter reading, calculates the usage difference compared to the previous month, and updates a Google Sheet with the billing details. Finally, the workflow sends a summary back via Telegram. Don’t hesitate to reach out if you have any questions or run into issues! 🙌 Requirements A Telegram bot token (created via BotFather). A Google account with access to Google Sheets. A Gemini API key (). A pre-created Google Sheet with the required columns. Google Sheet Setup Create a new Google Sheet with the following columns: Nama, Volume Sebelumnya, Volume Saat Ini, Harga/m³, Jumlah Bayar, Beban, Total Bayar, Tanggal Input Workflow Setup Instructions Connect Google Sheets Add your Google Sheets credentials in n8n. Link the workflow to your sheet with the structure above. Set Up Telegram Bot Create a Telegram bot via BotFather. Copy your bot token into the Telegram Trigger node. Configure Gemini AI Obtain a Gemini API key from Google AI Studio. Add it to your n8n credentials. The workflow will parse the meter reading from the uploaded image. Example Calculation Previous Volume: 535 m³ Current Volume: 545 m³ Usage: 10 m³ Price per m³: Rp3.000 Fixed cost: Rp3.000 Total Bill: Rp33.000 How It Works User sends a photo of the water meter with caption (name). Telegram Trigger receives the message. Gemini AI reads the meter number from the photo. Workflow fetches previous volume from Google Sheets. Usage and total bill are calculated. Data is stored back into Google Sheets. Bot replies in Telegram with detailed bill info. Customization Change Harga/m³ in the sheet to match your community’s water price. Update Beban if your community uses a different fixed fee. Edit the Telegram reply message node to adjust wording. With this workflow, you can streamline water billing for residents, ensure accuracy, and save time on manual calculations.
by Muhammad Bello
Description This n8n template demonstrates how to turn raw YouTube comments into research-backed content ideas complete with hooks and outlines. Use cases include: Quickly mining a competitor’s audience for video ideas. Generating hooks and outlines for your own channel’s comments. Validating content opportunities with live audience feedback. Good to know Apify is used to scrape YouTube comments (requires an API token). GPT-4.1-mini is used for both filtering and content generation. Tavily provides fresh research to ground the AI’s responses. All outputs are stored in Google Sheets, making it easy to manage and track ideas. How it works Trigger – Paste a YouTube URL into the chat trigger. Scrape Comments – Apify fetches all comments and metadata. Filter – GPT-4.1-mini decides if each comment could inspire a content idea. Store – Comments and “Yes/No” decisions are appended to Google Sheets. Research & Enrich – For “Yes” comments, Tavily provides context, and GPT generates a topic, hook, and outline. Update Sheet – The same row in Google Sheets is updated with enriched fields. Google Sheets Setup Your Google Sheet should include these columns (in this order): id | text | author | likes | isIdea | topic | research | hook | outline id** – unique identifier for each comment text** – the full YouTube comment author** – commenter’s name/handle likes** – number of likes on the comment isIdea** – “Yes” or “No” depending on GPT filter topic** – extracted video topic research** – 300–500 word background from Tavily hook** – engaging opening sentence for a video outline** – structured video outline Setup Steps Connect your Apify, OpenAI, Tavily, and Google Sheets credentials in n8n. Point the Google Sheets nodes to your own document and ensure the above headers exist. Replace sample API keys with your own stored in n8n Credentials. Time to set up: \~15–25 minutes for a first-time n8n user (less if you already have credentials handy). Customizing this workflow Filter logic** – Loosen the GPT filter to allow borderline ideas, or tighten it to only accept the best ones. Research depth** – Change Tavily’s search depth (e.g., depth: basic vs depth: advanced) to control how detailed the background research is. Notification channels* – Send new “Yes” ideas directly to *Slack* (#content-ideas), *Notion* (your content board), or *Email** (notify the content manager instantly). Alternative outputs** – Instead of hooks/outlines, generate: A script draft for YouTube Shorts. Blog post angles based on the same audience comments. A poll question for community engagement.
by Oneclick AI Squad
This n8n workflow monitors and alerts you about new construction projects in specified areas, helping you track competing builders and identify business opportunities. The system automatically searches multiple data sources and sends detailed email reports with upcoming projects. Good to know Email parsing accuracy depends on the consistency of request formats - use the provided template for best results. The workflow includes fallback mock data for demonstration when external APIs are unavailable. Government data sources may have rate limits - the workflow includes proper error handling. Results are filtered to show only upcoming/recent projects (within 3 months). How it works Email Trigger** - Detects new email requests with "Construction Alert Request" in the subject line Check Email Subject** - Validates that the email contains the correct trigger phrase Extract Location Info** - Parses the email body to extract area, city, state, and zip code information Search Government Data** - Queries government databases for public construction projects and permits Search Construction Sites** - Searches construction industry databases for private projects Process Construction Data** - Combines and filters results from both sources, removing duplicates Wait For Data** - Wait for Combines and filters results. Check If Projects Found** - Determines whether to send a results report or no-results notification Generate Email Report** - Creates a professional HTML email with project details and summaries Send Alert Email** - Delivers the construction project report to the requester Send No Results Email** - Notifies when no projects are found in the specified area The workflow also includes a Schedule Trigger that can run automatically on weekdays at 9 AM for regular monitoring. Email Format Examples Input Email Format To: alerts@yourcompany.com Subject: Construction Alert Request Area: Downtown Chicago City: Chicago State: IL Zip: 60601 Additional notes: Looking for commercial projects over $1M Alternative format: To: alerts@yourcompany.com Subject: Construction Alert Request Please search for construction projects in Miami, FL 33101 Focus on residential and mixed-use developments. Output Email Example Subject: 🏗️ Construction Alert: 8 Projects Found in Downtown Chicago 🏗️ Construction Project Alert Report Search Area: Downtown Chicago Report Generated: August 4, 2024, 2:30 PM 📊 Summary Total Projects Found: 8 Search Query: Downtown Chicago IL construction permits 🔍 Upcoming Construction Projects New Commercial Complex - Downtown Chicago 📍 Location: Downtown Chicago | 📅 Start Date: March 2024 | 🏢 Type: Mixed Development Description: Mixed-use commercial and residential development Source: Local Planning Department Office Building Construction - Chicago 📍 Location: Chicago, IL | 📅 Start Date: April 2024 | 🏢 Type: Commercial Description: 5-story office building with retail space Source: Building Permits [Additional projects...] 💡 Next Steps • Review each project for potential competition • Contact project owners for partnership opportunities • Monitor progress and timeline changes • Update your competitive analysis How to use Setup Instructions Import the workflow into your n8n instance Configure Email Credentials: Set up IMAP credentials for receiving emails Set up SMTP credentials for sending alerts Test the workflow with a sample email Set up scheduling (optional) for automated daily checks Sending Alert Requests Send an email to your configured address Use "Construction Alert Request" in the subject line Include location details in the email body Receive detailed project reports within minutes Requirements n8n instance** (cloud or self-hosted) Email account** with IMAP/SMTP access Internet connection** for API calls to construction databases Valid email addresses** for sending and receiving alerts API Integration Code Examples Government Data API Integration // Example API call to USA.gov jobs API const searchGovernmentProjects = async (location) => { const response = await fetch('https://api.usa.gov/jobs/search.json', { method: 'GET', headers: { 'Content-Type': 'application/json', }, params: { keyword: 'construction permit', location_name: location, size: 20 } }); return await response.json(); }; Construction Industry API Integration // Example API call to construction databases const searchConstructionProjects = async (area) => { const response = await fetch('https://www.construction.com/api/search', { method: 'GET', headers: { 'User-Agent': 'Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36', 'Accept': 'application/json' }, params: { q: ${area} construction projects, type: 'projects', limit: 15 } }); return await response.json(); }; Email Processing Function // Extract location from email content const extractLocationInfo = (emailBody) => { const lines = emailBody.split('\n'); let area = '', city = '', state = '', zipcode = ''; for (const line of lines) { if (line.toLowerCase().includes('area:')) { area = line.split(':')[1]?.trim(); } if (line.toLowerCase().includes('city:')) { city = line.split(':')[1]?.trim(); } if (line.toLowerCase().includes('state:')) { state = line.split(':')[1]?.trim(); } if (line.toLowerCase().includes('zip:')) { zipcode = line.split(':')[1]?.trim(); } } return { area, city, state, zipcode }; }; Customizing this workflow Adding New Data Sources Add HTTP Request nodes for additional APIs Update the Process Construction Data node to handle new data formats Modify the search parameters based on API requirements Enhanced Email Parsing // Custom email parsing for different formats const parseEmailContent = (emailBody) => { // Add regex patterns for different email formats const patterns = { address: /(\d+\s+[\w\s]+,\s[\w\s]+,\s[A-Z]{2}\s*\d{5})/, coordinates: /(\d+\.\d+),\s*(-?\d+\.\d+)/, zipcode: /\b\d{5}(-\d{4})?\b/ }; // Extract using multiple patterns // Implementation details... }; Custom Alert Conditions Modify the Check If Projects Found node to filter by: Project value/budget Project type (residential, commercial, etc.) Distance from your location Timeline criteria Advanced Scheduling // Set up multiple schedule triggers for different areas const scheduleConfigs = [ { area: "Downtown", cron: "0 9 * * 1-5" }, // Weekdays 9 AM { area: "Suburbs", cron: "0 14 * * 1,3,5" }, // Mon, Wed, Fri 2 PM { area: "Industrial", cron: "0 8 * * 1" } // Monday 8 AM ]; Integration with CRM Systems Add HTTP Request nodes to automatically create leads in your CRM when high-value projects are found: // Example CRM integration const createCRMLead = async (project) => { await fetch('https://your-crm.com/api/leads', { method: 'POST', headers: { 'Authorization': 'Bearer YOUR_TOKEN', 'Content-Type': 'application/json' }, body: JSON.stringify({ name: project.title, location: project.location, value: project.estimatedValue, source: 'Construction Alert System' }) }); }; Troubleshooting No emails received**: Check IMAP credentials and email filters Empty results**: Verify API endpoints and add fallback data sources Failed email delivery**: Confirm SMTP settings and recipient addresses API rate limits**: Implement delays between requests and error handling
by Fahmi Fahreza
Create Airtable records from new ClickUp Doc pages This workflow automates the process of turning content from ClickUp Docs into structured data in Airtable. When a new task is created in ClickUp with a link to a ClickUp Doc in its name, this workflow triggers, fetches the entire content of that Doc, parses it into individual records, and then creates a new record for each item in a specified Airtable base and table. Who's it for This template is perfect for content creators, project managers, and operations teams who use ClickUp Docs for drafting or knowledge management and Airtable for tracking and organizing data. It helps bridge the gap between unstructured text and a structured database. How it works Trigger: The workflow starts when a new task is created in a specific ClickUp Team. Fetch & Parse URL: It gets the new task's details and extracts the ClickUp Doc URL from the task name. Get Doc Content: It uses the URL to fetch the main Doc and all its sub-pages from the ClickUp API. Process Content: A Code node parses the text from each page. It's designed to split content by * * * and separate notes by looking for the "notes:" keyword. Find Airtable Destination: The workflow finds the correct Airtable Base and Table IDs by matching the names you provide. Create Records: It loops through each parsed content piece and creates a new record in your specified Airtable table. How to set up Configure the Set Node: Open the "Configure Variables" node and set the following values: clickupTeamId: Your ClickUp Team ID. Find it in your ClickUp URL (e.g., app.clickup.com/9014329600/...). airtableBaseName: The exact name of your target Airtable Base. airtableTableName: The exact name of your target Airtable Table. airtableVerticalsTableName: The name of the table in your base that holds "Vertical" records, which are linked in the main table. Set Up Credentials: Add your ClickUp (OAuth2) and Airtable (Personal Access Token) credentials to the respective nodes. Airtable Fields: Ensure your Airtable table has fields corresponding to the ones in the Create New Record in Airtable node (e.g., Text, Status, Vertical, Notes). You can customize the mapping in this node. Activate Workflow: Save and activate the workflow. Test: Create a new task in your designated ClickUp team. In the task name, include the full URL of the ClickUp Doc you want to process. How to customize the workflow Parsing Logic:* You can change how the content is parsed by modifying the JavaScript in the Parse Content from Doc Pages Code node. For example, you could change the delimiter from * * to something else. Field Mapping:** Adjust the Create New Record in Airtable node to map data to different fields or add more fields from the source data. Trigger Events:** Modify the Trigger on New ClickUp Task node to respond to different events, such as taskUpdated or taskCommentPosted.
by Țugui Dragoș
How It Works Story Generation – Your idea is transformed into a narrative split into scenes using DeepSeek LLM. Visuals – Each scene is illustrated with AI images via Replicate, then animated into cinematic video clips with RunwayML. Voice & Music – Narration is created using ElevenLabs (text-to-speech), while Replicate audio models generate background music. Final Assembly – All assets are merged into a professional video using Creatomate. Delivery – Everything is orchestrated by n8n, triggered from Slack with /render, and the final video link is delivered back instantly. Workflow in Action 1. Trigger from Slack Type your idea with /render in Slack - the workflow starts automatically. 2. Final Video Output Receive a polished cinematic video link in Slack. 3. Creatomate Template ⚠️ Important: You must create your own template in Creatomate. This is a one-time setup - the template defines where the voiceover, music, and video clips will be placed. The more detailed and refined your template is, the better the final cinematic result. Required APIs To run this workflow, you need accounts and API keys from the following services: DeepSeek – Story generation (LLM) Replicate – Images & AI music generation RunwayML – Image-to-video animations ElevenLabs – Text-to-speech voiceovers Creatomate – Video rendering and templates Dropbox – File storage and asset syncing Slack – Workflow trigger and video delivery Setup Steps Import the JSON workflow into your n8n instance. Add your API credentials for each service above. Create a Creatomate template (only once) – define layers for visuals, voice, and music. Trigger the workflow from Slack with /render Your Story Idea. Receive your final cinematic video link directly in Slack. Use Cases Automated YouTube Shorts / TikToks for faceless content creators. Scalable ad creatives and marketing videos for agencies. Educational explainers** and onboarding videos generated from text. Rapid prototyping** of cinematic ideas for developers & storytellers. With this workflow, you’re not just using AI tools – you’re running a full AI-powered studio in n8n.
by Club de Inteligencia Artificial Politécnico CIAP
🤖 Interactive Academic Chatbot (Telegram + MongoDB) Overview 📋 This project is a template for building a complete academic virtual assistant using n8n. It connects to Telegram, answers frequently asked questions by querying MongoDB, keeps the community informed about key dates (via web scraping), and collects user feedback for continuous improvement. How It Works Architecture and Workflow ⚙️ n8n: Orchestration of 3 workflows (chatbot, scraping worker, announcer). Telegram: Frontend for user interaction and sending announcements. MongoDB: Centralized database for FAQs, academic calendar, and feedback logs. Web Scraping: HTTP Request and HTML Extract nodes to read the university's web calendar. Cron: For automatic periodic executions (daily and weekly). Core Processes 🧠 Real-time reception of user queries via Telegram. Querying MongoDB collections for FAQ answers and calendar dates. Daily scraping of the university website to keep the calendar updated. Instant logging of user feedback (👍/👎) in MongoDB. Proactive sending of weekly announcements to the Telegram channel. Key Benefits ✅ Complete automation of student communication 24/7. An always-accurate academic calendar database without manual intervention. A built-in continuous improvement system through user feedback. Proactive communication of important events to the entire community. Use Cases 💼 Automation of student support in universities, colleges, and institutions. A virtual assistant for any organization needing to manage FAQs and a dynamic calendar. An automated announcements channel to keep a community informed. Requirements 👨💻 n8n instance (self-hosted or cloud). Credentials for a Telegram Bot (obtained from @BotFather). Credentials for a MongoDB database (Connection URI). URL of the academic calendar to be scraped. Authors 👥 Doménica Amores Nicole Guevara Adrián Villamar Mentor: Jaren Pazmiño Applicants to the CIAP Polytechnic Artificial Intelligence Club