by Parth Pansuriya
AI Meeting Summary Generator with Google Docs Integration Who’s it for Teams that record meetings and want fast, clear summaries without manual note-taking. Managers who need action items extracted automatically. Anyone using Google Drive + Google Docs as their central workspace. How it works / What it does This workflow automates meeting documentation: Watches a Google Drive folder for new audio/video meeting files. Downloads the file and transcribes speech into text using Gemini AI. Summarizes transcripts into Key Discussions and Action Items. Creates or updates a Google Doc with the formatted summary (title, bullets, checkmarks, styling). Sends final output to Docs with bold headings, bullets, and spacing for readability. How to set up Add your Google Drive Trigger to monitor a folder. Connect Gemini AI to handle transcription + summarization. Configure the Google Docs Tool to create/update your summary documents. (Optional) Use the Code Node + Docs API to apply bullet/checkmark formatting. Requirements Google Drive OAuth2 – for monitoring & downloading files Google Docs OAuth2 – for creating and updating documents Google Gemini API – for transcription + AI-powered summarization How to customize the workflow Change the Google Drive folder to monitor a different workspace. Edit the system prompt in the Summarizer to tweak summary style (e.g., more detail, decisions only, etc.). Modify the Code Node formatting rules (bullets, checkmarks, bold text). Add integrations (e.g., Slack, Email, Notion) to send summaries beyond Google Docs.
by Robert Breen
This n8n workflow template creates an intelligent data analysis chatbot that can answer questions about data stored in Google Sheets using OpenAI's GPT-5 Mini model. The system automatically analyzes your spreadsheet data and provides insights through natural language conversations. What This Workflow Does Chat Interface**: Provides a conversational interface for asking questions about your data Smart Data Analysis**: Uses AI to understand column structures and data relationships Google Sheets Integration**: Connects directly to your Google Sheets data Memory Buffer**: Maintains conversation context for follow-up questions Automated Column Detection**: Automatically identifies and describes your data columns 🚀 Try It Out! 1. Set Up OpenAI Connection Get Your API Key Visit the OpenAI API Keys page. Go to OpenAI Billing. Add funds to your billing account. Copy your API key into your OpenAI credentials in n8n (or your chosen platform). 2. Prepare Your Google Sheet Connect Your Data in Google Sheets Data must follow this format: Sample Marketing Data First row** contains column names. Data should be in rows 2–100. Log in using OAuth, then select your workbook and sheet. 3. Ask Questions of Your Data You can ask natural language questions to analyze your marketing data, such as: Total spend** across all campaigns. Spend for Paid Search only**. Month-over-month changes** in ad spend. Top-performing campaigns** by conversion rate. Cost per lead** for each channel. 📬 Need Help or Want to Customize This? 📧 rbreen@ynteractive.com 🔗 LinkedIn 🔗 n8n Automation Experts
by n8nwizard
📌 Overview This advanced multi-phase n8n workflow automates the complete research, analysis, and ideation pipeline for a YouTube strategist. It scrapes competitor channels, analyzes top-performing titles and thumbnails, identifies niche trends, gathers audience sentiment from comments, and produces data-driven content ideas—automatically writing them into a structured Google Sheets dashboard. This system is ideal for: YouTube creators Agencies Content strategists Automation engineers Anyone who wants to generate YouTube content ideas backed by real data 🧠 High‑Level Architecture The workflow is split into 5 phases, each building on the previous: Phase 1 — Niche Outliers (Input: User Form) Scrapes 3 high‑quality channels from your niche, extracts their outlier videos, and analyzes why they work. Phase 2 — Broad Niche Insights (Weekly) Scrapes the top trending content in your broad niche (e.g., "AI", "fitness", "personal finance") and logs weekly insights. Phase 3 — Niche Insights (Daily) Scrapes the top videos in your specific micro‑niche daily to keep track of content momentum. Phase 4 — Comment Analysis Analyzes real comments from your channel to understand what your audience likes, dislikes, and wants more of. Phase 5 — Content Ideation Generates 3 highly‑optimized title + thumbnail concepts using all prior insights. Everything is automatically logged into a Google Sheets dashboard. 🧩 Phase-by-Phase Breakdown ⭐ Phase 1 — Niche Outliers (Form Trigger) User enters 3 YouTube channel URLs in a form. Workflow scrapes each channel using Apify YouTube Scraper. Filters for top-performing videos. Extracts: title, views, likes, thumbnail, URL. AI analyzes: Power words in titles (OpenRouter/GPT 4.1-mini) Thumbnail attention hooks (OpenAI Vision) All insights are appended into the “Niche Outliers” sheet. Purpose: Understand what the best creators in your niche are doing. 🌐 Phase 2 — Broad Niche Insights (Weekly — Sundays @ 5 AM) Workflow scrapes the top videos for a broad niche (e.g., “artificial intelligence”). Analyzes: Title structure Power words Thumbnail cues Writes weekly insights to “Broad Niche Weekly” sheet. Purpose: Stay informed about macro‑level trends. 🎯 Phase 3 — Niche Insights (Daily @ 6 AM) Scrapes the top videos in your specific micro‑niche (e.g., “n8n automations”). Runs title + thumbnail analysis. Appends daily results to “Niche Daily”. Results feed directly into Phase 5. Purpose: Track daily momentum and trending formats. 💬 Phase 4 — Comment Analysis (Channel Feedback) Scrapes your channel’s latest 5 videos. Extracts up to 30 comments from each. Aggregates comments. AI identifies: What viewers love What viewers dislike What viewers are asking for Stores patterns in “Comment Analysis” sheet. Purpose: Understand real audience demand. 💡 Phase 5 — Content Ideation Using AI Using insights from all previous phases: Top titles Power words Thumbnail patterns Daily niche trends Audience comment analysis Channel positioning The Creative Agent produces: 3 optimized video titles 3 matching thumbnail concepts These are appended to the “Ideation” sheet. A Slack notification is sent when ideation is ready. Purpose: Fully automated content idea generation. 🗂️ Outputs in Google Sheets The workflow populates these tabs: 📌 Niche Outliers (top competitor videos) 📌 Broad Niche Weekly (weekly trend analysis) 📌 Niche Daily (daily trend analysis) 📌 Comment Analysis (audience sentiment) 📌 Ideation (final titles + thumbnails) 🔧 What This Workflow Automates ✔ Competitor analysis ✔ Thumbnail + title breakdowns ✔ Daily niche tracking ✔ Weekly niche tracking ✔ Viewer sentiment analysis ✔ Fully AI‑generated content ideas ✔ Automatic data logging to Google Sheets ✔ Slack notifications This is essentially a 24/7 AI YouTube strategist. ⚙️ Setup Requirements Apify API Key** (used in 5 scraper nodes) OpenRouter API Key** (for GPT 4.1-mini intelligence) OpenAI API Key** (for thumbnail image analysis) Google Sheets OAuth2 Credential** Make a copy of the provided sheet template Fill Set nodes: Phase II: Broad niche (e.g., "AI") Phase III: Micro niche (e.g., "n8n automations") Phase IV: Your Channel URL Phase V: Your Channel Description 🧪 Testing Guide Test the Form Trigger with 3 competitor channel URLs. Test both Schedule Triggers (weekly + daily) manually. Verify Sheets are receiving rows. Run the full pipeline end‑to‑end. Confirm Slack notification. Everything should chain together smoothly. 🎉 Final Result By the end of this workflow, you have a: 🧠 Data‑driven YouTube strategy system that: studies your niche finds outliers understands your audience detects trends generates smart content ideas every day
by PollupAI
Who it’s for Built for Customer Success and Account Management teams focused on proactive retention. This workflow helps you automatically identify at-risk customers – before they churn – by combining CRM, usage, and sentiment data into one actionable alert. What it does This end-to-end workflow continuously monitors customer health by consolidating data from HubSpot and Google Sheets. Here’s how it works: Fetch deals from HubSpot. Collect context — linked support tickets and feature usage from a Google Sheet. Run sentiment analysis on the tickets to generate a customer health score. Evaluate risk — an AI agent reviews deal age, sentiment score, and usage trends against predefined thresholds. Send alerts — if churn risk is detected, it automatically sends a clear, data-driven email to the responsible team member with next-step recommendations. How to set it up To get started, configure your credentials and parameters in the following nodes: Credentials: HubSpot: Connect your account (HubSpot: Get All Deals). LLM Model: Add credentials for your preferred provider (Config: Set LLM for Agent & Chains). Google Sheets: Connect your account (Tool: Get Feature Usage from Sheets). Email: Set up your SMTP credentials (Email: Send Churn Alert). Tool URLs: In Tool: Calculate Sentiment Score, enter the Webhook URL from the Trigger: Receive Tickets for Scoring node within this same workflow. In Tool: Get HubSpot Data, enter the Endpoint URL for your MCP HubSpot data workflow. (Note: This tool *does call an external workflow)*. Google Sheet: In Tool: Get Feature Usage from Sheets, enter the Document ID for your own Google Sheet. Email Details: In Email: Send Churn Alert, change the From and To email addresses. Requirements HubSpot account with Deals API access LLM provider account (e.g. OpenAI) Google Sheets tracking customer feature usage n8n with LangChain community nodes enabled A separate n8n workflow set up to act as an MCP endpoint for fetching HubSpot data (called by Tool: Get HubSpot Data). How to customize it Tailor this workflow to match your business logic: Scoring logic:** Adjust the JavaScript in the Code: Convert Sentiment to Score node to redefine how customer scores are calculated. Alert thresholds:** Update the prompt in the AI Chain: Analyze for Churn Risk node to fine-tune when alerts trigger (e.g. deal age, score cutoff, or usage drop). Data sources:** Swap HubSpot or Google Sheets for your CRM or database of choice — like Salesforce or Airtable. ✅ Outcome: A proactive customer health monitoring system that surfaces risks before it’s too late — keeping your team focused on prevention, not firefighting.
by Sabrina Ramonov 🍄
Description Fully automated pipeline where you send an email to yourself with a rough idea (subject contains “thread”), n8n’s Gmail trigger picks it up, OpenAI ChatGPT rewrites/apply a viral-thread template, and Blotato posts the long-form thread to X/Twitter, Bluesky, and Meta Threads (optionally schedule or include images/videos). Template is easily extensible to other social platforms. Who Is This For? Digital creators, content marketers, social media managers, agencies, entrepreneurs, and influencers who want fast, automated long-form thread posting. 📄 Documentation Full Step-by-Step Tutorial How It Works 1. Trigger: Gmail Connect your Gmail account. n8n monitors emails sent from you and filters for subjects containing the word “thread”. 2. AI Thread Writer: OpenAI ChatGPT Connect your OpenAI account. Prompt ChatGPT to clean up your draft and format a long-form viral thread. 3. Publish to Social Media via Blotato Connect your Blotato account and choose social accounts (X/Twitter, Threads, Bluesky). Schedule or post immediately. Supports optional image/video URLs via a mediaUrls array (publicly accessible URLs). Example email to trigger the workflow: Email Subject: thread Email Body: I'm obsessed with voice AI apps. Super Whisper is my current favorite because it runs locally and keeps my voice data private. I talk to it instead of typing. Way faster. Setup & Required Accounts Gmail account (used as trigger) n8n Gmail OAuth doc: https://docs.n8n.io/integrations/builtin/credentials/google/oauth-single-service OpenAI Platform account (access to ChatGPT) Blotato account: https://blotato.com Generate Blotato API Key: Settings > API > Generate API Key (paid feature only) Sign in to Blotato and create an API Key (required for posting) n8n: Ensure "Verified Community Nodes" enabled in your n8n Admin Panel Install the "Blotato" community node and create Blotato credentials Optional: Media & Style Tweaks Attach images/videos: insert publicly accessible URLs into the mediaUrls array (advanced). To emulate a specific tone/structure, provide ChatGPT examples of your favorite viral threads or replace the example viral-thread prompt with your preferred example. Voice-to-text tip: record ideas (e.g., Superwhispr) and send the transcript by email — ChatGPT will clean it up. Tips & Tricks During testing, use “Scheduled Time” in Blotato instead of immediate posting to preview before going live. Start with a single social platform while testing. If your script is long or includes media, processing may take longer. Many users prefer speaking their ideas (voice notes) then letting AI edit — faster than typing. Troubleshooting Check your Blotato API Dashboard to inspect each request, response, and error. Confirm API key validity, n8n node credentials, and that emails sent have subject containing “thread”. Need Help? In the Blotato web app, click the orange support button in the bottom right to access Blotato support.
by Weiser22
Shopify Multilingual Product Copy with n8n & Gemini 2.5 Flash-Lite Use for free Created by <Weiser22> · Last update 2025-09-02 Categories: E-commerce, Product Content, Translation, Computer Vision Description Generate language-specific Shopify product copy (ES, DE, EN, FR, IT, PT) from each product’s main image and metadata. The workflow performs a vision analysis to extract objective, verifiable details, then produces product names, descriptions, and handles per language, and stores the results in Google Sheets for review or publishing. Good to know Model:** models/gemini-2.5-flash-lite (supports image input). Confirm pricing/limits in your account before scaling. Image requirement:** products should have images[0].src; add a fallback if some products lack a primary image. Sheets mapping:** the sheet node uses Auto-map; ensure your matching column aligns with the field you emit (id vs product_id). Strict output:** the Agent enforces a multilingual JSON contract (es,de,en,fr,it,pt), each with shopify_product_name, shopify_description, handle. How it works Manual Trigger:** start a test run on demand. Get many products (Shopify):** fetch products and their images. Analyze image (Gemini Vision):** send images[0].src with an objective, 3–5 sentence prompt. AI Agent (Gemini Chat):** merge Shopify fields + vision text under anti-hallucination rules and a strict JSON schema. Structured Output Parser:** validates the exact JSON shape. Expand Languages & Sanitize (Code):** split into 6 items and normalize handles/HTML content as needed. Append row in sheet (Google Sheets):** add one row per language to your spreadsheet. Requirements Shopify Access Token with product read permissions. Google AI Studio (Gemini) API key for Vision + Chat Model nodes. Google Sheets credentials (OAuth or Service Account) with access to the target spreadsheet. How to use Connect credentials: Shopify, Gemini (same key for Vision and Chat), and Google Sheets. Configure nodes: Get many products: adjust limit/filters. Analyze image: verify ={{ $json.images[0].src }} resolves to a public image URL. AI Agent & Parser: keep the strict JSON contract as provided. Code (Expand & Sanitize): emits product_id, lang, handle, shopify_product_name, shopify_description, base_handle_es. Google Sheets (Append): set documentId and tab name; confirm the matching column. Run a test: execute the workflow and confirm six rows per product (one per language) appear in the sheet. Data contract (Agent output) { "es": {"shopify_product_name": "", "shopify_description": "", "handle": ""}, "de": {"shopify_product_name": "", "shopify_description": "", "handle": ""}, "en": {"shopify_product_name": "", "shopify_description": "", "handle": ""}, "fr": {"shopify_product_name": "", "shopify_description": "", "handle": ""}, "it": {"shopify_product_name": "", "shopify_description": "", "handle": ""}, "pt": {"shopify_product_name": "", "shopify_description": "", "handle": ""} } Customising this workflow Publish to Shopify:** after review in Sheets, add a product.update step to write finalized copy/handles. Handle policy:** tweak slug rules (diacritics, separators, max length) in the Code node to match store conventions. No-image fallback:** add an IF/Switch to skip vision when images[0].src is missing and generate copy from title + body only. Tone/length:** adjust temperature and token limits on the Chat Model for brand-fit. Troubleshooting No rows in Sheets:** confirm spreadsheet ID, tab name, Auto-map status, and that the matching column matches your emitted field. Vision errors:** ensure images[0].src is reachable. Parser failures:* the Agent must return *bare JSON** with the six root keys and three fields per language—no extra text.
by Satoshi
Overview The workflow automatically gathers weekly user and page view metrics. It then uses AI to analyze, compare, and compile a summary report. Finally, it sends the report to the manager's email. How it works Get Data from GA Automatically retrieve data from Google Analytics (GA) for the two most recent weeks. Compare the data and calculate the variances between the two weeks. Generate Report Automatically analyze the data and generate reports using Artificial Intelligence (AI). Generate charts to visualize the data. Export the report to PDF. Send Report Send the report via email to the manager. Set up steps Google cloud account Create the credentials and replace them in the workflow. Please enable the following APIs: Gmail API Google Analytics Admin API Google Analytics Data API HTML to PDF account You need to install node HTML to PDF. Get API key and replace in the workflow.
by n8n Team
This workflow provides a simple example of how to use itemMatching(itemIndex: Number) in the Code node to retrieve linked items from earlier in the workflow.
by Lucas Peyrin
How it works This workflow is an interactive, hands-on tutorial designed to teach you the absolute basics of JSON (JavaScript Object Notation) and, more importantly, how to use it within n8n. It's perfect for beginners who are new to automation and data structures. The tutorial is structured as a series of simple steps. Each node introduces a new, fundamental concept of JSON: Key/Value Pairs: The basic building block of all JSON. Data Types: It then walks you through the most common data types one by one: String (text) Number (integers and decimals) Boolean (true or false) Null (representing "nothing") Array (an ordered list of items) Object (a collection of key/value pairs) Using JSON with Expressions: The most important step! It shows you how to dynamically pull data from a previous node into a new one using n8n's expressions ({{ }}). Final Exam: A final node puts everything together, building a complete JSON object by referencing data from all the previous steps. Each node has a detailed sticky note explaining the concept in simple terms. Set up steps Setup time: 0 minutes! This is a tutorial workflow, so there is no setup required. Simply click the "Execute Workflow" button to run it. Follow the instructions in the main sticky note: click on each node in order, from top to bottom. For each node, observe the output in the right-hand panel and read the sticky note next to it to understand what you're seeing. By the end, you'll have a solid understanding of what JSON is and how to work with it in your own n8n workflows.
by Mohamed Salama
Let AI agents fetch communicate with your Bubble app automatically. It connects direcly with your Bubble data API. This workflow is designed for teams building AI tools or copilots that need seamless access to Bubble backend data via natural language queries. How it works Triggered via a webhook from an AI agent using the MCP (Model-Chain Prompt) protocol. The agent selects the appropriate data tool (e.g., projects, user, bookings) based on user intent. The workflow queries your Bubble database and returns the result. Ideal for integrating with ChatGPT, n8n AI-Agents, assistants, or autonomous workflows that need real-time access to app data. Set up steps Enable access to your Bubble data or backend APIs (as needed). Create a Bubble admin token. Add your Bubble node/s to your n8n workflow. Add your Bubble admin token. Configer your Bubble node/s. Copy the generated webhook URL from the MCP Server Trigger node and register it with your AI tool (e.g., LangChain tool loader). (Optional) Adjust filters in the “Get an Object Details” node to match your dataset needs. Once connected, your AI agents can automatically retrieve context-aware data from your Bubble app, no manual lookups required.
by Tobi Adeleke
How it works Downloads markdown documents from a Google Drive folder containing private information, uses the Ollama model to identify and extract sensitive data, and stores the sanitized text in a single combined markdown document. Setup Steps Google cloud OAuth credentials for accessing google drive. A local Ollama model (currently using Ollama 3.1). Local instance of N8N (for example, running in docker).
by Vitaliy
Pearl Hybrid Intelligence: AI draft + expert verification (across 100+ expert domains) Deliver higher-confidence answers by combining an AI assistant with professional human verification. Pearl provides access to a network of 20,000 licensed experts (including JDs, MDs, PhDs, DVMs, CPAs, engineers) across 100+ domains - Legal and Health are just examples, and the prompts can be tuned for many other areas. What this template does Lets users ask questions in a conversational format (your app sends the conversation history). Collects any missing intake details before answering. Generates a clear draft answer using AI. Sends the draft to a Pearl expert for professional verification. Returns the verified answer back to the user, including who verified it. How it works (non-technical) A user asks a question. If the assistant needs more context, it asks one focused follow-up question. Once the assistant has enough information, it drafts a complete answer. That draft is automatically reviewed by a qualified Pearl expert. The user receives the verified answer, along with expert attribution. Setup steps (10–15 minutes) Connect your OpenAI credential in n8n (used for intake + drafting). Add your Pearl MCP Server API key (used for expert verification). You can request a demo access key here: https://www.pearl.com/enterprise/contact-get-started Activate the workflow and send a test request using the example payload below. Customization The intake and draft answer prompts can be adjusted to match end-user needs (domain, tone, risk policy, compliance requirements). You can also modify the response format (fields returned, disclaimers, attribution formatting) without changing the core flow. Input payload format (conversation history) Send a JSON body like: { "model": "gpt-4o-mini", "messages": [ { "role": "user", "content": "My question..." } ] }