by Paul
Gmail AI Email Manager - Setup Guide 🎯 Workflow Overview This workflow will create an intelligent Gmail email manager that can: Monitor incoming emails via webhook Analyze email content using AI Categorize emails automatically Generate smart responses Take actions based on email content Send notifications for important emails 📋 Pre-Setup Checklist Before we build the workflow, let me gather the necessary information and validate our approach. Phase 1: Discovery & Planning [ ] Search for Gmail nodes [ ] Find AI analysis nodes [ ] Identify webhook trigger options [ ] Check notification nodes Phase 2: Configuration Requirements [ ] Gmail API credentials [ ] AI service (OpenAI/Claude) API key [ ] Webhook URL setup [ ] Email classification rules 🔧 Setup Instructions Step 1: Gmail API Setup Go to Google Cloud Console Create new project or select existing Enable Gmail API Create OAuth 2.0 credentials Add authorized redirect URI: https://your-n8n-instance.com/rest/oauth2-credential/callback Step 2: AI Service Setup Choose one of the following: OpenAI**: Get API key from platform.openai.com Claude**: Get API key from console.anthropic.com Local AI**: Set up Ollama or similar Step 3: n8n Credentials Gmail OAuth2: Add client ID, secret, and scopes AI Service: Add API key Webhook: Configure webhook URL Gmail AI Email Manager - Setup Guide 🔧 Quick Setup Checklist 1. Google Cloud Console [ ] Enable Gmail API [ ] Create OAuth2 credentials [ ] Add redirect URI: https://your-n8n.com/rest/oauth2-credential/callback [ ] Set up Gmail push notifications with Pub/Sub 2. API Keys [ ] Get OpenAI API key from platform.openai.com [ ] Create Google Sheets for logging (optional) 3. n8n Credentials [ ] Gmail OAuth2: Client ID, Secret, Scopes: gmail.readonly,gmail.modify,gmail.compose [ ] OpenAI API: Your API key 4. Gmail Labels (Create these) [ ] URGENT (red) [ ] IMPORTANT (orange) [ ] PROMOTIONAL (purple) [ ] PERSONAL (green) [ ] WORK (blue) [ ] SPAM (gray) 5. Update Workflow Values [ ] High Priority Alert: Change notification email [ ] Spreadsheet Log: Update sheet ID (if using) [ ] Webhook: Copy URL after saving workflow 6. Test [ ] Save & activate workflow [ ] Send test email to Gmail [ ] Check execution log [ ] Verify auto-categorization works That's it! Your AI email manager is ready! 🚀
by Jimleuk
This n8n workflow demonstrates how to automate indexing of images to build a object-based image search. By utilising a Detr-Resnet-50 Object Classification model, we can identify objects within an image and store these associations in Elasticsearch along with a reference to the image. How it works An image is imported into the workflow via HTTP request node. The image is then sent to Cloudflare's Worker AI API where the service runs the image through the Detr-Resnet-50 object classification model. The API returns the object associations with their positions in the image, labels and confidence score of the classification. Confidence scores of less the 0.9 are discarded for brevity. The image's URL and its associations are then index in an ElasticSearch server ready for searching. Requirements A Cloudflare account with Workers AI enabled to access the object classification model. An ElasticSearch instance to store the image url and related associations. Extending this workflow Further enrich your indexed data with additional attributes or metrics relevant to your users. Use a vectorstore to provide similarity search over the images.
by Lucas Walter
Who's it for Content creators, social media managers, and marketing teams who want to automatically extract the most engaging clips from long-form YouTube videos and identify content with high viral potential. What it does This workflow analyzes any YouTube video using Vizard AI's clipping technology and automatically generates up to 8 short clips with viral score ratings. It then filters for the highest-scoring clips (9/10 or above) and posts them to a designated Slack channel for team review and distribution. How it works Video submission: Enter a YouTube URL through a user-friendly form AI analysis: Submits the video to Vizard AI for automated clipping and viral score analysis Smart polling: Waits for processing completion and retrieves results Quality filtering: Only surfaces clips with viral scores of 9/10 or higher Team notification: Posts results to Slack with clip titles, scores, and download links Requirements Vizard AI API credentials (sign up at vizard.ai) Slack workspace with OAuth app configured How to set up Configure Vizard AI credentials: Add your Vizard AI API key to the HTTP Request nodes Set up Slack integration: Configure the Slack OAuth2 credentials and select your target channel Customize filtering: Adjust the viral score threshold in the filter node (currently set to 9/10) Test the workflow: Submit a test YouTube URL to ensure everything works properly How to customize the workflow Adjust clip quantity**: Modify the maxClipNumber parameter (currently 8) in the initial API request Change viral score threshold**: Update the filter condition to match your quality standards Extend with automation**: Connect to social media posting tools or caption generation workflows for full automation Add scheduling**: Integrate with webhook triggers, scheduled triggers, or RSS feeds for batch processing videos
by Don Jayamaha Jr
Instantly access real-time decentralized exchange (DEX) insights directly in Telegram! This workflow integrates the DexScreener API with GPT-4o-powered AI and Telegram, allowing users to fetch the latest blockchain token analytics, liquidity pools, and trending tokens effortlessly. Ideal for crypto traders, DeFi analysts, and investors who need actionable market data at their fingertips. How It Works A Telegram bot listens for user queries about tokens or trading pairs. The workflow interacts with the DexScreener API (no API key required) to fetch real-time data, including: Token fundamentals (profiles, images, descriptions, and links) Trending and boosted tokens (hyped projects, potential market movers) Trading pair analytics (liquidity, price action, volumes, volatility) Order and payment activity (transaction insights, investor movements) Liquidity pool depth (market stability, capital flows) Multi-chain pair comparisons (performance tracking across networks) An AI-powered language model (GPT-4o-mini) enhances responses for better insights. The workflow logs session data to improve user interaction tracking. The requested DEX insights are sent back via Telegram in an easy-to-read format. What You Can Do with This Agent This AI-driven Telegram bot enables you to: ✅ Track trending and boosted tokens before they gain mainstream traction. ✅ Monitor real-time liquidity pools to assess token stability. ✅ Analyze active trading pairs across different blockchains. ✅ Identify transaction trends by checking paid orders for tokens. ✅ Compare market activity with detailed trading pair analysis. ✅ Receive instant insights with AI-enhanced responses for deeper understanding. Set Up Steps Create a Telegram Bot Use @BotFather on Telegram to create a bot and obtain an API token. Configure Telegram API Credentials in n8n Add your Telegram bot token under Telegram API credentials. Deploy and Test Send a query (e.g., "SOL/USDC") to your Telegram bot and receive real-time insights instantly! 🚀 Unlock powerful, real-time DEX insights directly in Telegram—no API key required! 📺 Setup Video Tutorial Watch the full setup guide on YouTube:
by David Olusola
🔍 What This Workflow Does This RAG Pipeline in n8n automates document ingestion from Google Drive, vectorizes it using OpenAI embeddings, stores it in Pinecone, and enables chat-based retrieval using LangChain agents. Main Functions: 📂 Auto-detects new files uploaded to a specific Google Drive folder. 🧠 Converts the file into embeddings using OpenAI. 📦 Stores them in a Pinecone vector database. 💬 Allows a user to query the knowledge base through a chat interface. 🤖 Uses a GPT-4o-mini model with LangChain to generate intelligent responses using retrieved context. ⚙️ Setup Instructions Connect Accounts Ensure these services are connected in n8n: ✅ Google Drive (OAuth2) ✅ OpenAI ✅ Pinecone You can do this in n8n > Credentials > New and use the matching names from the file: Google Drive: "Google Drive account 2" OpenAI: "OpenAi success" Pinecone: "PineconeApi account 2" Folder Setup Upload your documents to this folder in Google Drive: 📁 Power Folder The workflow is triggered every minute when a new file is uploaded. Workflow Overview A. File Ingestion Path Google Drive Trigger — detects new file. Google Drive (Download) — downloads the new file. Recursive Text Splitter — splits text into chunks. Default Data Loader — loads content as LangChain documents. OpenAI Embeddings — converts text chunks into embeddings. Pinecone Vector Store — stores them in "ragfile" index. B. Chat Retrieval Path When chat message received — AI Agent — LangChain agent managing tools. OpenAI Chat Model (GPT-4o-mini) — generates replies. Pinecone Vector Store (retrieval) — retrieves matching content. Embeddings OpenAI1 — helps match queries to document chunks.
by Automate With Marc
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 🚀 Start-up Ideas Generator – From Idea to Executive Summary in Minutes Description: This AI-powered n8n workflow helps you brainstorm and validate start-up ideas, perform market research using Perplexity Sonar, and automatically generate a professional business plan — all within minutes. It’s designed for entrepreneurs, accelerators, venture studios, and ideation teams looking to go from a single prompt to a detailed proposal ready to pitch or develop. Watch step-by-step build video of n8n workflows like this: https://www.youtube.com/@Automatewithmarc Setup Instructions Required Credentials: OpenAI API Key – For GPT-4.1 Perplexity Sonar Access Token – For deep research Anthropic Claude API Key – For business plan writing Google Service Account Credentials – To write to Docs How It Works: 💬 Chat Trigger Start the process by typing a simple request like “Give me a few start-up ideas in AI tech.” 🧠 AI Research Agent (OpenAI + Perplexity Sonar) The system defines the research scope, taps into Perplexity AI for deep market scans, and outputs high-potential business opportunities including: Market size Customer pain points Competition overview Differentiation strategy 📄 Business Case Generator (Claude) Each opportunity is expanded into a complete business plan using Claude Sonnet, covering: Executive summary Market analysis Product description Competitor gap Business model & GTM Financials & roadmap 📃 Google Docs Export The full business plan is automatically inserted into a connected Google Doc for easy sharing, editing, or pitching. Google Docs Configuration: Create a Google Doc titled "Startup Business Plan" (or adjust the title in the node settings) Share the document with your Google service account email Update the Document ID field in the Google Docs node accordingly Ensure that the structure accepts plain text input — formatting is handled by the node Tools & Models Used: LangChain Chat Trigger OpenAI GPT-4.1 (Research Prompt Structuring) Perplexity Sonar Deep Research (Market Research) Anthropic Claude Sonnet (Business Plan Writing) Google Docs Node (Formatted Output) Use Cases: Rapid ideation for venture building or incubators Validating start-up ideas before prototyping Automating market research + proposal writing Generating investor-ready pitch materials
by keisha kalra
Try It Out! This n8n template creates a fully automated Instagram content schedule using AI and Google Sheets. It is perfect for content creators, marketing teams, or local businesses looking to organize and scale their social media posting. How it works The workflow starts by reading two sets of inputs from a Google Sheet: Your content strategy inputs (Pillar, Objective, Frequency, Format, Structure, Examples). A list of scraped blog posts with title, URL, and description (fetched from your website). Blog posts are scraped using Apify and parsed to extract key fields, which are stored in a tab labeled "Input (blog month)". You can assign a preferred posting month for each blog (e.g. fall blog posts get tagged for September). The workflow then merges both inputs and extracts the relevant information for further information added by ChatGPT. AI Scheduling & Personalization Once merged, the workflow loops through each content item and: Identifies if the scheduled post falls on or near a holiday (like Mother’s Day) and adjusts the content accordingly. A reference tool is attached to guide structure and tone, based on a library of post examples. Sends the content to an AI Agent (using GPT-4, but customizable) that generates: A compelling Instagram caption A visual description Hashtags Suggested post date, day, content pillar, and format (carousel, reel, image, etc.) Output All generated content including captions, structure, dates, hashtags, and pillar is exported into a tab titled Output in your Google Sheet. The final schedule is ready for manual review, editing, or publishing to social media. How to use The workflow uses a manual trigger to start, but you can replace it with a Webhook, cron job, or form submission. Add/edit your content strategy in Google Sheets. How to Set-Up Initial Input Tab Define your content pillars and structure Create a tab named "Input" or "Strategy" Include these columns: Pillar: e.g., Family images Objective: e.g., Showcase images Frequency: e.g., Bi-weekly Content Form: e.g., Images, Reels Structure: brief description of expected layout (e.g., carousel Q&A, singular photo) Examples: prompts or questions to guide AI (e.g., Why do you think families should do a session?) Input (blog month) Tab – Store scraped blog content Include these columns: URL: direct link to blog post Title: blog post title Description: short summary of the post Preferred Month: month you want it posted (e.g., August, September) This sheet is partially auto-filled by the workflow (except for Preferred Month) Output Tab – Final scheduled content Include these columns: Date: scheduled posting date (YYYY-MM-DD) Day: day of the week Pillar: content category assigned Format: e.g., Images, Reels, Carousel Description: visual summary Caption: Instagram-ready caption Hashtags: complete hashtag block To use the Apify HTTP Request node: Drag in an HTTP Request node into your n8n workflow. Set the Method and URL based on how you're using Apify: Use POST if you want to run an actor live with dynamic input (e.g. scrape blog posts in real time). Use GET if you want to retrieve results from a completed or static dataset run (faster and cheaper if you're reusing previous data). Configure query or body parameters: Include your Apify API token for authentication (e.g. token=YOUR_API_KEY) For POST: include an input object with any required actor settings (e.g., blog URL to scrape). For GET: specify the dataset ID in the URL Test the node to ensure you're retrieving the blog titles, descriptions, and URLs as expected. Requirements Apify account for scraping blog posts OpenAI key (e.g. GPT-4) or another model of your choice Google Sheets Credentials Example Use Cases A photographer repurposing blogs into Instagram carousels A nonprofit automatically generating seasonal posts A small team managing multi-pillar content across weeks or months Need Help? Join the n8n Discord or ask in the n8n Forum! Happy Content Making ! 📅✨
by LukaszB
Crypto Price Alert – n8n Workflow A simple and effective crypto alert system for anyone who wants to stay up to date with coin price changes — without refreshing charts all day. This workflow checks the current price of your chosen cryptocurrency (via CoinGecko) and sends you an alert on Discord if it goes above or below your target range. It’s lightweight, easy to set up, and runs on autopilot. What the Workflow Does Checks the live price of a selected coin using the CoinGecko API. Compares it to the max/min prices you define manually. Decides if the price is too high or too low. Sends an alert message to Discord depending on the result. How It Works The flow is triggered manually or on a schedule (your choice). It pulls the current price of the coin you set. Compares that price with your min and max values. Sends a “high” or “low” message to your Discord webhook. Setup Steps Enter your coin ID and price thresholds in the “Set Low and High” node. Paste your Discord webhook URLs in the "Message High" and "Message Low" nodes. Optional: Adjust the schedule trigger to run every X minutes/hours. Run once manually to test — takes under 1 minutes. Full instructions and config tips are in sticky notes inside the workflow.
by Yaron Been
Automate expense reviews with AI-powered CFO-level analysis. This workflow monitors Airtable expense submissions, uses GPT-4 to analyze expenses like an experienced CFO, flags suspicious expenses with detailed reasoning, and maintains comprehensive audit trails in Pinecone vector database. 🚀 What It Does Smart Monitoring**: Watches Airtable for new expense submissions AI CFO Analysis**: GPT-4 applies financial expertise to review amounts, categories, and descriptions Intelligent Flagging**: Automatically identifies policy violations and suspicious patterns Audit Trail**: Stores all decisions in Pinecone for compliance and searchability Auto Updates**: Updates Airtable records with AI decisions and detailed reasoning 🎯 Perfect For Finance teams needing intelligent expense oversight CFOs wanting to automate expense policy enforcement Growing companies scaling expense management Businesses requiring compliance documentation ⚙️ Key Benefits ✅ 99% faster expense processing vs manual review ✅ CFO-level intelligence applied to every expense ✅ Complete audit trail for compliance ✅ Real-time fraud detection and policy enforcement ✅ Detailed explanations for every decision 🔧 What You Need Airtable base with expense data (template included) OpenAI API for GPT-4 analysis Pinecone account for audit trail storage Basic expense submission process 📊 Sample Results Input: $4,500 business class flight to Tokyo AI Decision: "Flagged - Amount exceeds typical travel thresholds. Requires verification against travel policies and client justification for premium travel." 🛠️ Setup & Support Quick Setup: Deploy in 60 minutes with included templates and documentation YouTube: https://www.youtube.com/@YaronBeen/videos 💼 Expert Support LinkedIn: https://www.linkedin.com/in/yaronbeen/ 📧 Direct Help Email: Yaron@nofluff.online Transform expense management from manual bottleneck to intelligent automation. Let AI handle policy compliance while your finance team focuses on strategy.
by Olek
How it works This workflow will activate and deactivate a selected other workflow on schedule. > ⚠️ Warning! > This approach won't work for trial users as it requires n8n API that is not available to trial users. > See https://docs.n8n.io/api/ for details. Set up steps Adjust activation/deactivation schedule per your needs. Custom (cron) interval is a recommended approach. Set targeted Workflow ID. You will find it in the URL of the workflow you want to manage. Set n8n API credentials: Create an API key: how to Create n8n credentials using the API key: how to This workflow uses n8n node. #DevOps #workflow-management Other useful stuff Need a universal Error workflow to catch both execution and trigger errors? Here you go: Error handling: Send email via Gmail on execution or trigger-level errors More stuff by Olek and do not forget to backup your workflows often by automating.
by Agus Narestha
🔒 SSL Certificate Monitoring & Expiry Alert with Spreadsheet [FREE APIs] ✅ What This Workflow Does This n8n template automatically monitors SSL certificates of websites listed in a Google Sheet and sends email alerts if any are expiring within 14 days. It helps ensure you avoid downtime, security issues, and trust warnings due to expired certificates. 🧩 Key Features 📅 Weekly Automation: Runs every Monday at 7:00 AM (configurable). 📄 Google Sheets Integration: Fetches and updates data in a spreadsheet. 🔍 SSL Check via API: Uses ssl-checker.io to get certificate details. ⚠️ SSL Expiry Filter: Identifies certificates expiring within 14 days. 📧 Email Alerts: Sends notifications for certificates close to expiration. 📂 Input Spreadsheet Format Your Google Sheet should have the following columns: | No | Name | Link | SSL Issued On | SSL Expired On | SSL Status | |----|-----------------|-----------------------|-------------------|-------------------|------------| | 1 | Example Site | https://example.com | 2024-07-01 | 2025-07-01 | Valid | | 2 | My Blog | https://myblog.org | 2024-07-05 | 2024-07-20 | Expiring | Each row should include a valid website URL in the Link column. 🛠️ How It Works Scheduled Trigger Executes weekly (Monday 7:00 AM). Fetch Website List Reads all website entries from the Google Sheet. Check SSL Certificates Uses ssl-checker.io API to retrieve certificate details for each website. Update Spreadsheet Writes "Issued On" and "Expired On" fields back to the spreadsheet. Evaluate SSL Expiry Filters for certificates expiring within 14 days. Check Condition Determines whether to send alerts based on filtered results. Send Email Alert Notifies via email if any certificates are expiring soon. 📬 Example Email Output Subject: ⚠️ ALERT!! SSL EXPIRED SSL certificates expiring soon: example.com (expires in 5 days) anotherdomain.net (expires in 3 days) 🧰 Setup Requirements A Google Sheet with the correct columns and website links. SMTP credentials to send alert emails.
by Taiki
Workflow Setup Guide This workflow collects the most-viewed videos from specified YouTube channels and saves the data to a Google Sheet. Follow these steps to set it up: 1. Credentials Setup Google Sheets:** You need to have a Google Sheets credential configured in your n8n instance. If you don't have one, go to the 'Credentials' section in n8n and add a new credential for Google Sheets. YouTube API Key:** You need a YouTube Data API v3 key. Go to the Google Cloud Console. Create a new project or select an existing one. Go to 'APIs & Services' > 'Library' and enable the YouTube Data API v3. Go to 'APIs & Services' > 'Credentials', click 'Create Credentials', and choose 'API key'. Copy the generated API key. 2. Google Sheet Setup You will need one Google Sheet with two separate sheets (tabs) inside it. Input Sheet Use Template This sheet provides the list of YouTube channels to process. Required Columns:** Create a sheet with the following two columns: ChannelID: The ID of the YouTube channel (e.g., T7M3PpjBZzw). video_num_to_get: The number of top videos to retrieve for that channel (e.g., 5). Output Sheet This sheet is where the results will be saved. Required Columns:** The workflow will automatically append data to the following columns. You can create them beforehand or let the workflow do it. channelName title videoId videoLink 3. Node Configuration Read Channel Info from Sheet:** Select your Google Sheets credential. Enter your Spreadsheet ID. Enter the name of your Input Sheet. Fetch Most-Viewed Videos via YouTube API:** Replace YOUR_YOUTUBE_API_KEY with the API key you generated in Step 1. Append Video Details to Sheet:** Select your Google Sheets credential. Enter your Spreadsheet ID (the same one as before). Enter the name of your Output Sheet.