by Jaruphat J.
Who is this for? This workflow is perfect for digital content creators, marketers, and social media managers who regularly create engaging short-form videos featuring inspirational or motivational quotes. While the workflow is universally applicable, it specifically highlights Thai as an example to demonstrate effective language and font integration. What problem is this workflow solving? Creating consistent and engaging multilingual video content manually, including attractive fonts and proper video formatting, is time-consuming and repetitive. Additionally, managing files, background music, and updating statuses manually can be tedious and prone to errors. What this workflow does Automatically fetches background video and music files stored on Google Drive. Randomly selects a quote (demonstrated with Thai language) and author information from Google Sheets. Dynamically combines the selected quote and author text using appealing fonts, such as the Thai font "Kanit," directly onto the video using FFmpeg on your n8n local environment. Creates visually engaging videos with a 9:16 aspect ratio, optimized for YouTube Shorts and other vertical video platforms. Automatically uploads the finalized video to YouTube. Updates the status and YouTube URL back into your Google Sheet, ensuring you have up-to-date records. Setup Requirements: This workflow requires a self-hosted n8n instance, as the execution of FFmpeg commands is not supported on n8n Cloud. Ensure FFmpeg is installed on your self-hosted environment. Google Sheets Setup: Your Google Sheet must include at least these columns: Index: (Unique identifier for each quote) Quote: (Text of the quote) Author: (Author of the quote) CreateStatus: (Track video creation status; values like 'DONE' or blank for pending) YoutubeURL: (Automatically updated after upload) To help you get started quickly, you can use this template spreadsheet. Next steps: Organize your video and music files in separate folders in Google Drive. Authenticate your Google Sheets, Google Drive, and YouTube accounts in n8n. Ensure fonts compatible with your target languages (such as Kanit for Thai) are available in your FFmpeg installation. How to customize this workflow to your needs Fonts:** Adjust font styles and sizes within the workflow's code node. Ensure the fonts you choose fully support the language you wish to use. Quote Management:** Easily add or remove quotes and authors in your Google Sheets document. Media Files:** Change or update background videos and music by modifying the files in your Google Drive folders. Video Specifications:** Customize video dimensions, text positioning, opacity, and music volume directly in the provided FFmpeg commands. Benefits of Using Localized Fonts and Quotes Utilizing fonts specific to your target language, as demonstrated with Thai, significantly increases audience engagement by making your content more relatable, shareable, and visually appealing. Ensure you select fonts that properly support the language you're targeting.
by Jimleuk
This n8n template monitors an Outlook mailbox for invoices, automatically parses/extracts data from them and then uploads the output to an Excel Workbook. One of my top workflow requests, this template can save many hours of manual labour for you or your finance/accounts team. How it works A scheduled trigger is set to fetch recent Outlook messages to the Accounts receivable mailbox. Each message is analysed to determine whether or not it from a supplier and is issuing/contains an invoice. For each valid message, the attachments are downloaded and non-invoice documents are filtered out via AI Vision classification. Invoices are then processed through a AI vision model again to extract the details. The extracted data can then be used for reconciliation or otherwise. For this demonstration, we'll just append the row to an Excel sheet for now. How to use Ensure your Microsoft365 credential points to the correct mailbox. If a shared folder is used, toggle "shared folder" option to "on" and for the principal ID, use the email address. If you receive lots of other types of messages such as replies and forwards, you may want to implement additional checks to prevent processing invoices twice. The "remove duplicates" node can help with this. Requirements Outlook for Mailbox Google Gemini for Document Understanding and Invoice Extraction Excel for Data Storage Customising this workflow Note the assumption for this template is that all invoices will come as a PDF attachment. In real life, this is rarely the case! Adding in document conversion to cover all invoice formats. Human feedback is also an important factor in AI workflows. Try tagging emails as a way to notify team members that the invoice was processed.
by Don Jayamaha Jr
Analyze exchange data, market indexes, and community sentiment from CoinMarketCap—powered by AI. This sub-agent provides access to exchange listings, token holdings, metadata, and high-level metrics like the CMC 100 Index and the Fear & Greed Index. It’s designed for use within your larger CoinMarketCap AI Analyst system or as a standalone workflow. This agent can be triggered by a supervisor or manually used with message and sessionId inputs. Supported Tools (5 Total) 🔍 Exchange Map Get CoinMarketCap IDs, names, and slugs for exchanges (used as lookup before deeper queries). 🧾 Exchange Info Metadata including launch date, social links, country, and operational status. 💰 Exchange Assets Token balances, wallet addresses, and total USD value held by a specific exchange. 📈 CoinMarketCap 100 Index Constituents and weights of the CMC 100 Index, updated live. 😱 Fear & Greed Index Market sentiment score updated daily, ranging from Extreme Fear to Extreme Greed. What You Can Do with This Agent 🔹 Map exchanges to retrieve their ID and slug 🔹 Analyze exchange holdings by token and blockchain 🔹 Pull metadata for major CEXs like Binance or Coinbase 🔹 Compare global sentiment using the Fear & Greed Index 🔹 Access index data to understand CMC’s top 100 crypto asset breakdown Example Queries You Can Use ✅ "What is the latest Fear and Greed Index reading?" ✅ "Get a list of all exchanges on CoinMarketCap." ✅ "What tokens are held by Binance?" ✅ "Retrieve metadata for Coinbase." ✅ "Show me the top assets in the CMC 100 Index." Agent Architecture AI Brain**: GPT-4o-mini Memory**: Window buffer memory using sessionId Tools**: 5 API-connected nodes Trigger**: External input via message and sessionId Setup Instructions Get a CoinMarketCap API Key Apply here: https://coinmarketcap.com/api/ Configure n8n Credentials Use HTTP Header Auth to store your CoinMarketCap API key. Optional: Trigger from a Supervisor Connect to a parent agent using Execute Workflow with message and sessionId inputs. Test Sample Prompts “Get all exchanges”, “Fetch CMC index”, “Show Binance token holdings” Sticky Notes Included Exchange & Community Guide – Explains agent purpose and component connections Usage & Examples – Walkthrough for sample use cases Error Handling & Licensing – Includes API error code reference and licensing details ✅ Final Notes This agent is part of a broader CoinMarketCap AI Analyst System. Visit my Creator profile to download all available sub-agents and supervisor flows. Understand exchange behavior and community sentiment—automated with AI and CoinMarketCap.
by Don Jayamaha Jr
Meet your AI-powered crypto data analyst—fully integrated with CoinMarketCap APIs. This workflow acts as the supervisor agent for a multi-agent architecture built in n8n, connecting three powerful sub-agents to extract real-time insights from centralized and decentralized markets. It’s the ultimate tool for crypto traders, analysts, developers, and researchers who need strategic multi-source intelligence—all through Telegram. This workflow requires 3 sub-agent templates to function correctly. See below. 🔌 Required Sub-Workflows (Install First) CoinMarketCap Crypto Agent Tool → Token prices, metadata, conversions, listings CoinMarketCap Exchange & Community Agent Tool → Exchange info, token holdings, Fear & Greed index CoinMarketCap DEXScan Agent Tool → DEX trading pairs, liquidity, OHLCV data Download all from my Creator Profile: https://n8n.io/creators/don-the-gem-dealer/ What Makes This Workflow Special? This is not just another API wrapper—it’s an intelligent routing agent powered by GPT-4o-mini, capable of: Understanding complex user queries Choosing the appropriate tool workflow Structuring the API request Executing sub-workflows Formatting the output Returning insights via Telegram It connects three domains of market data: Cryptocurrencies (CEX)** Exchanges & Sentiment** DEX trading data** 🔍 What You Can Do 💰 Token Intelligence Get token metadata, price, volume, supply Compare rankings and conversions 🏦 Exchange Insights View assets held by exchanges Track the CMC 100 Index and Fear & Greed Score 🌐 DEX Market Analysis Analyze pair quotes, historical OHLCV, live trades Discover the top DEXs by volume across blockchains ✅ Example Questions to Ask “What’s the market cap of Ethereum today?” “Show liquidity and volume for SOL/USDT on Solana” “Get token holdings for Binance” “Compare BTC price on Uniswap vs Binance” “What’s the Fear & Greed index right now?” 🛠️ Setup Instructions Create Telegram Bot Use @BotFather to get your bot token. Get CoinMarketCap API Key Apply here: https://coinmarketcap.com/api/ Install Sub-Agent Templates Required: Crypto Agent Tool Exchange & Community Tool DEXScan Tool Configure Credentials in n8n Add both Telegram and CoinMarketCap keys as HTTP Header Auth. Deploy & Test Ask your Telegram bot: “Top 10 tokens by 24h volume” or “Convert 5 ETH to USD” Workflow Architecture AI Brain**: GPT-4o-mini Memory**: Windowed buffer memory via sessionId Tool Agents**: toolWorkflow() → routes requests to the appropriate sub-agent Executes real-time API queries and returns structured output Included Sticky Notes System Overview** Error Handling Guide (200, 400, 401, 429, 500)** Step-by-Step Usage Instructions** Prompt Examples + API Docs** Legal & Licensing Notes** Your crypto insights—smarter, faster, and all in one Telegram message.
by Don Jayamaha Jr
Access real-time cryptocurrency prices, market rankings, metadata, and global stats—powered by GPT-4o and CoinMarketCap! This modular AI-powered agent is part of a broader CoinMarketCap multi-agent system designed for crypto analysts, traders, and developers. It uses the CoinMarketCap API and intelligently routes queries to the correct tool using AI. This agent can be used standalone or triggered by a supervisor AI agent for multi-agent orchestration. Supported API Tools (6 Total) This agent intelligently selects from the following tools to answer your crypto-related questions: 🔍 Tool Summary Crypto Map – Lookup CoinMarketCap IDs and active coins Crypto Info – Get metadata, whitepapers, and social links Crypto Listings – Ranked coins by market cap CoinMarketCap Price – Live prices, volume, and supply Global Metrics – Total market cap, BTC dominance Price Conversion – Convert between crypto and fiat What You Can Do with This Agent 🔹 Get live prices and volume for tokens (e.g., BTC, ETH, SOL) 🔹 Convert crypto → fiat or fiat → crypto instantly 🔹 Retrieve whitepapers, logos, and website links for any token 🔹 Analyze total market cap, BTC dominance, and circulating supply 🔹 Discover new tokens and track their CoinMarketCap IDs 🔹 View the top 100 coins ranked by market cap or volume Example Queries ✅ "What is the CoinMarketCap ID for PEPE?" ✅ "Show me the top 10 cryptocurrencies by market cap." ✅ "Convert 5 ETH to USD." ✅ "What’s the 24h volume for ADA?" ✅ "Get the global market cap and BTC dominance." AI Architecture AI Brain**: GPT-4o-mini Memory**: Session buffer with sessionId Agent Type**: Subworkflow AI tool Connected APIs**: 6 CoinMarketCap endpoints Trigger Mode**: Executes when called by a supervisor (via message and sessionId inputs) Setup Instructions Get a CoinMarketCap API Key Register here: https://coinmarketcap.com/api/ Configure Credentials in n8n Use HTTP Header Auth with your API key for each connected endpoint Connect This Agent to a Supervisor Workflow (Optional) Trigger this agent using Execute Workflow with inputs message and sessionId Test Prompts Try asking: “Convert 1000 DOGE to BTC” or “Top 5 coins in EUR” Included Sticky Notes Crypto Agent Guide – Agent overview, node map, and endpoint details Usage Instructions – Step-by-step usage and sample prompts Error Handling & Licensing – Troubleshooting and IP rights ✅ Final Notes This agent is part of the CoinMarketCap AI Analyst System, which includes multiple specialized agents for cryptocurrencies, exchanges, community data, and DEX insights. Visit my Creator profile to find the full suite of tools. Get smarter about crypto—analyze the market in real time with AI and CoinMarketCap.
by Don Jayamaha Jr
Gain full visibility into decentralized exchanges using CoinMarketCap’s DEXScan API—powered by AI. This workflow is part of the CoinMarketCap AI Analyst system and delivers real-time and historical insights on spot trading pairs, DEX liquidity, trading activity, and OHLCV data across chains like Ethereum, Polygon, Solana, and more. Use this workflow as a sub-agent triggered by a parent supervisor workflow, or run it manually with inputs sessionId and message. 🔧 Supported Tools (8 Total) DEX Metadata → Static info (name, launch date, logo, URLs) DEX Networks List → All supported DEX chains + network metadata DEX Listings Quotes → Ranked list of DEXs with live trading volume, market share DEX Pair Quotes (Latest) → Real-time liquidity, price, and buy/sell stats DEX OHLCV Historical → Time-series data (daily/hourly/1m) DEX OHLCV Latest → Today’s price, volume, open/close for pairs DEX Trades Latest → Up to 100 recent trades for any DEX pair DEX Spot Pairs Latest → Active token pairs across DEXs + filters (volume, liquidity, volatility) Agent Architecture AI Model**: gpt-4o-mini Context Memory**: Window buffer using sessionId Trigger Input**: message, sessionId Execution**: Via Execute Workflow or parent AI supervisor Design**: Tool-based LangChain agent with CMC DEXScan endpoints 💡 Use Cases 🔹 Find top DEXs by 24h volume 🔹 Get spot pairs with highest liquidity on a specific network 🔹 Track historical OHLCV for Uniswap pairs 🔹 View latest trades for SOL/USDC pool 🔹 Analyze tax, pooled % and holders for specific pairs 🔹 Filter pairs by 24h volume, percent change, liquidity, or number of transactions ✅ Example Queries ✅ "Top 5 DEXs by 24h volume on Ethereum" ✅ "Get historical OHLCV for SOL-USDC on Solana" ✅ "Latest trades for a PancakeSwap pair" ✅ "Show all spot pairs with over $500K in liquidity on Polygon" ✅ "Retrieve metadata for Uniswap and SushiSwap" 🛠️ Setup Instructions Get a CoinMarketCap API Key Sign up at: https://coinmarketcap.com/api/ Add API Key to Credentials in n8n Use HTTP Header Auth method Trigger from Parent Workflow (Optional) Use Execute Workflow and pass message and sessionId Test Prompt Ideas Try: "Compare liquidity of Uniswap and Curve pairs on Ethereum" Sticky Notes Included DEXScan Agent Guide – Workflow architecture + supported tools Usage & API Call Examples – Prompts, test inputs, setup flow Error Codes + Licensing – 400/401/429/500 troubleshooting, IP rights ✅ Final Notes This agent is part of the CoinMarketCap AI Analyst System, which includes multiple specialized agents for cryptocurrencies, exchanges, and community data. Visit my Creator profile to find the full suite of tools. Master DEX analytics with AI—get powerful liquidity, trading, and pair insights in seconds.
by irfan saeed
Auto-Generate YouTube Chapters with AI-Powered Transcript Analysis Overview This workflow uses YouTube Data API v3 and Google Gemini 1.5 Flash AI to automatically generate timestamped chapters for videos by analyzing SRT captions. It enhances viewer navigation, improves SEO , and saves creators time by automating manual tasks. Prerequisites YouTube API Setup Create a Google Cloud Project Go to the Google Cloud Console. Click Select a project > New Project and name it (e.g., "YouTube Chapters Automation") . Enable YouTube Data API v3 Navigate to APIs & Services > Library. Search for "YouTube Data API v3" and click Enable . Configure OAuth Consent Screen Go to APIs & Services > OAuth consent screen. Select External (public) or Internal (testing), then add required details (app name, support email) . Generate OAuth 2.0 Credentials Under Credentials, click Create Credentials > OAuth client ID. Choose Web app, then download the JSON key file . Add Credentials to n8n Other Requirements Google Gemini API**: Configure access for the gemini-1.5-flash-8b-exp-0924 model by getting the api key. Workflow Steps Set Video ID Input the target video ID (e.g., r1wqsrW2vmE) using the Set Video ID node. Fetch Video Metadata Use the YouTube API node to retrieve the video’s title, category, and existing description . Download SRT Captions Get Caption ID: Call https://www.googleapis.com/youtube/v3/captions to fetch the caption track ID . Download Transcript: Use the ID to retrieve SRT data via https://www.googleapis.com/youtube/v3/captions/{{ID}}?tfmt=srt . Analyze Transcript with Gemini AI Process the SRT file with Google Gemini AI to identify chapters using a prompt like: "Classify this transcript into timestamped chapters (e.g., 00:00 - Introduction)." Validate output with a structured parser (e.g., Structured Captions node) . Update Video Description Append chapters to the description using the YouTube API’s videos.update method . Value Proposition Viewer Experience**: Chapters improve navigation and reduce drop-off rates . SEO Benefits**: Structured descriptions enhance search visibility . Time Savings**: Eliminates manual chapter creation .
by Sagar
This template streamlines your Instagram content posting workflow by connecting Google Drive for image storage, using OpenAI for AI-generated captions, and leveraging Facebook Graph API for automated publishing. Pre-requisites Before setting up this workflow, ensure you have: A Google account with access to Google Drive An OpenAI API key for AI caption generation A Facebook Developer account with Instagram Graph API access An Instagram Business or Creator account connected to a Facebook Page n8n.io account with workflow access Setup Instructions Configure Data Source Create a Google Sheet with the following columns: Name: Filename of your image in Google Drive Caption: Optional custom caption (leave empty for AI-generated captions) URL: your Video Reel or Image in Google Drive Connect Google Drive Add your Google Drive credentials in the "Google Drive" node Specify the folder path where your Instagram image/Video are stored Configure the node to retrieve image files based on filenames from your Google Sheet Set Up OpenAI Integration Add your OpenAI API key to the credentials Configure the OpenAI node to generate engaging captions based on image content Adjust temperature and model parameters for desired creativity level Configure Facebook Graph API Connect your Facebook account with Instagram access Set up the Facebook Graph API node to post to your Instagram Business/Creator account Ensure proper image formatting (1:1, 4:5, or 16:9 aspect ratios supported by Instagram) Workflow Automation Setup Configure the scheduler node to run at your preferred frequency Set up error handling to notify you of any posting failures Add conditional nodes to use either custom or AI-generated captions Execution Instructions After completing all connections, test the workflow with a single image Monitor the execution in the n8n dashboard to ensure proper functioning View the "Executions" tab to track successful posts and troubleshoot any errors Adjust posting frequency and scheduling as needed This template saves hours of manual Instagram posting work while maintaining an authentic presence. Perfect for social media managers, content creators, and businesses looking to maintain consistent Instagram activity without the daily manual effort. The workflow handles image retrieval, caption generation or customization, proper Instagram API formatting, scheduled posting, and execution tracking - all in one automated solution.
by Łukasz
Who is it for? If you are getting a lot of emails into your Gmail inbox, then probably some of those can be solved easly by replying or by doing specific short tasks. But analyzing whole email thread content just to catch up with multiple threads can be very wasteful. So by using AI you can actually get simple propositions of what should be done before closing this specific email and actual proposed answer to that email. This is especially useful if you need to do some actions before replying to email. In that case you can simply assign task to specific person, await until it's done, copy-paste AI answer when it's done, and close. Another good use would be if on one inbox there are working multiple people. It can make the process much more streamlined. How It Works? Script runs on your selected trigger. If you are using section "Read and Star", then you may use "Email Trigger". Automation is looking for exiting open Todoist tasks, that have the same title as email If task does not exist, then we are asking AI to analyze thread and give output that is Todoist-API-ready: having summary of email content having proposed actions to be taken having proposed answer to this email If email was unstarred for some reason but task was not closed, then task is being closed automatically. Script FOR PURPOSE is not trying to unstar messagess which have closed tasks, because this could lead to some inconsistencies. How to set up? Select and setup your triggers, depending on your needs Setup connections using N8N instructions. You will need: Gmail Todoist AI (in this workflow OpenAI is used) (Optional) Remove "Read and Star" section if you don't want tasks automatically read and starred. (Optional) Adjust AI node - especially useful if you want to use different model or have response in different language NOTE Chat does not heave memory attached on purpose. The purpose is that it should analyze each inbox message separately, not in thread. When using memory, it can get lost easily. NOTE2 You might want to adjust limits on nodes "Get Unread From Inbox", "Get Starred From Inbox" and "Get Open Tasks", especially if having issues with model complying to output structure. And that's it. I hope that this automation will make your Gmail <-> Todoist process much more streamlined! What's More? There is actually more that you could do with this automation, but it really depends on your needs. For example, you could add Form trigger to handle incoming support requests. Another thing is that you could replace Todoist with Asana or any database (like NocoDB) if you are using it for your task management.
by Ranjan Dailata
Who this is for? This workflow is designed for professionals and teams who need real-time, structured insights from Google Search results without manual effort. What problem is this workflow solving? This n8n workflow solves the problem of automating Google Search result extraction, cleanup, summarization, and AI-enhanced formatting for downstream use like sending the results to a webhook or another system. What this workflow does Automates Google Search via Bright Data Uses Bright Data’s proxy-based SERP API to run a Google Search query programmatically. Makes the process repeatable and scriptable with different search terms and regions/zones. Cleans and Extracts Useful Content The Google Search Data Extractor uses LLM based cleaning to remove HTML/CSS/JS from the response and extract pure text data. Converts messy, unstructured web content into structured, machine-readable format. Summarizes Search Results Through the Gemini Flash + Summarization Chain, it generates a concise summary of the search results. Ideal for users who don’t have time to read full pages of search results. Formats Data Using AI Agent The AI Agent acts like a virtual assistant that: Understands search results Formats them in a readable, JSON-compatible form Prepares them for webhook delivery Delivers Results to Webhook Sends the final summary + structured search result to a webhook (could be your app, a Slack bot, Google Sheets, or CRM). Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. A Google Gemini API key (or access through Vertex AI or proxy). Update the Google Search query as you wish by navigating to the Set Google Search Query node. Update the Webhook HTTP Request node with the Webhook endpoint of your choice. How to customize This Workflow to your needs 1. Change the Search Input Default: It searches a fixed query or dataset. Customize: Accept input from a Google Sheet, Airtable, or a form. Auto-trigger searches based on keywords or schedules. 2. Customize Summarization Style (LLM Output) Default: General summary using Google Gemini or OpenAI. Customize: Add tone: formal, casual, technical, executive-summary, etc. Focus on specific sections: pricing, competitors, FAQs, etc. Translate the summaries into multiple languages. Add bullet points, pros/cons, or insight tags. 3.Choose Where the Results Go Options: Email, Slack, Notion, Airtable, Google Docs, or a dashboard. Auto-create content drafts for WordPress or newsletters. Feed into CRM notes or attach to Salesforce leads.
by Ranjan Dailata
Who this is for? This workflow automates the process of Wikipedia data extraction using the Bright Data Web Unlocker, parsing and cleaning the data, and then sending the results to a specified webhook URL for downstream processing, reporting, or integration. What problem is this workflow solving? Researchers who need structured information from Wikipedia pages regularly. Data Engineers building knowledge bases or enriching datasets with factual data. Digital Marketers or Content Writers automating fact-checking or content sourcing. Automation Enthusiasts who want to trigger external systems with rich context from Wikipedia. What this workflow does This workflow addresses the challenges of manually retrieving, structuring, and using data from Wikipedia at scale. Workflow Breakdown Trigger Type: Scheduled or Manual Purpose: Starts the workflow either on a fixed schedule (e.g., daily) or on-demand via a manual trigger or incoming webhook. Bright Data Wikipedia Scraping Tool Used: Bright Data Web Unlocker Action: Scrape the HTML content of one or multiple Wikipedia article URLs. Parse & Extract Structured Data The Basic LLM Chain node is responsible for producing a human readable content. Summarization Summarize the Wikipedia content by utilizing the Summarization Chain node. Send to Webhook Initiates a Webhook notification to the specified URL as part of the "Summary Webhook Notifier" node. Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. In n8n, configure the Google Gemini(PaLM) Api account with the Google Gemini API key (or access through Vertex AI or proxy). Update the Set Wikipedia URL with Bright Data Zone node with the Wikipedia URL and Bright Data Zone. Update the Summary Webhook Notifier node with the Webhook endpoint of your choice. How to customize this workflow to your needs Update Wikipedia URL Replace with your own Wikipedia URL of your interest. Make sure to set the Wikipedia URL as part of the "Set Wikipedia URL with Bright Data Zone" node. Modify Data Extraction Logic Extract entire article content or just specific sections by extending the "LLM Data Extractor" node prompt. Extend AI Summarization Extract key bullet points or entities. Create short-form summaries by extending the "Concise Summary Generator" node. Extend Summary Webhook Notifier Send to Slack, Discord, Telegram, MS Teams via the Webhook notification mechanism. Connect to your internal database/API via the Webhook notification mechanism.
by merfy
Use Case Manually extracting images from PDF files for analysis is often slow and inefficient. Many users resort to taking screenshots of each page, uploading them to an AI tool like OpenAI for image analysis, and then manually copying the insights into a document. This manual process is time-consuming and prone to errors. This workflow streamlines the entire process by automatically extracting images from a PDF, analyzing them using the GPT-4o model, and saving the results in seconds—eliminating the need for manual effort. What This Workflow Does Extracts all images from the uploaded PDF file automatically The workflow scans each page of the PDF and identifies embedded images without manual intervention. Uses the GPT-4o model to analyze each extracted image Each image is processed through GPT-4o to generate descriptive insights, summaries, or context-specific analysis depending on the use case. Saves the analysis results to a .txt file, including image URLs The final output is a plain text file containing both the image URLs (e.g., hosted on cloud storage) and the corresponding GPT-4o analysis, ready for further use or sharing. Setup 1.Set up your credentials when you first open the workflow. You’ll need accounts for OpenAI, Convert API, and Google Drive. 2.Convert API does not rate-limit your API, sometimes you may receive 503 service unavailable error. Nevertheless, it doesn’t mean that you cannot convert your file. It simply means that you should retry the conversion in a few seconds. 3.Upload a PDF with images to Google Drive. 4.Remove unnecessary parts and retrieve image-related information. 5.Integrate image and image analysis information together. 6.Analyze each image using the OPENAI GPT-4o model. 7.Retrieve all image analysis content and image URL 8.Integrate multiple image URLs and analysis content 9.Output content to a .txt file. Template was created in n8n v1.83.2 How to Customize Replace the manual trigger with a Google Drive trigger or other automation triggers Change the image analysis model (e.g., switch or fine-tune GPT-4o) Send the results to other platforms (e.g., Slack, Telegram, LINE, etc.) instead of saving to a .txt file