by noda
🧩 What this template does This workflow builds a 120-minute local date course around your starting point by querying Google Places for nearby spots, selecting the top candidates, fetching real-time weather data, letting an AI generate a matching emoji, and drafting a friendly itinerary summary with an LLM in both English and Japanese. It then posts the full bilingual plan with a walking route link and weather emoji to Slack. 👥 Who it’s for Makers and teams who want a plug-and-play bilingual local itinerary generator with weather awareness — no custom code required. ⚙️ How it works Trigger – Manual (or schedule/webhook). Discovery – Google Places nearby search within a configurable radius. Selection – Rank by rating and pick the top 3. Weather – Fetch current weather (via OpenWeatherMap). Emoji – Use an AI model to match the weather with an emoji 🌤️. Planning – An LLM writes the itinerary in Markdown (JP + EN). Route – Compose a Google Maps walking route URL. Share – Post the bilingual itinerary, route link, and weather emoji to Slack. 🧰 Requirements n8n (Cloud or self-hosted) Google Maps Platform (Places API) OpenWeatherMap API key Slack Bot (chat:write) LLM provider (e.g., OpenRouter or DeepL for translation) 🚀 Setup (quick) Open Set → Fields: Config and fill in coords/radius/time limit. Connect Credentials for Google, OpenWeatherMap, Slack, and your LLM. Test the workflow and confirm the bilingual plan + weather emoji appear in Slack. 🛠 Customize Adjust ranking filters (type, min rating). Modify translation settings (target language or tone). Change output layout (side-by-side vs separated). Tune emoji logic or travel mode. Add error handling, retries, or logging for production use.
by MANISH KUMAR
Shopify Digital Product Automation (from just and image to complete Shopify product page.) This Shopify Digital Product Automation is an advanced n8n-powered workflow that leverages AI (Google Gemini), Airtable, and Shopify API to generate product details from images and automatically post them to Shopify. It fully automates the process — from uploading images to publishing Shopify products — with minimal manual effort. 💡 Key Advantages Our Shopify Digital Product Automation offers five core advantages: 🔗 Shopify Product Sync — Automatically posts product details including title, description, SEO fields, and matched category to Shopify. ✍️ AI-Powered Product Generation — Gemini analyzes uploaded images and generates engaging, SEO-friendly product titles, descriptions, and metadata. 🗂️ Structured Output — Outputs JSON-ready product data compatible with Shopify, ensuring smooth automation. 📄 Airtable Integration — Tracks uploaded images, analyzed data, and generated products to prevent duplication. 📤 End-to-End Automation — Handles the complete workflow from image upload to Shopify posting without manual intervention. ⚙️ How It Works The workflow follows a step-by-step automated process: Step-by-Step Process Upload Images – Add your digital artwork or poster images to Google Drive and record them in Airtable. Image Analysis – AI fetches new images and analyzes visual elements like characters, series, poster text, and style. Store Analysis Results – Updates analyzed data in Airtable and marks images as Used. Fetch Shopify Collections – Retrieves current collections from your Shopify store for category matching. Generate Product Details – Gemini generates product title, description, matched category, SEO page title, meta description, and URL handle. Save Generated Products – Stores generated product details in Airtable and marks them as generated. Post Products to Shopify – Automatically creates new products in Shopify using the API. Update Status – Marks products as posted in Airtable after successful posting. 🛠️ Setup Steps Required Node Configuration To implement this workflow, configure the following n8n nodes: Trigger Node** – Start workflow manually or via scheduler. Airtable Node** – Fetch raw images and store processed product details. Google Drive Node** – Access image files. HTTP Request Node** – Fetch Shopify collections. Code Node** – Refine AI outputs and format product data. Split & Limit Nodes** – Process images and products in batches. LangChain / Gemini Node** – Generate product titles, descriptions, and SEO data. Shopify Node** – Create products via Shopify API. Status Update Node** – Update Airtable with processing and posting status. 🔐 Credentials Required Before running the workflow, ensure you have the following credentials configured: Shopify Access Token** – For posting products and fetching collections. Gemini API Key** – For AI-powered product generation. Airtable API Key** – For storing and tracking workflow data. Google Drive OAuth** – To access image files. 👤 Ideal For This automation workflow is designed for: Shopify store owners managing hundreds of digital products Ecommerce teams automating product listings Marketing teams needing scalable, AI-driven product content workflows 💬 Bonus Tip The workflow is fully modular and customizable. You can extend it to: Automatically assign prices or discounts Multi-language product description generation Social media promotion of new products Email campaign integration All extensions can be implemented within the same n8n flow, making it a complete digital product automation solution.
by Cheng Siong Chin
Introduction This workflow connects to OpenAI, Anthropic, and Groq, processing requests in parallel with automatic performance metrics. Ideal for testing speed, cost, and quality across models. How It Works Webhooks trigger parameter extraction and routing. Three AI agents run simultaneously with memory and parsing. Responses merge with detailed metrics. Workflow Template Webhook → Extract Parameters → Router ├→ OpenAI Agent ├→ Anthropic Agent ├→ Groq Agent → Merge → Metrics → Respond Workflow Steps Webhook receives POST with prompt and settings. Parameters extracted and validated. Router directs by cost, latency, or type. AI agents run in parallel. Results merged with metadata. Metrics compute time, cost, and quality. Response returns outputs and recommendation. Setup Instructions Activate Webhook with authentication. Add API keys for all providers. Define models, tokens, and temperature. Adjust Router logic for selection. Tune Metrics scoring formulas. Prerequisites n8n v1.0+ instance API keys: OpenAI, Anthropic, Groq HTTP client for testing Customization Add providers like Gemini or Azure OpenAI. Enable routing by cost or performance. Benefits Auto-select efficient providers and compare model performance in real time.
by Muhammad Faaz Khan
This workflow automatically researches a lead’s website, understands what the business actually does, and generates a highly personalized cold outreach subject line and icebreaker using AI. Instead of guessing or relying on shallow placeholders, it scrapes real website content, summarizes it intelligently, and feeds that context into an LLM to produce outreach that feels relevant and human. If a website is broken or unreachable, the workflow safely flags it so you can identify faulty leads early. Sticky Notes No API keys are hardcoded. Add all credentials via n8n’s credential manager. Website scraping uses markdown and strict character limits to keep token usage under control. Unresponsive websites return no content, helping you identify broken or invalid leads. The database can be swapped easily. Baserow is used here, but Airtable or Google Sheets can be used with minimal changes. How It Works 1. Fetch Leads from Database The workflow starts by pulling leads from Baserow, including company name and website URL. 2. Visit and Scrape the Website The lead’s website is fetched and converted into HTML. If the site fails to load or respond, the workflow records no content and continues without breaking. 3. Extract and Filter Internal Links All links are extracted from the page, then filtered so only links belonging to the same website are kept. 4. Scrape Multiple Pages The workflow scrapes up to five pages in total, including the main website page and up to four internal pages. This provides enough context while avoiding unnecessary data. 5. Convert to Markdown and Trim Content Each page is converted to markdown to reduce token usage and trimmed to a maximum of 5,000 characters to control LLM costs. 6. Aggregate Website Content All processed markdown content is combined into a single structured input. 7. Generate a Business Overview An LLM analyzes the aggregated content and generates a concise overview of the company and its offering. 8. Generate Subject Line and Icebreaker A second LLM uses the company name, lead name where available, and the generated business overview to create a highly personalized subject line and icebreaker for outreach. 9. Update the Database The final outputs are written back to the database, keeping each lead enriched and ready for outreach. Use Cases Cold email personalization at scale Lead research and enrichment Sales and SDR workflows Agency outreach systems Founder-led outbound without manual research Requirements Apify API key for website scraping Baserow credentials (can be replaced with Airtable or Google Sheets) LLM API key (OpenRouter is used in this workflow) Why This Template Is Useful Most outreach fails because it is generic. This workflow solves that by grounding every message in real website content while staying fast, efficient, and cost-conscious.
by Daniel Shashko
This workflow automates the process of monitoring multiple RSS feeds, intelligently identifying new articles, maintaining a record of processed content, and delivering timely notifications to a designated Slack channel. It leverages AI to ensure only truly new and relevant articles are dispatched, preventing duplicate alerts and information overload. 🚀 Main Use Cases Automated News Aggregation:** Continuously monitor industry news, competitor updates, or specific topics from various RSS feeds. 📈 Content Curation:** Filter and deliver only new, unprocessed articles to a team or personal Slack channel. 🎯 Duplicate Prevention:** Maintain a persistent record of seen articles to avoid redundant notifications. 🛡️ Enhanced Information Delivery:** Provide a streamlined and intelligent way to stay updated without manual checking. 📧 How it works The workflow operates in distinct, interconnected phases to ensure efficient and intelligent article delivery: 1. RSS Feed Data Acquisition 📥 Initiation:** The workflow is manually triggered to begin the process. 🖱️ RSS Link Retrieval:** It connects to a Baserow database to fetch a list of configured RSS feed URLs. 🔗 Individual Feed Processing:** Each RSS feed URL is then processed independently. 🔄 Content Fetching & Parsing:** An HTTP Request node downloads the raw XML content of each RSS feed, which is then parsed into a structured JSON format for easy manipulation. 📄➡️🌳 2. Historical Data Management 📚 Seen Articles Retrieval:** Concurrently, the workflow queries another Baserow table to retrieve a comprehensive list of article GUIDs or links that have been previously processed and notified. This forms the basis for duplicate detection. 🔍 3. Intelligent Article Filtering with AI 🧠 Data Structuring for AI:** A Code node prepares the newly fetched articles and the list of already-seen articles into a specific JSON structure required by the AI Agent. 🏗️ AI-Powered Filtering:** An AI Agent, powered by an OpenAI Chat Model and supported by a Simple Memory component, receives this structured data. It is precisely prompted to compare the new articles against the historical "seen" list and return only those articles that are genuinely new and unprocessed. 🤖 Output Validation:** A Structured Output Parser ensures that the AI Agent's response adheres to a predefined JSON schema, guaranteeing data integrity for subsequent steps. ✅ JSON Cleaning:** A final Code node takes the AI's raw JSON string output, parses it, and formats it into individual n8n items, ready for notification and storage. 🧹 4. Notification & Record Keeping 🔔 Persistent Record:** For each newly identified article, its link is saved to the Baserow "seen products" table, marking it as processed and preventing future duplicate notifications. 💾 Slack Notification:** The details of the new article (title, content, link) are then formatted and sent as a rich message to a specified Slack channel, providing real-time updates. 💬 Summary Flow: Manual Trigger → RSS Link Retrieval (Baserow) → HTTP Request → XML Parsing | Seen Articles Retrieval (Baserow) → Data Structuring (Code) → AI-Powered Filtering (AI Agent, OpenAI, Memory, Parser) → JSON Cleaning (Code) → Save Seen Articles (Baserow) → Slack Notification 🎉 Benefits: Fully Automated:** Eliminates manual checking of RSS feeds and Slack notifications. ⏱️ Intelligent Filtering:** Leverages AI to accurately identify and deliver only new content, avoiding duplicates. 💡 Centralized Data Management:** Utilizes Baserow for robust storage of RSS feed configurations and processed article history. 🗄️ Real-time Alerts:** Delivers timely updates directly to your team or personal Slack channel. ⚡ Scalable & Customizable:** Easily adaptable to monitor various RSS feeds and integrate with different Baserow tables and Slack channels. ⚙️ Setup Requirements: Baserow API Key:** Required for accessing and updating your Baserow databases. 🔑 OpenAI API Key:** Necessary for the AI Agent to function. 🤖 Slack Credentials:** Either a Slack OAuth token (recommended for full features) or a Webhook URL for sending messages. 🗣️ Baserow Table Configuration:** A table with an rssLink column to store your RSS feed URLs. A table with a Nom column to store the links of processed articles. For any questions or further assistance, feel free to connect with me on LinkedIn: https://www.linkedin.com/in/daniel-shashko/
by Amirul Hakimi
Advanced AI Lead Enrichment & Cold Email Personalization with n8n, Airtable, Apify, and LLMs Automated B2B Lead Nurturing: Hyper-Personalization for High-Converting Cold Email Campaigns This powerful n8n automation workflow is designed to execute advanced B2B lead enrichment and hyper-personalization for cold email outreach. By orchestrating a complex chain of data scraping, AI analysis (via LLMs/GPT-4.1), and CRM synchronization (using Airtable), this workflow ensures every lead receives a highly tailored and relevant outreach message, maximizing conversion rates and minimizing manual effort. Workflow Execution & Key Features Airtable Trigger & Lead Qualification: The workflow is triggered by an Airtable webhook, pulling a new lead record (including name, email, and company URLs). Email Validation* is performed using *NeverBounce** to filter out invalid contacts. Initial Lead Filtering screens for key demographic criteria (e.g., US: Yes or No? and target Headcount: >5, <30?). Only qualified B2B leads proceed, ensuring optimal resource allocation. Deep Web & Social Scraping (Apify Integration): LinkedIn Company Scraper* and a *LinkedIn Profile Scraper* (via *Apify**) extract raw data from the lead's company and personal profiles. Company Homepage Scraper** pulls the main website content for analysis. Scrape Personal LinkedIn Posts** node retrieves recent activity for the ultimate personalization hook. AI-Powered Data Synthesis & Variable Determination: Multiple OpenAI (GPT-4.1-mini/4.1) nodes analyze and structure the raw, cleaned text (Remove HTML nodes ensure clean inputs). Determine Valuable URLs** uses an LLM to smartly categorize and select key company pages (e.g., ==/about==, ==/solutions==, ==/case-studies==) for deeper scraping. Analyze Company/Mission, Analyze Offerings & Positioning, Analyze Process & Differentiation, and **Analyze Proof of Success nodes create factual, structured business summaries for the ultimate ICP research. Determine Variables* nodes create *pre-written, personalized cold email variables** (==company_specialty==, ==ICPofLead==, ==PainPointLeadSolves==, etc.) for different outreach strategies. LinkedIn Post Personalization: An LLM (Craft Opening Line - Posts) analyzes recent LinkedIn activity to generate a hyper-specific, conversation-starting opener (e.g., "Saw your LinkedIn post about..."). Conditional logic (Posts Available?) determines whether to use the post-based opener or fall back to the standard, company-based personalization. CRM Update & Campaign Launch (Instantly.ai): Finalized, enriched lead data and the crafted personalization variables are synchronized back to the Airtable CRM for record-keeping and lead status updates (Update Lead W/ Enrichment). The lead is then seamlessly pushed to the Instantly.ai outbound platform, injecting the AI-generated custom variables directly into the cold email sequence for mass deployment. This blueprint automates the tedious, high-effort task of prospect research and personalization, providing a scalable lead generation solution that increases both outreach quality and sales velocity. Stop sending generic emails—start leveraging AI automation today.
by Club de Inteligencia Artificial Politécnico CIAP
🤖 Interactive Academic Chatbot (Telegram + MongoDB) Overview 📋 This project is a template for building a complete academic virtual assistant using n8n. It connects to Telegram, answers frequently asked questions by querying MongoDB, keeps the community informed about key dates (via web scraping), and collects user feedback for continuous improvement. How It Works Architecture and Workflow ⚙️ n8n: Orchestration of 3 workflows (chatbot, scraping worker, announcer). Telegram: Frontend for user interaction and sending announcements. MongoDB: Centralized database for FAQs, academic calendar, and feedback logs. Web Scraping: HTTP Request and HTML Extract nodes to read the university's web calendar. Cron: For automatic periodic executions (daily and weekly). Core Processes 🧠 Real-time reception of user queries via Telegram. Querying MongoDB collections for FAQ answers and calendar dates. Daily scraping of the university website to keep the calendar updated. Instant logging of user feedback (👍/👎) in MongoDB. Proactive sending of weekly announcements to the Telegram channel. Key Benefits ✅ Complete automation of student communication 24/7. An always-accurate academic calendar database without manual intervention. A built-in continuous improvement system through user feedback. Proactive communication of important events to the entire community. Use Cases 💼 Automation of student support in universities, colleges, and institutions. A virtual assistant for any organization needing to manage FAQs and a dynamic calendar. An automated announcements channel to keep a community informed. Requirements 👨💻 n8n instance (self-hosted or cloud). Credentials for a Telegram Bot (obtained from @BotFather). Credentials for a MongoDB database (Connection URI). URL of the academic calendar to be scraped. Authors 👥 Doménica Amores Nicole Guevara Adrián Villamar Mentor: Jaren Pazmiño Applicants to the CIAP Polytechnic Artificial Intelligence Club
by 小林幸一
Generate personalized sales emails from Google Maps results to Google Sheets This workflow automates the entire process of lead generation and personalized outreach drafting for local businesses. It utilizes Apify to scrape business data from Google Maps based on your search criteria (e.g., "Cafes in Shibuya"), visits each business's website to extract content, and uses OpenAI to generate a highly personalized sales email that connects the business's unique characteristics with your service's value proposition. Finally, it saves the business details, scraped data, and the generated email draft into Google Sheets. This template is perfect for reducing the manual effort required to research leads and write initial cold outreach emails. Who is this for Sales Representatives** looking to automate lead sourcing and initial drafting. Marketing Agencies** doing outreach for local businesses. Freelancers** offering web design, SEO, or reservation system services to brick-and-mortar stores. What it does Configuration: You define your search query (e.g., "Gyms in London"), the number of leads to fetch, and details about the service you are selling. Lead Scraping: The workflow triggers an Apify actor (Google Maps Scraper) to find businesses matching your criteria. Website Analysis: It checks if the business has a website, fetches the HTML, and extracts relevant text to understand the business's vibe and offerings. AI Email Generation: OpenAI analyzes the scraped website text and generates a specific, personalized email subject and body promoting your service. Data Storage: All data (Business Name, Phone, Address, Website, Scraped Info, and Email Draft) is appended to a Google Sheet. Requirements n8n** (v1.0 or later) Apify Account**: You need an Apify account and the compass/google-maps-scraper actor. OpenAI Account**: An API key for generating the email content. Google Cloud Platform**: A project with the Google Sheets API enabled. How to set up Credentials: Set up your credentials for Apify, OpenAI, and Google Sheets in n8n. Google Sheet: Create a new Google Sheet and add the following headers in the first row: 店舗名 (Store Name) 住所 (Address) Webサイト (Website) 電話番号 (Phone Number) サイトから取得した情報 (Info from Website) 生成されたメール件名 (Generated Subject) 生成されたメール本文 (Generated Body) Workflow Configuration Node: Open the first "Workflow Configuration" node and update the following values: searchQuery: The location and keyword you want to target. serviceName: The name of the product you are selling. serviceStrength: The USP (Unique Selling Proposition) of your product. Google Sheets Node: Open the "Save to Google Sheets" node and select the file you created in step 2. How to customize Change the Prompt**: Open the "Generate Personalized Email" (OpenAI) node to modify the system prompt. You can change the tone, language (currently set to Japanese context in your example), or structure of the sales email. Filter Results**: You can add logic to the "Check Website URL Exists" node to filter out specific types of businesses or domains. Limit Scraping**: Adjust the maxPlaces value in the Configuration node to control how many leads you process per run to save on API credits.
by Yoshino Haruki
Who’s it for This template is ideal for busy professionals, students, or anyone with a dynamic schedule who wants to optimize their brief periods of free time. If you frequently find yourself with unexpected gaps between meetings and wish for intelligent, personalized suggestions on where to grab a coffee or get some work done, this workflow is for you. How it works / What it does The workflow begins by checking your Google Calendar for your next event at a scheduled time. It then calculates the travel time from your current location to your next event's venue using Google Maps. This allows it to determine your actual "gap time" – the usable free time before you need to start moving. If you have a sufficient gap (e.g., 30 minutes or more), the workflow fetches your preferred cafe criteria from a Google Sheet and searches for nearby cafes using Google Places. An AI agent then processes this information, along with your available gap time, to recommend the best cafe suited to your needs. This recommendation, complete with ratings and a Google Maps link, is sent directly to your Slack channel. Conversely, if the gap time is too short to comfortably visit a cafe, the workflow sends an urgent Slack alert, reminding you to prepare for your next appointment and providing essential details. How to set up Import the Workflow: Import this workflow into your n8n instance. Configure API Keys: In the "Workflow Configuration" node, replace the placeholders for googleMapsApiKey and googlePlacesApiKey with your actual API keys. Ensure these keys have access to the Google Maps Distance Matrix API and Google Places API (Nearby Search). Also, update currentLocation with your default or most frequent starting location (latitude/longitude or address). Google Calendar Credentials: Authenticate the "Get Next Calendar Event" node with your Google Calendar account. Select the calendar you wish to monitor. Google Sheets Credentials: Authenticate the "Get User Preferences" node with your Google Sheets account. Create a Google Sheet to store your cafe preferences (e.g., "Likes quiet places", "Prefers espresso", "Needs Wi-Fi"). Update the "Document ID" and "Sheet Name" in this node to point to your preference sheet. OpenRouter Credentials: Authenticate the "OpenRouter Chat Model" with your OpenRouter API key. Slack Credentials: Authenticate both "Send Slack Notification" and "Send Urgent Move Alert (Slack)" nodes with your Slack account. In both Slack nodes, update the channelId to your desired Slack channel where you want to receive notifications (e.g., #general, or a specific DM channel). Activate the Workflow: Once all configurations are complete, activate the workflow. Requirements An n8n instance (self-hosted or cloud). Google Account with Google Calendar and Google Sheets. Google Cloud Project with activated Google Maps Platform APIs (Distance Matrix API, Places API) and corresponding API Keys. An OpenRouter API Key. A Slack Workspace and API Token (or Webhook URL). How to customize the workflow Scheduling: Adjust the "Schedule Trigger" node to run at different intervals or specific times that best suit your daily routine. Minimum Gap Time: Modify the minimumGapMinutes variable in the "Workflow Configuration" node to set a different threshold for cafe recommendations. Cafe Search Radius: In the "Search Nearby Cafes (Google Places API)" node, you can change the radius parameter to search for cafes within a larger or smaller area. User Preferences: Expand your Google Sheet with more detailed preferences to give the AI agent better context for recommendations (e.g., "vegan options," "good for meetings," "strong coffee"). AI Prompt: Refine the prompt in the "AI Agent" node to guide the AI towards specific types of recommendations or output formats. Slack Message Customization: Edit the text fields in the Slack nodes to personalize the notification messages.
by Itunu
CoinMarketCap Token Discovery (Free API) Automatically discover cryptocurrency tokens from CoinMarketCap, clean the data, enrich it with official websites, and store the results in your preferred database or sheet. This workflow is designed to be safe for free API usage, easy to understand, and ready for extension. What This Workflow Does This workflow runs on a schedule and: Randomly selects pages from CoinMarketCap listings Fetches token data using the free CoinMarketCap API Cleans and normalizes token fields Enriches each token with official website data Processes tokens in safe batches with delays Outputs clean, structured token records Optionally saves results to a database or sheet Who This Is For This workflow is useful if you are: Doing crypto research or market discovery Building token datasets Running crypto outreach or lead generation Learning how to work with APIs in n8n Looking for a clean, real-world n8n example No advanced n8n knowledge is required. Setup Instructions (Required) Follow these steps before running the workflow: 1. Get a CoinMarketCap API Key Create a free account on CoinMarketCap Generate an API key from your dashboard 2. Add Your API Key Open the HTTP Request nodes Add your API key to the request headers: X-CMC_PRO_API_KEY = YOUR_API_KEY 3. Connect Storage Replace the storage node with your preferred option: Google Sheets Airtable PostgreSQL Webhook Add your own credentials before running the workflow. 4. Activate the Workflow Enable the workflow Let it run automatically based on the schedule How the Workflow Is Structured Trigger:** Runs every few days to avoid API limits Random Page Generator:** Prevents bias toward only top tokens Batch Processing:** Controls memory and request volume Delay Logic:** Keeps the workflow stable and API-friendly Cleaning Steps:** Removes messy or invalid data Final Output:** Clean, simple token records ready for use Output Example Each valid token produces a clean record like: Token name Symbol (ticker) Official website Source (CoinMarketCap) Timestamp Invalid or incomplete entries are automatically skipped. Customization Ideas You can easily extend this workflow to: Add social media scraping Track new tokens over time Trigger alerts for specific tokens Combine with other crypto APIs Feed data into outreach or analytics pipelines Important Notes This workflow uses CoinMarketCap’s free API tier Do not remove batch limits or delays unless you upgrade your API plan Replace sample storage with your own before production use License This workflow is provided for educational and practical use. You are free to modify and adapt it for your own projects. Author Built and shared by Itunu Ola n8n automation builder focused on practical, production-ready workflows.
by vinci-king-01
How it works This workflow automatically processes bank statements from various formats and extracts structured transaction data with intelligent categorization using AI. Key Steps File Upload - Accepts bank statements via webhook upload (PDF, Excel, CSV formats). Smart Format Detection - Automatically routes files to appropriate processors (PDF text extraction or spreadsheet parsing). AI-Powered Extraction - Uses GPT-4 to extract account details, transactions, and balances from statement data. Data Processing & Categorization - Cleans, validates, and automatically categorizes transactions into expense categories. Database Storage - Saves processed data to PostgreSQL database for analysis and reporting. API Response - Returns structured summary with transaction counts, expense totals, and category breakdowns. Set up steps Setup time: 8-12 minutes Configure OpenAI credentials - Add your OpenAI API key for AI-powered data extraction. Set up PostgreSQL database - Connect your PostgreSQL database and create the required table structure. Configure webhook endpoint - The workflow provides a /upload-statement endpoint for file uploads. Customize transaction categories - Modify the AI prompt to include your preferred expense categories. Test the workflow - Upload a sample bank statement to verify the extraction and categorization process. Set up database table - Ensure your PostgreSQL database has a bank_statements table with appropriate columns. Features Multi-format support**: PDF, Excel, CSV bank statements AI-powered extraction**: GPT-4 extracts account details and transactions Automatic categorization**: Expenses categorized as groceries, dining, gas, shopping, utilities, healthcare, entertainment, income, fees, or other Data validation**: Cleans and validates transaction data with error handling Database storage**: PostgreSQL integration for data persistence API responses**: Clean JSON responses with transaction summaries and category breakdowns Smart routing**: Automatic format detection and appropriate processing paths
by giangxai
Automated AI Video Avatar Workflow for Shorts using n8n & HeyGen Automatically generate AI avatar short videos and publish them to social platforms using n8n and HeyGen. This workflow turns viral ideas or prepared scripts into fully rendered AI avatar videos and handles publishing and tracking without manual editing. Who is this for? This template is ideal for: Content creators producing Shorts, Reels, or TikTok videos at scale Marketers using AI avatars for faceless or branded content Affiliate marketers running automated short-form content funnels AI automation builders combining LLMs, avatars, and no-code workflows Teams that want a repeatable system to generate and publish short videos If you want to automate AI avatar video creation instead of manually scripting, rendering, and uploading videos, this workflow is for you. What problem is this workflow solving? Creating short-form video content with AI avatars usually involves many manual steps and tools. Common challenges include: Manually writing scripts for each video Switching between AI tools for script generation and video rendering Waiting for avatar videos to render and checking status manually Uploading videos to social platforms one by one Tracking which ideas have been processed or published This workflow connects all steps into a single automated pipeline and removes repetitive manual work. What this workflow does This automated AI video avatar workflow: Triggers on a defined schedule Loads viral content ideas or pending rows from Google Sheets Analyzes content and generates optimized scripts using AI Creates AI avatar videos automatically via HeyGen Waits for video rendering and checks completion status Retrieves the final video once rendering is complete Publishes the video to social platforms Updates Google Sheets with publishing status and results The entire process runs end-to-end with minimal human intervention. Setup 1. Prepare Google Sheets Create a Google Sheet to manage your content pipeline with columns such as: idea / topic – Source idea or viral reference script – Generated or custom script (optional) status – pending / processing / published / error video_url – Final rendered video link publish_result – Publishing status or notes Only rows marked as pending will be processed by the workflow. 2. Connect Google Sheets Authenticate your Google Sheets account in n8n Select the spreadsheet in the content loading nodes Ensure update nodes can write back to the same sheet for status tracking 3. Configure AI & HeyGen Add credentials for your AI model (e.g. Gemini or OpenRouter) Add your HeyGen API credentials Configure avatar settings such as voice, language, and style Test video creation once before running the workflow at scale. 4. Configure Publishing & Schedule Set up publishing credentials for your target social platforms Open the Schedule trigger and define how often the workflow runs The schedule controls how frequently new AI avatar videos are created and published How to customize this workflow to your needs You can adapt this workflow without changing the core structure: Replace viral idea sources with your own content inputs Add approval steps (Slack, Telegram, Email) before video creation Customize scripts per platform or language Disable publishing and use the workflow only for video generation Add retry logic for failed renders or publishing steps Extend the workflow with analytics or performance tracking Best practices Start with a small batch of test rows Keep status values consistent in Google Sheets Use short, clear scripts optimized for AI avatars Monitor render and publish status nodes regularly Adjust schedule frequency based on rendering limits 📄 Documentation For a full walkthrough and advanced customization ideas, see Video Guide