by Jimleuk
This n8n workflow automates the process of parsing and extracting data from PDF invoices. With this workflow, accounts and finance people can realise huge time and cost savings in their busy schedules. Read the Blog: https://blog.n8n.io/how-to-extract-data-from-pdf-to-excel-spreadsheet-advance-parsing-with-n8n-io-and-llamaparse/ How it works This workflow will watch an email inbox for incoming invoices from suppliers It will download the attached PDFs and processing them through a third party service called LlamaParse. LlamaParse is specifically designed to handle and convert complex PDF data structures such as tables to markdown. Markdown is easily to process for LLM models and so the data extraction by our AI agent is more accurate and reliable. The workflow exports the extracted data from the AI agent to Google Sheets once the job complete. Requirements The criteria of the email trigger must be configured to capture emails with attachments. The gmail label "invoice synced" must be created before using this workflow. A LlamaIndex.ai account to use the LlamaParse service. An OpenAI account to use GPT for AI work. Google Sheets to save the output of the data extraction process although this can be replaced for whatever your needs. Customizing this workflow This workflow uses Gmail and Google Sheets but these can easily be swapped out for equivalent services such as Outlook and Excel. Not using Excel? Simple redirect the output of the AI agent to your accounting software of choice.
by Don Jayamaha Jr
Stay on top of the latest crypto news and market sentiment instantly, all inside Telegram! This workflow aggregates articles from the top crypto news sources, filters for your topic of interest, and summarizes key news and market sentiment using GPT-4o AI. Ideal for crypto traders, investors, analysts, and market watchers needing fast, intelligent news briefings. > 💬 Just type a coin name (e.g., "Bitcoin", "Solana", "DeFi") into your Telegram AI Agent—and get a smart news digest. How It Works Telegram Bot Trigger User sends a keyword (e.g., "Ethereum") of questions to the Telegram AI Agent. Keyword Extraction (AI-Powered) An AI agent identifies the main topic for better targeting. News Aggregation Pulls articles from 9 major crypto news RSS feeds: Cointelegraph Bitcoin Magazine CoinDesk Bitcoinist NewsBTC CryptoPotato 99Bitcoins CryptoBriefing Crypto.news Filtering Finds and merges articles relevant to the user's keyword. AI Summarization GPT-4o generates a 3-part summary: News Summary Market Sentiment Analysis List of Article Links Telegram Response Sends a structured, easy-to-read digest back to the user. 🔍 What You Can Do with This Workflow 🔹 Summarize breaking news for any crypto project or keyword 🔹 Monitor real-time market sentiment on Bitcoin, DeFi, NFTs, and more 🔹 Stay ahead of FUD, bullish trends, and major news events 🔹 Quickly brief yourself or your team via Telegram 🔹 Use it as a foundation for more advanced crypto alert bots ✅ Example User Inputs ✅ "Bitcoin" → Latest Bitcoin news and sentiment summary ✅ "Solana" → Updates on Solana projects, price movements, and community trends ✅ "NFT" → Aggregated news about NFT markets and launches ✅ "Layer 2" → Insights on Optimism, Arbitrum, and other L2s 🛠️ Setup Instructions Create a Telegram Bot Use @BotFather and obtain the Bot Token. Configure Telegram Credentials in n8n Add your bot token under Telegram API Credentials. Configure OpenAI API Add your OpenAI credentials for GPT-4o access. Update Telegram Send Node In the Telegram Send node, replace the placeholder chatId with your real Telegram user or group chat ID. Deploy and Test Start chatting with your bot: e.g., "Ethereum" or "DeFi". 📌 Workflow Highlights 9 major crypto news sources combined** Smart keyword matching** with AI query parsing Summarized insights** in human-readable format Reference links** included for deeper reading Instant delivery** via Telegram 🚀 Get ahead of the crypto market—automate your news and sentiment monitoring with AI inside Telegram!
by HoChien Chang
How it works This workflow automates the backup of all your n8n workflows to a specified Google Drive folder. It operates in two main phases: Orchestration (Scheduled Task): The workflow is initiated by a Schedule Trigger (e.g., daily at 1:30 AM by default). It then uses an n8n API Node to fetch a list of all existing workflows in your n8n instance. A Loop Over Items node processes each fetched workflow individually. For every workflow in the list, an Execute Workflow node calls the "worker" part of this same workflow, passing the individual workflow's data. Individual Workflow Backup (Worker Task): This part is triggered by the When Executed by Another Workflow node (called by the orchestrator part). It first retrieves the passed workflow data. A Parameters node defines the target Google Drive directory. The Get Google Drive File Data node searches the specified Google Drive folder to see if a backup file for the current workflow (named as WorkflowName_WorkflowID.json) already exists. An IF Node (ifDriveEmpty) checks the result: If a backup file exists, the workflow data is converted to a binary JSON file using a Code Node, and then the existing file on Google Drive is updated (Backup to Google Drive2). If a backup file does not exist, the workflow data is first formatted as JSON, converted to a binary JSON file using another Code Node, and then uploaded as a new file to Google Drive (Backup to Google Drive4). If any Google Drive upload/update operation fails, a Failure Email is sent. After the loop in the orchestration phase completes (all workflows processed), a Limit node ensures only one signal proceeds to send a Success Email and a Discord message indicating the overall backup process was completed. Setup Clone Workflow: Clone this workflow into your n8n environment. Credentials: n8n API: In the "Get all n8n Workflows" node, select or create n8n API credentials. This allows the workflow to list all your other workflows. Google Drive: In the "getDriveFileData", "Backup to Google Drive2", and "Backup to Google Drive4" nodes, select or create your Google Drive OAuth2 API credentials. Gmail: In the "successEmail" and "failureEmail" nodes, select or create your Gmail credentials. Discord (Optional): If you wish to use Discord notifications, configure your Discord Bot API credentials in the "Discord" node. Configuration: Schedule: Open the "Schedule Trigger" node and adjust the trigger interval (e.g., time of day) as needed. Google Drive Folder: Open the "Parameters" node (connected after "Workflow Data"). Edit the directory value to the full URL of the Google Drive folder where you want to store your backups (e.g., https://drive.google.com/drive/folders/YOUR_FOLDER_ID). Email Recipients: Open the "successEmail" and "failureEmail" nodes and update the "Send To" field with your desired recipient email address(es). Discord Channel (Optional): If using Discord, open the "Discord" node and set the "Channel ID" for notifications. Sub-Workflow ID: The "Execute Workflow" node is pre-configured to call this workflow itself using its ID. If you import this workflow and its ID changes, you may need to update the workflowId in the "Execute Workflow" node to the new ID of this imported workflow. However, n8n usually handles this if it's self-referential within the same imported workflow. How to use Activate: After completing the setup steps, activate the workflow. It will automatically run according to the defined schedule. Manual Execution: You can also manually trigger the workflow by clicking the "Execute Workflow" play button on the "Schedule Trigger" node to perform an immediate backup of all workflows. Check Backups: Your n8n workflows will be saved as .json files (named WorkflowName_WorkflowID.json) in the Google Drive folder you specified in the "Parameters" node. Notifications: You will receive an email (and optionally a Discord message) upon successful completion of the entire backup process, or individual failure emails if a specific workflow backup to Google Drive fails. Author & Credits Creator:** Hochien Chang YouTube Channel (Chinese):** HC AI說人話 YouTube Channel (English):** HC HumanizeAI Original Explanation Video:** https://youtu.be/PA15H5qunC0 Based on:** n8n Workflow Backup to Google Drive
by Joseph
Note: Now includes an Apify alternative for Rapid API (Some users can't create new accounts on Rapid API, so I have added an alternative for you. But immediately you are able to get access to Rapid API, please use that option, it returns more detailed data). *Scroll to bottom for APify setup guide* This n8n workflow automates LinkedIn lead generation, enrichment, and activity analysis using Apollo.io, RapidAPI, Google Sheets and Mail.so. Perfect for sales teams, founders, B2B marketers, and cold outreach pros who want personalized lead insights to drive better conversion rates. ⚙️ How This Workflow Works The workflow is broken down into several key steps, each designed to help you build and enrich a valuable list of LinkedIn leads: 1. 🔑 Lead Discovery (Keyword Search via Apollo) Pulls leads using Apollo.io's API based on keywords, industries, or job titles. Saves lead name, title, company, and LinkedIn URL to your Google Sheet. You can replace the trigger node from the form node to a webhook, whatsapp, telegram, etc, any way for you to send over your query variables over to initiate the workflow. 2. 🧠 Username Extraction (from LinkedIn URL) Extracts the LinkedIn username from profile URLs using a simple script node. This is required for further enrichment via RapidAPI. 3. ✉️ Email Lookup (via Apollo User ID) Uses the Apollo User ID to retrieve the lead’s verified work email. Ensures high-quality leads with reliable contact info. To double check that the email is currently valid, we use the mail.so api and filter out emails that fail deliverability and mx-record check. We don't wanna risk sending emails to no longer existent addresses, right? 4. 🧾 Profile Summary Enrichment (via RapidAPI) Queries the LinkedIn Data API to fetch a lead’s profile summary/bio. Gives you a deeper understanding of their background and expertise. 5. 📰 Recent Activity Collection (Posts & Reposts) Retrieves recent posts or reposts from each lead’s profile. Great for tailoring outreach with reference to what they’re currently talking about. 6. 🗂️ Leads Database Update All enriched data is written to the same Google Sheet. New columns are filled in without overwriting existing data. ✅ Smart Retry & Row Status Logic Every subworkflow includes a fail-safe mechanism to ensure: ✅ Each row has status columns (e.g., done, failed, pending). 🕒 A scheduled retry workflow resets failed rows to pending after 2 weeks (customizable). 💬 This gives failed enrichments another chance to be processed later, reducing data loss. 📋 Google Sheets Setup Template 1: Apollo Leads Scraping & Enrichment Template 2: Enriched Leads Database Make a copy to your Drive and use. Columns will be filled as each subworkflow runs (email, summary, interests, etc.) 🔐 Required API Keys To use this workflow, you’ll need the following credentials: 🧩 Apollo.io Sign up and get your key here: Apollo.io API Keys ⚠️ Important: Toggle the “Master API Key” option to ON when generating your key. This ensures the same key can be used for all Apollo endpoints in this workflow. 🌐 RapidAPI (LinkedIn Data API) Subscribe to the API here: LinkedIn Data API on RapidAPI Use the key in the x-rapidapi-key header in the relevant nodes. ✉️ Mail.so Sign up and get your key here: Mail.so API > 💡 For both APIs, set up the credentials in n8n as “Generic Credential” types. This way, you won’t need to reconfigure the headers in each node. 🛠️ Customization Options Modify the Apollo filters (location, industry, seniority) to target your ideal customers. Change retry interval in the scheduler (e.g., weekly instead of 2 weeks). Connect the database to your email campaign tool like Mailchimp or Instantly.ai. Replace the AI nodes with your desired AI agents and customize the system messages further to get desired results. 🆕 Apify Update Guide To use this workflow, you’ll need the following credentials: Login to Apify, then open this link; https://console.apify.com/actors/2SyF0bVxmgGr8IVCZ/ Click on integrations and scroll down to API Solutions and select "Use API endpoints". Scroll to "Run Actor synchronously and get dataset items" and copy the actor endpoint url then paste it in the placeholder inside the http node of Apify alternative flow "apify-actor-endpoint". That's it, you are set to go. I am available for custom n8n workflows, if you like my work, please get in touch with me on email at joseph@uppfy.com
by Davide
How It Works This workflow creates an AI-powered Telegram chatbot with session management, allowing users to: Start new conversations** (/new). Check current sessions** (/current). Resume past sessions** (/resume). Get summaries** (/summary). Ask questions** (/question). Key Components: Session Management**: Uses Google Sheets to track active/expired sessions (storing SESSION IDs and STATE). /new creates a session; /resume reactivates past ones. AI Processing**: OpenAI GPT-4 generates responses with contextual memory (via Simple Memory node). Summarization: Condenses past conversations when requested. Data Logging**: All interactions (prompts/responses) are saved to Google Sheets for audit and retrieval. User Interaction**: Telegram commands trigger specific actions (e.g., /question [query] fetches answers from session history). Main Advantages 1. Multi-session Handling Each user can create, manage, and switch between multiple sessions independently, perfect for organizing different conversations without confusion. 2. Persistent Memory Conversations are stored in Google Sheets, ensuring that chat history and session states are preserved even if the server or n8n instance restarts. 3. Commands for Full Control With intuitive commands like /new, /current, /resume, /summary, and /question, users can manage sessions easily without needing a web interface. 4. Smart Summarization and Q&A Thanks to OpenAI models, the workflow can summarize entire conversations or answer specific questions about past discussions, saving time and improving the chatbot’s usability. 5. Easy Setup and Scalability By using Google Sheets instead of a database, the workflow is easy to clone, modify, and deploy — ideal for quick prototyping or lightweight production uses. 6. Modular and Extensible Each logical block (new session, get current session, resume, summarize, ask question) is modular, making it easy to extend the workflow with additional features like analytics, personalized greetings, or integrations with CRM systems. Setup Steps Prerequisites: Telegram Bot Token**: Create via BotFather. Google Sheets**: Duplicate this template. Two sheets: Session (active/inactive sessions) and Database (chat logs). OpenAI API Key**: For GPT-4 responses. Configuration: Telegram Integration: Add your bot token to the Telegram Trigger and Telegram Send nodes. Google Sheets Setup: Authenticate the Google Sheets nodes with OAuth. Ensure sheet names (Session, Database) and column mappings match the template. OpenAI & Memory: Add your API key to the OpenAI Chat Model nodes. Adjust contextWindowLength in the Simple Memory node for conversation history length. Testing: Use Telegram commands to test: /new: Starts a session. /question [query]: Tests AI responses. /summary: Checks summarization. Deployment: Activate the workflow; the bot will respond to Telegram messages in real-time. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Max aka Mosheh
How it works • Automates multi-platform social media posting (Instagram, YouTube, TikTok, etc.) using AI-generated content • Integrates Airtable, n8n, and Blotato for full content scheduling and publishing • Supports both image and video uploads with dynamic text and account routing Set up steps • Takes ~15–30 minutes to set up depending on how many platforms you connect • Requires Airtable personal access token and Blotato API key • Uses sticky notes throughout the workflow to explain config, tokens, and troubleshooting clearly
by Harsh Maniya
✨ Automate Daily Hindu Festival Posts on X (Twitter) with AI 🐦 This workflow automates the entire process of creating and publishing culturally rich social media content about Hindu festivals. It starts by building a comprehensive festival calendar for the year in a Google Sheet, then runs daily to post engaging, bilingual updates on X (formerly Twitter). 🗓️ The workflow uses a sophisticated dual-AI system: Google Gemini acts as a content generator creating multiple post options ✍️, while OpenAI's GPT-4o Mini acts as a discerning social media manager, selecting the very best post for publication. 🧠 This ensures your content is not only automated but also high-quality and optimized for engagement. How it works ⚙️ This workflow operates in two distinct stages: Part 1: Data Population (One-Time Manual Run) 🔍 Fetch Festival Data: Manually trigger the workflow to scrape a list of 2025 Hindu festivals from a public calendar using the Jina AI Reader API. ✨ Enrich with AI: For each festival, a Google Gemini model researches and extracts key details: The festival's name in Hindi. A concise description of its significance in English. A Hindi translation of the description. 📝 Store in Google Sheets: The enriched data for the entire year is then systematically organized and saved in a designated Google Sheet, creating a content calendar. Part 2: Daily Automated Posting ⏰ Daily Trigger: A Schedule Trigger node activates the workflow every morning at 8 AM. ✔️ Check for Festivals: The workflow gets today's date and checks the Google Sheet to see if there is a corresponding festival. 🎨 Generate Post Options: If a festival is scheduled for the day, Google Gemini generates three distinct and engaging post options for X. Each post is crafted to be concise, use a mix of English and Hindi, and include relevant emojis and hashtags. 🏆 Select the Best Post: OpenAI's GPT-4o Mini then evaluates the three generated posts based on criteria like clarity, engagement potential, and effective use of language. It selects the single most impactful post. 🚀 Publish to X: The winning post is automatically published to your connected X account. Features ⭐ 🤖 Fully Automated Content Pipeline: From data collection to final publication, no manual intervention is needed after the initial setup. 🧠 Dual-AI System: Leverages Google Gemini for creative generation and OpenAI GPT-4o Mini for critical selection, ensuring high-quality output. 🗣️ Bilingual Content: Creates posts that blend English and Hindi to enhance cultural connection and broaden audience reach. 🎯 Dynamic and Contextual: Posts are automatically tailored to the specific festival of the day. 🗓️ Centralized Content Calendar: Uses Google Sheets as a reliable, easy-to-manage database for your yearly social media plan. Prerequisites 🛠️ Before you can use this workflow, you will need to: Have an n8n instance set up. Create a new, empty Google Sheet. Obtain credentials for the following services: Jina AI: Get a free Bearer Token from the Jina AI API page. Google: Set up Google credentials (OAuth2) for the Google Sheets and Google Gemini nodes. OpenAI: Get an API key from your OpenAI Platform dashboard. X (Twitter): Set up X credentials (OAuth2) to allow n8n to post on your behalf. How to use this template 🚀 🔑 Set up Credentials: In n8n, go to the "Credentials" section and add new credentials for Jina AI, Google (for both Sheets and Gemini), OpenAI, and X using the API keys and tokens you obtained. 📊 Configure the Google Sheet: Create a new Google Sheet. In the first row, create the following headers exactly as written: Name of the Festival Date English Description (Note the trailing space) Hindi Description Open the "Add all Rows at once" and "Fetch Data of Matched Date" nodes in the workflow and connect them to your Google account and select the Sheet you just created. ▶️ Populate the Data (Manual Step): Click the "Execute workflow" button on the When clicking ‘Execute workflow’ node. This will run the first part of the workflow, filling your Google Sheet with festival data for 2025. This only needs to be done once. ✅ Activate the Workflow: Save the workflow and then activate it using the toggle at the top right of the n8n canvas. The workflow will now run automatically every day to post about the day's festival. Extending the Workflow 💡 🖼️ Add Image Generation: Integrate a node like DALL-E or Midjourney to generate a unique image for each festival and include it in the tweet. 🌐 Cross-Platform Posting: Duplicate the final "Post to X" node and adapt it to post on other platforms like Facebook, LinkedIn, or Telegram. 🎨 Change the Tone: Modify the prompts in the "Generate Posts" and "Select Best Post" nodes to change the style of your social media content—make it more formal, humorous, or poetic. 📅 Use a Different Year: Update the URL in the "Get Festival Data" node to fetch data for a different year. The current URL is https://r.jina.ai/https://www.calendarlabs.com/2025-hindu-calendar.
by Joseph LePage
This workflow creates an automated system for monitoring and receiving notifications about new videos from your favorite YouTube channels through RSS feeds, with customizable email and Telegram notifications. 🌟 Key Features 📡 RSS Feed Management Accepts custom YouTube channel IDs or uses default channels Automatically creates RSS feeds for each YouTube channel Monitors channels for new video uploads Labels and filters recent videos within a 3-day window (change this as required) 📨 Multi-Channel Notification System Sends Telegram notifications with video thumbnails and links Delivers customized email notifications in two formats: Individual emails for each new video Single digest email containing all new videos ⚙️ Content Processing Fetches detailed video information using YouTube API Creates responsive HTML email templates with video previews Includes video thumbnails, titles, descriptions, and direct links Maintains professional formatting across different email clients 🛠️ Setup Requirements 🔑 API Configuration YouTube Data API credentials Gmail account for sending notifications Telegram bot token and chat ID OpenAI API key for content processing 📋 Channel Management Add YouTube channel IDs through form input Configure default channel list Set notification preferences Adjust monitoring schedule This workflow is perfect for content creators, marketers, or anyone wanting to stay updated with their favorite YouTube channels through automated, professionally formatted notifications delivered via email and Telegram.
by JPres
👥 Who Is This For? Content creators, marketing teams, and channel managers who need to streamline video publishing with optimized metadata and scheduled releases across multiple videos. 🛠 What Problem Does This Solve? Manual YouTube video publishing is time-consuming and often results in inconsistent descriptions, tags, and scheduling. This workflow fully automates: Extracting video transcripts via Apify for metadata generation Creating SEO-optimized descriptions and tags for each video Setting videos to private during initial upload (critical for scheduling) Implementing scheduled publishing at strategic times Maintaining consistent branding and formatting across all content 🔄 Node-by-Node Breakdown | Step | Node Purpose | |------|--------------| | 1 | Every Day (Scheduler) | Trigger workflow on a regular schedule | | 2 | Get Videos to Harmonize | Retrieve videos requiring metadata enhancement | | 3 | Get Video IDs (Unpublished) | Filter for videos that need publishing | | 4 | Loop over Video IDs | Process each video individually | | 5 | Get Video Data | Retrieve metadata for the current video | | 6 | Loop over Videos with Parameter IS | Set parameters for processing | | 7 | Set Videos to Private | Ensure videos are private (required for scheduling) | | 8 | Apify: Get Transcript | Extract video transcript via Apify | | 9 | Fetch Latest Videos | Get most recent channel content | | 10 | Loop Over Items | Process each video item | | 11 | Generate Description, Tags, etc. | Create optimized metadata from transcript | | 12 | AP Clean ID | Format identifiers | | 13 | Retrieve Generated Data | Collect the enhanced metadata | | 14 | Adjust Transcript Format | Format transcript for better processing | | 15 | Update Video's Metadata | Apply generated description and tags to video | ⚙️ Pre-conditions / Requirements n8n with YouTube API credentials configured Apify account with API access for transcript extraction YouTube channel with upload permissions Master templates for description formatting Videos must be initially set to private for scheduling to work ⚙️ Setup Instructions Import this workflow into your n8n instance. Configure YouTube API credentials with proper channel access. Set up Apify integration with appropriate actor for transcript extraction. Define scheduling parameters in the Every Day node. Configure description templates with placeholders for dynamic content. Set default tags and customize tag generation rules. Test with a single video before batch processing. 🎨 How to Customize Adjust prompt templates for description generation to match your brand voice. Modify tag selection algorithms based on your channel's SEO strategy. Create multiple publishing schedules for different content categories. Integrate with analytics tools to optimize publishing times. Add notification nodes to alert when videos are successfully scheduled. ⚠️ Important Notes Videos MUST be uploaded as private initially - the Publish At logic only works for private videos that haven't been published before. Publishing schedules require videos to remain private until their scheduled time. Transcript quality affects metadata generation results. Consider YouTube API quotas when scheduling large batches of videos. 🔐 Security and Privacy API credentials are stored securely within n8n. Transcripts are processed temporarily and not stored permanently. Webhook URLs should be protected to prevent unauthorized triggering. Access to the workflow should be limited to authorized team members only.
by PollupAI
This n8n workflow streamlines the onboarding process for new customers by automating personalized email communication, calendar scheduling, and contact assignment in HubSpot. It is perfect for businesses looking to ensure a smooth and personalized onboarding experience for new clients. 🧑💼 Who is this for? Customer success teams who need to onboard new clients efficiently. Sales teams who want to ensure smooth transitions from prospect to customer. Small businesses that want to automate customer onboarding without complex systems. 🧩 What problem is this workflow solving? This workflow reduces the manual effort involved in onboarding new customers by: Automatically sending personalized welcome emails. Scheduling a welcome meeting using a calendar tool. Assigning the customer to a Customer Success Manager (CSM) in HubSpot. ⚙️ What this workflow does Trigger via Webhook or HubSpot: The workflow can be triggered either by a webhook (direct API call) or a HubSpot trigger (e.g., when a new contact is created). HubSpot Connection: Retrieves the list of HubSpot owners (users with contact access). Identifies the owner of the new contact. Calendar Management: Utilizes a Calendar Agent to schedule a welcome meeting with the new customer. The Calendar Agent can create, update, or delete events as needed. Personalized Email Creation: Uses an AI-powered Email Writer (OpenAI) to generate a personalized welcome email. Transforms the email text into HTML for a polished format. Email Sending via Gmail: Sends the personalized email to the customer using Gmail. Sets the new contact’s owner in HubSpot for further communication tracking. 🛠️ Setup Webhook Setup in n8n: Create a new workflow and add a Webhook node. Set the Webhook URL path (e.g., /webhook-customer-onboarding). Make sure the workflow is active. Webhook Setup in HubSpot: Go to HubSpot Developer Account. Navigate to Settings > Integrations > Webhooks. Create a new webhook and set the URL as the n8n Webhook URL. Choose POST as the request method. Test the webhook to ensure it triggers the workflow in n8n. Calendar Agent Configuration: The Calendar Agent can be configured to create, update, or delete events. Connect it to your calendar tool (Google Calendar, Outlook, etc.). Customize the calendar event details (title, description, time). Email Writer Setup: Customize the AI prompt in the Email Writer node to match your brand’s voice. Adjust the email text format for your specific needs. Gmail Integration: Connect your Gmail account in n8n. Set the recipient email to the new customer’s email address. ✏️ How to customize this workflow to your needs Modify the AI-Powered Email**: Adjust the email prompt for the AI model to create a different welcome message. Change the email format or add custom variables (e.g., customer name, service details). Customize Calendar Settings**: Set default time slots for welcome meetings. Specify which calendar to use for scheduling. Add Additional Steps**: Extend the workflow to automatically assign the customer to a specific HubSpot list. Add a follow-up email or survey after the welcome meeting. This workflow is perfect for businesses seeking an efficient and personalized onboarding process, ensuring new customers feel welcomed and supported from day one.
by AlexWantMoreB
🚀 What this flow does • 🔎 Selects the least-used WordPress category (tracked in PostgreSQL) • 🤖 Uses GPT (4-mini or better) to generate a fully formatted SEO article with headings, TOC, lists, CTA, and Yoast blocks • 🖼️ Creates a placeholder cover image and uploads it to WordPress Media • 📬 Publishes the final post via /wp-json/wp/v2/posts with correct category + featured image • 🧠 Logs the used category for future rotation (zero duplicates!) ⚙️ Setup in 3 mins 🏷️ Add your WordPress domain with a simple Set node: domain=https://yourdomain.com 🔐 Create these 3 credentials in n8n: YOUR_WORDPRESS_CREDENTIAL — for /media, /posts YOUR_POSTGRES_CREDENTIAL — for category tracking YOUR_OPENAI_CREDENTIAL — GPT-4-mini or better 🧱 Run the SQL from docs to create the used_categories table ✅ Manually test first 3–5 nodes to check WP auth, OpenAI response, and DB connection 🕒 Then just schedule it and let the bot write for you. 🎯 Why it's awesome This is your personal AI content writer + publisher — perfect for: • 📰 SEO content farms • 📈 Affiliate blogs • 🧰 Micro niche sites • 🤫 PBNs with rotation-safe automation No more manual uploads, broken categories, or GPT spam. Every post is structured, beautiful, and intelligently categorized.
by Reza Gholizade
This n8n workflow template uses community nodes and is only compatible with the self-hosted version of n8n. Conversational Kubernetes Management with GPT-4o and MCP Integration This workflow enables you to manage Kubernetes clusters conversationally using OpenAI’s GPT-4o and a secure MCP (Model Context Protocol) server. It transforms natural language queries into actionable Kubernetes commands via a lightweight MCP API gateway — perfect for developers and platform engineers seeking to simplify cluster interaction. 🚀 Setup Instructions Import the Workflow Upload this template to your n8n instance. Configure Required Credentials OpenAI API Key: Add your GPT-4o API key in the credentials. MCP Client Node: Set the URL and auth for your MCP server. Test Kubernetes Access Ensure your MCP server is correctly configured and has access to the target Kubernetes cluster. 🧩 Prerequisites n8n version 0.240.0 or later Access to GPT-4o via OpenAI A running MCP server Kubernetes cluster credentials configured in your MCP backend ⚠️ Community Nodes Disclaimer This workflow uses custom community nodes (e.g., MCP Client). Make sure to review and trust these nodes before running in production. 🛠️ How It Works A webhook or chat input triggers the conversation. GPT-4o interprets the message and generates structured Kubernetes queries. MCP Client securely sends requests to your cluster. The result is returned and formatted for easy reading. 🔧 Customization Tips Tweak the GPT-4o prompt to match your tone or technical level. Extend MCP endpoints to support new Kubernetes actions. Add alerting or monitoring integrations (e.g., Slack, Prometheus). 🖼️ Template Screenshot 🧠 Example Prompts Show me all pods in the default namespace. Get logs for nginx pod in kube-system. List all deployments in staging. 📎 Additional Resources MCP Server on GitHub OpenAI Documentation n8n Docs Build smarter Kubernetes workflows with the power of AI !