by Oneclick AI Squad
This guide walks you through setting up an AI-driven workflow to automate flight and hotel reservation processes using a conversational travel booking system. The workflow accepts booking requests, processes them via APIs, and sends confirmations, enabling a seamless travel booking experience. What’s the Goal? Automatically accept and process booking requests for flights and hotels via HTTP POST. Use AI to understand natural language requests and route them to appropriate data processors. Search for flights and hotels using external APIs and process booking confirmations. Send confirmation emails and return structured booking data to users. Enable an automated system for efficient travel reservations. By the end, you’ll have a self-running system that handles travel bookings effortlessly. Why Does It Matter? Manual booking processes are time-consuming and prone to errors. This workflow offers: Zero Human Error**: AI ensures accurate request parsing and booking processing. Time-Saving Automation**: Automates the entire booking lifecycle, boosting efficiency. Seamless Confirmation**: Sends automated emails and responses without manual intervention. Enhanced User Experience**: Provides a conversational interface for bookings. Think of it as your reliable travel booking assistant that keeps the process smooth and efficient. How It Works Here’s the step-by-step flow of the automation: Step 1: Trigger the Workflow Webhook Trigger**: Accepts incoming booking requests via HTTP POST, initiating the workflow. Step 2: Parse the Request AI Request Parser**: Uses AI to understand natural language booking requests (e.g., flight or hotel) and extracts relevant details. Step 3: Route Booking Type Booking Type Router**: Determines whether the request is for a flight or hotel and routes it to the respective data processor. Step 4: Process Flight Data Flight Data Processor**: Handles flight-specific data and prepares it for the search API. Step 5: Search Flight API Flight Search API**: Searches for available flights based on parameters (e.g., https://api.aviationstack.com) and returns results. Step 6: Process Hotel Data Hotel Data Processor**: Handles hotel-specific data and prepares it for the search API. Step 7: Search Hotel API Hotel Search API**: Searches for available hotels based on parameters (e.g., https://api.booking.com) and returns results. Step 8: Process Flight Booking Flight Booking Processor**: Processes flight bookings and generates confirmation details. Step 9: Process Hotel Booking Hotel Booking Processor**: Processes hotel bookings and generates confirmation details. Step 10: Generate Confirmation Message Confirmation Message Generator**: Creates structured confirmation messages for the user. Step 11: Send Confirmation Email Send Confirmation Email**: Sends booking confirmation via email to the user. Step 12: Send Response Send Response**: Returns structured booking data to the user, completing the workflow. How to Use the Workflow? Importing the workflow in n8n is a straightforward process. Follow these steps to import the Conversational Travel Booker workflow: Download the Workflow: Obtain the workflow file (e.g., JSON export from n8n). Open n8n: Log in to your n8n instance. Import Workflow: Navigate to the workflows section, click "Import," and upload the workflow file. Configure Nodes: Adjust settings (e.g., API keys, webhook URLs) as needed. Execute Workflow: Test and activate the workflow to start processing bookings. Requirements n8n account and instance setup. Access to flight and hotel search APIs (e.g., Aviationstack, Booking.com). Email service integration for sending confirmations. Webhook URL for receiving booking requests. Customizing this Workflow Modify the AI Request Parser to handle additional languages or booking types. Update API endpoints in Flight Search API and Hotel Search API nodes to match your preferred providers. Adjust the Send Confirmation Email node to include custom email templates or additional recipients. Schedule the Webhook Trigger to align with your business hours or demand peaks.
by Davide
📩🤖 This workflow automatically processes emails received in Gmail, extracts their attachments, and organizes them into specific folders in Google Drive based on the sender's email address. Note: The workflow avoids duplicates by checking folder existence before creation. Benefits: ✅ Automated Organization: No need to manually sort or download email attachments. 📁 Sender-based Categorization: Files are stored in clearly labeled folders per sender, improving traceability and reducing clutter. ⏱ Time-saving: Reduces repetitive administrative tasks by automating the workflow end-to-end. 🔁 Modular and Scalable: Can be easily extended or reused with other services (e.g., Dropbox, S3) or integrated into larger document workflows. 🔐 Secure Cloud Storage: Attachments are safely backed up in Google Drive, minimizing the risk of data loss from email. How It Works Trigger: The workflow can be triggered manually ("When clicking ‘Execute workflow’) or automatically (via Gmail Trigger polling emails every minute). Email Processing: Fetches emails (with attachments) from Gmail within a date range (default: July 6–9, 2025). For each email, checks if it contains attachments (via IF node). Folder Management: Searches Google Drive for a folder named after the sender’s email address (under parent folder "Email Attachments"). Creates the folder if it doesn’t exist. Attachment Handling: Splits out binary attachments, extracts filenames, and uploads each file to the sender’s dedicated folder in Google Drive. Sub-Workflow Execution: Uses Execute Workflow to modularize the upload process (reusable for other workflows). Set Up Steps Google Services: Connect Gmail and Google Drive nodes to your accounts via OAuth2. Ensure the parent folder "Email Attachments" (ID: 1EitwWVd5rKZTlvOreB4R-6xxxxxx) exists in Google Drive. Adjust Date Range: Modify receivedAfter/receivedBefore in the Get emails node to target specific emails. Test: Run manually to verify folder creation and attachment uploads. Activate Automation: Enable the Gmail Trigger for real-time processing (currently active: false). Need help customizing? Contact me for consulting and support or add me on Linkedin.
by ConvertAPI
Who is this for? For developers and organizations that need to convert PPTX files to PDF. What problem is this workflow solving? The file format conversion problem. What this workflow does Downloads the PPTX file from the web. Converts the PPTX file to PDF. Stores the PDF file in the local file system. How to customize this workflow to your needs Open the HTTP Request node. Adjust the URL parameter (all endpoints can be found here). Add your secret to the Query Auth account parameter. Please create a ConvertAPI account to get an authentication secret. Optionally, additional Body Parameters can be added for the converter.
by Vuong Nguyen
How it works This workflow generates an 8-second product advertising video from a single input image. It downloads the image from Google Drive, converts it to base64 for the API request, analyzes it with Gemini (Creative Visualiser), then turns the description into a short video script/prompt. The prompt + image are sent to Veo to start a long-running video generation job. The workflow polls until a video URI is available, downloads the MP4, and uploads it back to Google Drive. Setup 1) Connect credentials used in this workflow: Google Drive + Google Gemini, and an API key for the Veo HTTP requests. 2) Set the input image file in Download ad image. 3) Set the output folder in Upload to Drive. 4) (Optional) Adjust aspectRatio, resolution, and durationSeconds in Generate Video, then execute the workflow.
by Sk developer
YouTube Transcript Summarization in Any Language for Social Media This n8n workflow automates the process of: Retrieving YouTube Video Transcripts: It fetches the transcript for any YouTube video URL provided using the YouTube Transcript API from RapidAPI. Generating a Concise Summary in Any Language: The workflow uses Google Gemini (PaLM) to create a concise summary of the transcript in the language specified by the user (e.g., English, Spanish, etc.). Storing the Summary in Google Docs: The generated summary is inserted into a predefined Google Document, making it easy for users to share or edit. Features: Language Flexibility:** Summaries are created in the desired language. Fully Automated:** From fetching the transcript to updating Google Docs, the process is fully automated. Social Media Ready:** The summary is formatted and stored in a Google Doc, ready for use in social media posts. This workflow integrates with YouTube Transcript API via RapidAPI, allowing you to easily fetch video transcripts and summarize them with AI. The entire process is automated and seamless. Powered by RapidAPI: API Used:* YouTube Transcript API via *RapidAPI** to get the transcript data. Benefits: Saves Time:** Automates the transcript summarization process, eliminating the need for manual content extraction and summarization. Customizable Language Support:** Provides summaries in any language, enabling accessibility and engagement for a global audience. Streamlined Content Creation:** Automatically generates concise, engaging summaries that are ready for social media use. Google Docs Integration:** Saves summaries directly into a Google Doc for easy sharing, editing, and content management. Challenges Addressed: Manual Transcript Extraction:** Problem: Manually transcribing and summarizing YouTube videos for social media can be time-consuming and error-prone. Solution: This workflow fully automates the process, saving hours of manual work using the YouTube Transcript API. Lack of Language Support in Summaries:** Problem: Many automated tools only summarize content in a single language, limiting their accessibility. Solution: With language flexibility, the workflow creates summaries in the language of your choice, helping you cater to diverse audiences. Inconsistent Video Quality & Transcript Accuracy:** Problem: Not all YouTube videos have well-structured or accurate transcripts, leading to incomplete or inaccurate summaries. Solution: The workflow can process and format even imperfect transcripts, ensuring that the generated summaries are still accurate and useful. Managing Content Across Platforms:** Problem: Transcripts and summaries often need to be stored in multiple locations for social media posts, which can be cumbersome. Solution: The workflow integrates with Google Docs to automatically store and manage summaries in one place, making it easier to share and reuse content.
by Harshil Agrawal
This workflow allows you to send position updates of the ISS every minute to a topic in MQTT using the MQTT node. Cron node: The Cron node will trigger the workflow every minute. HTTP Request node: This node will make a GET request to the API https://api.wheretheiss.at/v1/satellites/25544/positions to fetch the position of the ISS. This information gets passed on to the next node in the workflow. Set node: We will use the Set node to ensure that only the data that we set in this node gets passed on to the next nodes in the workflow. AWS SQS: This node will send the data from the previous node to the iss-position topic. If you have created a topic with a different one, you can use that topic instead.
by Alex Halfborg
BACKGROUND Malaysia's Inland Revenue (LHDN) provides an API to get the tax id number for a business entity, based on a given Business Registration number (BRN or SSM), or NRIC (MyKad). PROBLEM However, the API only allows one search at a time. SOLUTION This free workflow lets you do a batch search to get TIN for multiple SSM or NRIC. This is useful if you need to prepare your internal DB for e-invoicing PRE-REQUISITES 1) Get your connection client id and client secret from myhasil.gov.my website 2) Prepare your Google Sheet containing a list of SSM and NRIC you want to get the TIN 3) Create N8N credential to connect to your google sheet above SUPPORT Questions? Ask alex at halfborg dot com
by Markhah
Overview This n8n workflow is a modular AI analyst system that provides real-time insights from CoinMarketCap’s centralized and decentralized data sources. Using GPT-based AI, the system interprets natural language questions about the crypto market and delegates them to specialized agent workflows. It supports Telegram chat input and returns structured results such as coin quotes, DEX liquidity, exchange info, and community sentiment—all integrated from the CoinMarketCap API ecosystem. Prerequisites a. OpenAI or Gemini account (via GPT-4o-mini or equivalent LLM). b. Telegram Bot API token (for message input/output). c. Valid CoinMarketCap API key. 📦 Required subflows: CoinMarketCap_Crypto_Agent_Tool CoinMarketCap_Exchange_and_Community_Agent_Tool CoinMarketCap_DEXScan_Agent_Tool d. All tools must be installed and configured before use. Each one acts as a specialized endpoint wrapper for CoinMarketCap APIs. How It Works Telegram Input Users send a query to the bot (e.g. “Top DEX pairs on Ethereum”). Session Memory & Agent Brain Session is tracked via chat.id GPT-4o-mini interprets the query, routes to sub-agents Sub-Agent Workflows Crypto Agent: prices, rankings, conversions Exchange Agent: community sentiment, exchange token holdings DEX Agent: OHLCV data, liquidity pools, trades Multi-Agent Coordination AI can combine queries across tools (e.g., get token ID → fetch quote → analyze liquidity) Ensures valid parameters and avoids API errors Telegram Output Final analysis is sent back to the user as a formatted message. Troubleshooting Tips Error Code Meaning Fix 400 Bad request Check symbol/slug/ID validity 401 Unauthorized Verify CoinMarketCap API key 429 Rate limit exceeded Throttle or upgrade API tier 500 Server error Retry with backoff or report to CMC Example Telegram Queries “Show me top 5 coins by market cap” “Get price of ETH on Uniswap and Binance” “How much liquidity is in SOL-USDC pair?” “Fear & Greed Index and trending tokens” SEO Tags (ẩn hoặc ghi chú riêng): coinmarketcap, n8n crypto analyst, crypto ai telegram bot, dex liquidity, CMC price tracker, gpt-4o crypto market, token sentiment dashboard, fear and greed index
by Nazmy
Bearer Token Validation This n8n template helps you manage and validate tokens easily using: n8n as your backend workflow engine Airtable as your lightweight token store 🚀 What It Does Stores user tokens securely in Airtable with expiry or usage metadata. Validates incoming tokens in your workflows (e.g., webhook APIs). Rejects invalid or expired tokens automatically for security. Can be extended to generate, rotate, or revoke tokens for user management. How It Works Webhook node receives requests with a Bearer header. Airtable Query looks up the provided token. Validation Logic (Code node): Checks if the token exists. Verifies expiry or usage limits if configured. Returns success if valid, or error if error with describing the issue. Note: This is the simplest way to do auth, just for simplification Why Use This No need for a full backend to manage secure token validation. Clean, modular, and ready for your SaaS workflows. Enjoy building secure automations with n8n + Airtable! 🚀 Built by: Nazmy
by Meak
Google Maps Email Scraper System Most lead generation tools charge $2–$5 per lead and lock you into expensive subscriptions. This workflow lets you scrape unlimited business emails from Google Maps for free — no paid APIs required. Benefits Zero API costs – scrape data directly from Google Maps Unlimited leads – extract thousands of emails per day Geographic targeting – search by city, region, or business type Complete automation – from search to clean email list Built-in data cleaning – removes duplicates & invalid entries How It Works Reads search queries from a Google Sheet (e.g., "Calgary dentist") Sends HTTP requests to Google Maps and scrapes business listings Extracts website URLs with custom JavaScript regex Visits each site, scrapes HTML, and finds email addresses Cleans and validates data Exports organized lead list back to Google Sheets Who Is This For B2B sales teams generating leads for outreach Marketing agencies building client lead databases Local businesses researching competitors & partners Real estate professionals analyzing target neighborhoods Franchise developers scouting new markets Setup Create a Google Sheet with two tabs: “searches” & “emails” Add search queries to the “searches” tab (one per row) Connect Google Sheets OAuth credentials in n8n Configure HTTP request nodes with SSL ignore enabled Add custom JavaScript regex code for URL and email extraction ROI & Monetization $0 per lead vs. $2–$5 from paid tools Generate 1,000+ leads per day without hitting API limits Sell lead lists or offer as a $500–$2,000 per niche/location service Perfect upsell for agencies offering outreach or local SEO Strategy Insights In the YouTube walkthrough, I show how to: Write custom JavaScript + regex for clean URL extraction Build a robust loop system with error handling & rate limiting Avoid IP blocking with batching & delays Sell lead generation as a high-margin recurring service Automate outreach to monetize the leads you scrape Check Out My Channel For more advanced AI automation systems that generate real business results, check out my YouTube channel where I share the exact strategies I use to build automation agencies, sell high-value services, and scale to $20k+ monthly revenue.
by Simeon
Google Calendar MCP – Context-Aware Calendar Operations This n8n template implements an MCP (Model Context Protocol)-compliant module for managing Google Calendar events in a context-aware, conflict-free manner. 🧠 What It Does This MCP enables structured interaction with Google Calendar based on context and intent, ensuring reliable, reusable operations with awareness of existing data and state. ✅ Core Capabilities Context-aware event creation** Prevents overlapping by validating time availability before creating new events. Gap validation** Checks if a time range is busy or free, enabling smarter scheduling decisions. Conditional updates** Only updates events after confirming their existence and current state. Safe deletion** Removes events using MCP principles of validation and traceability. 🚀 How to Use To use this MCP in your context-aware systems: Deploy the template in your n8n instance. Locate the Server node in the workflow — it exposes a Server-Sent Events (SSE) URL. Copy that SSE URL. Use that URL as the entry point for your MCP client or orchestrator. This URL acts as the communication bridge, allowing you to interact with the MCP-compliant Google Calendar logic using standard MCP semantics.
by Julien DEL RIO
This template is inspired by Save your workflows into a GitHub repository by hikerspath and Back Up Your n8n Workflows To Github by jon-n8n. Basic Retrieve all workflows from an n8n instance and save it on a gitlab project. If the workflow exist, il will only save the changes. Flow What the workflow does : Sets custom parameters Gets workflows Iterates through each workflow one by one Get the file from Gitlab if exists Compare the files as objects (not as strings) Return a status on the workflow Create, Edit or ignore the file depending on the status Return a list of status for each workflow Configuration Select a credential in each Gitlab nodes. Edit the data in node "Globals" : repo.owner : slug of the user or team owning the repo repo.name : slug of the repository repo.branch : branch to commit on repo.path : from root of the repository. Should end with / Comments Error on gitlab nodes will not stop the run but will list the current workflow as error in the results Some fields are ignored to determined if there is changes : updatedAt : should be ignored if only ignores fields are changed globals : it's running information, no need to follow the changes