by Oneclick AI Squad
This automated n8n workflow continuously monitors airline schedule changes by fetching real-time flight data, comparing it with stored schedules, and instantly notifying both internal teams and affected passengers through multiple communication channels. The system ensures stakeholders are immediately informed of any flight delays, cancellations, gate changes, or other critical updates. Good to Know Flight data accuracy depends on the aviation API provider's update frequency and reliability Critical notifications (cancellations, major delays) trigger immediate passenger alerts via SMS and email Internal Slack notifications keep operations teams informed in real-time Database logging maintains a complete audit trail of all schedule changes The system processes only confirmed schedule changes to avoid false notifications Passenger notifications are sent only to those with confirmed tickets for affected flights How It Works Schedule Trigger - Automatically runs every 30 minutes to check for flight schedule updates Fetch Airline Data - Retrieves current flight information from aviation APIs Get Current Schedules - Pulls existing schedule data from the internal database Process Changes - Compares API data with database records to identify schedule changes Check for Changes - Determines if any updates require processing and notifications Update Database - Saves schedule changes to the internal flight database Notify Slack Channel - Sends operational updates to the flight operations team Check Urgent Notifications - Identifies critical changes requiring immediate passenger alerts Get Affected Passengers - Retrieves contact information for passengers on changed flights Send Email Notifications - Dispatches detailed schedule change emails via SendGrid Send SMS (Critical Only) - Sends urgent text alerts for cancellations and major delays Update Internal Systems - Syncs changes with other airline systems via webhooks Log Sync Activity - Records all synchronization activities for audit and monitoring Data Sources The workflow integrates with multiple data sources and systems: Aviation API (Primary Data Source) Real-time flight status and schedule data Departure/arrival times, gates, terminals Flight status (on-time, delayed, cancelled, diverted) Aircraft and route information Internal Flight Database flight_schedules table - Current schedule data with columns: flight_number (text) - Flight identifier (e.g., "AA123") departure_time (timestamp) - Scheduled departure time arrival_time (timestamp) - Scheduled arrival time status (text) - Flight status (active, delayed, cancelled, diverted) gate (text) - Departure gate number terminal (text) - Terminal identifier airline_code (text) - Airline IATA code origin_airport (text) - Departure airport code destination_airport (text) - Arrival airport code aircraft_type (text) - Aircraft model updated_at (timestamp) - Last update timestamp created_at (timestamp) - Record creation timestamp passengers table - Passenger contact information with columns: passenger_id (integer) - Unique passenger identifier name (text) - Full passenger name email (text) - Email address for notifications phone (text) - Mobile phone number for SMS alerts notification_preferences (json) - Communication preferences created_at (timestamp) - Registration timestamp updated_at (timestamp) - Last profile update tickets table - Booking and ticket status with columns: ticket_id (integer) - Unique ticket identifier passenger_id (integer) - Foreign key to passengers table flight_number (text) - Flight identifier flight_date (date) - Travel date seat_number (text) - Assigned seat ticket_status (text) - Status (confirmed, cancelled, checked-in) booking_reference (text) - Booking confirmation code fare_class (text) - Ticket class (economy, business, first) created_at (timestamp) - Booking timestamp updated_at (timestamp) - Last modification timestamp sync_logs table - Audit trail and system logs with columns: log_id (integer) - Unique log identifier workflow_name (text) - Name of the workflow that created the log total_changes (integer) - Number of schedule changes processed sync_status (text) - Status (completed, failed, partial) sync_timestamp (timestamp) - When the sync occurred details (json) - Detailed log information and changes error_message (text) - Error details if sync failed execution_time_ms (integer) - Processing time in milliseconds Communication Channels Slack - Internal team notifications SendGrid - Passenger email notifications Twilio - Critical SMS alerts Internal webhooks - System integrations How to Use Import the workflow into your n8n instance Configure aviation API credentials (AviationStack, FlightAware, or airline-specific APIs) Set up PostgreSQL database connection with required tables Configure Slack bot token for operations team notifications Set up SendGrid API key and email templates for passenger notifications Configure Twilio credentials for SMS alerts (critical notifications only) Test with sample flight data to verify all notification channels Adjust monitoring frequency and severity thresholds based on operational needs Monitor sync logs to ensure reliable data synchronization Requirements API Access Aviation data provider (AviationStack, FlightAware, etc.) SendGrid account for email delivery Twilio account for SMS notifications Slack workspace and bot token Database Setup PostgreSQL database with flight schedule tables Passenger and ticket management tables Audit logging tables for tracking changes Infrastructure n8n instance with appropriate node modules Reliable internet connection for API calls Proper credential management and security Customizing This Workflow Modify the Process Changes node to adjust change detection sensitivity, add custom business rules, or integrate additional data sources like weather or airport operational data. Customize notification templates in the email and SMS nodes to match your airline's branding and communication style. Adjust the Schedule Trigger frequency based on your operational requirements and API rate limits.
by InfraNodus
This template can be used to generate research questions from PDF documents (e.g. research papers, market reports) based on the content gaps found in text using the InfraNodus knowledge graph GraphRAG knowledge graph representation. Simply upload several PDF files (research papers, corporate or market reports, etc) and generate a research question / AI prompt in seconds. The template is useful for: generating research questions generating AI prompts that drive research further finding blind spots in any discourse and generating ideas that address them. avoiding the generic bias of LLM models and focusing on what's important in your particular context Using Content Gaps for Generating Research Questions Knowledge graphs represent any text as a network: the main concepts are the nodes, their co-occurrences are the connections between them. Based on this representation, we build a graph and apply network science metrics to rank the most important nodes (concepts) that serve as the crossroads of meaning and also the main topical clusters that they connect. Naturally, some of the clusters will be disconnected and will have gaps between them. These are the topics (groups of concepts) that exist in this context (the documents you uploaded) but that are not very well connected. Addressing those gaps can help you see which groups of concepts you could connect with your own ideas. This is exactly what InfraNodus does: builds the structure, finds the gaps, then uses the built-in AI to generate research questions that bridge those gaps. How it works 1) Step 1: First, you upload your PDF files using an online web form, which you can run from n8n or even make publicly available. 2) Steps 2-4: The documents are processed using the Code and PDF to Text nodes to extract plain text from them. 3) Step 5: This text is then sent to the InfraNodus GraphRAG node that creates a knowledge graph, identifies structural gaps in this graph, and then uses built-in AI to research questions / prompts. 4) Step 6: The ideas are then shown to the user in the same web form. Optionally, you can hook this template to your own workflow and send the question generated to an InfraNodus expert or your own AI model / agent for further processing. If you'd like to sync this workflow to PDF files in a Google Drive folder, you can copy our Google Drive PDF processing workflow for n8n. How to use You need an InfraNodus GraphRAG API account and key to use this workflow. Create an InfraNodus account Get the API key at https://infranodus.com/api-access and create a Bearer authorization key. Add this key into the InfraNodus GraphRAG HTTP node(s) you use in this workflow. You do not need any OpenAI keys for this to work. Optionally, you can change the settings in the Step 4 of this workflow and enforce it to always use the biggest gap it identifies. Requirements An InfraNodus account and API key Note: OpenAI key is not required. You will have direct access to the InfraNodus AI with the API key. Customizing this workflow You can use this same workflow with a Telegram bot or Slack (to be notified of the summaries and ideas). You can also hook up automated social media content creation workflows in the end of this template, so you can generate posts that are relevant (covering the important topics in your niche) but also novel (because they connect them in a new way). Check out our n8n templates for ideas at https://n8n.io/creators/infranodus/ Also check the full tutorial with a conceptual explanation at https://support.noduslabs.com/hc/en-us/articles/20454382597916-Beat-Your-Competition-Target-Their-Content-Gaps-with-this-n8n-Automation-Workflow Also check out the video introduction to InfraNodus to better understand how knowledge graphs and content gaps work: For support and help with this workflow, please, contact us at https://support.noduslabs.com
by Taiki
Workflow Setup Guide This workflow collects the most-viewed videos from specified YouTube channels and saves the data to a Google Sheet. Follow these steps to set it up: 1. Credentials Setup Google Sheets:** You need to have a Google Sheets credential configured in your n8n instance. If you don't have one, go to the 'Credentials' section in n8n and add a new credential for Google Sheets. YouTube API Key:** You need a YouTube Data API v3 key. Go to the Google Cloud Console. Create a new project or select an existing one. Go to 'APIs & Services' > 'Library' and enable the YouTube Data API v3. Go to 'APIs & Services' > 'Credentials', click 'Create Credentials', and choose 'API key'. Copy the generated API key. 2. Google Sheet Setup You will need one Google Sheet with two separate sheets (tabs) inside it. Input Sheet Use Template This sheet provides the list of YouTube channels to process. Required Columns:** Create a sheet with the following two columns: ChannelID: The ID of the YouTube channel (e.g., T7M3PpjBZzw). video_num_to_get: The number of top videos to retrieve for that channel (e.g., 5). Output Sheet This sheet is where the results will be saved. Required Columns:** The workflow will automatically append data to the following columns. You can create them beforehand or let the workflow do it. channelName title videoId videoLink 3. Node Configuration Read Channel Info from Sheet:** Select your Google Sheets credential. Enter your Spreadsheet ID. Enter the name of your Input Sheet. Fetch Most-Viewed Videos via YouTube API:** Replace YOUR_YOUTUBE_API_KEY with the API key you generated in Step 1. Append Video Details to Sheet:** Select your Google Sheets credential. Enter your Spreadsheet ID (the same one as before). Enter the name of your Output Sheet.
by Rex Lui
Easily generate QR codes from any URL! This workflow lets users submit a URL via a simple form and instantly receive a downloadable QR code image—perfect for quick sharing or promotions. Setup is fast and user-friendly, so you’ll be up and running in minutes! 🚀 How it works The end user submits a URL through a simple online form. The workflow automatically sends the submitted URL to a QR code generation API. The user receives a downloadable QR code image corresponding to their URL. ⚙️ Setup instruction Import Workflow: Click "Import from JSON" in your n8n environment and paste the provided workflow JSON. Click "Save" and activate the workflow. Double click the "On form submission" node to obtain the production URL. You may now use this URL to do QR code generation.
by Akash Kankariya
All-in-One Portfolio Tracker & Telegram Finance Updates Workflow for n8n: Multi-Broker, Real-Time, Global 🚀 Overview Take control of all your investments—across multiple brokers and platforms—in one place, with live updates sent directly to your Telegram! 🌍💸 This n8n template brings together Google Sheets and Telegram so you can track your complete finance portfolio with ease, whether you’re in the US market, India, or anywhere in the world. 🔧 Built By - akash@codescale.tech How This Workflow Works Tracks your investments** across multiple brokers, platforms, or asset types. Automatically sends updates to your Telegram account**—see daily Profit & Loss (P&L), changes, and total returns in a rich, emoji-filled report. Works globally**, with a sample provided for the US market, but can be configured for any country and broker. Schedule automated updates** (e.g., market close/open) or get real-time insights on demand with Telegram commands. Highlights & Features 📊 Unified Dashboard: Integrate all your broker data in one Google Sheet for effortless monitoring (Google Sheet Link - https://docs.google.com/spreadsheets/d/1dakq9EhU8GrDgBsk82KvAen0N1P3FySAwNHFtG2lsLI/edit?usp=sharing) 🤖 Interactive Telegram Bot: Send /total or a specific broker’s name in the Telegram chat to get instant, formatted portfolio summaries. ⏰ Automatic Notifications: Receive scheduled P&L summaries at market open and close. 🗂️ Customizable for Any Region or Broker: Just update your Google Sheet with the platforms or brokers you use—including those in the US, Europe, Asia, etc. 🔐 Secure and Private: Only your pre-set Telegram user or chat receives the sensitive financial update. Example (For US Market) Let’s imagine you have portfolios with Robinhood, E*TRADE, and Charles Schwab. Every day at 10AM and 4PM Eastern Time, or whenever you send the /total command, you get this on Telegram: 📊 Daily P&L Report 🔹 Robinhood Invested: $5,000.00 P&L: $250.00 (5.00%) Change: $30.00 (0.60%) Current Value: $5,250.00 🔹 E*TRADE Invested: $8,000.00 P&L: $400.00 (5.00%) Change: $45.00 (0.56%) Current Value: $8,400.00 📈 Total Portfolio Total Invested: $13,000.00 Total P&L: $650.00 (5.00%) Today's Change: $75.00 (0.58%) 💰 Overall Value: $13,650.00 📈 Overall Return: 5.00% 💸 Overall P&L: $650.00 Easy Setup Steps Copy the Template to Your n8n Instance: Just import the provided workflow JSON. Configure Your Google Sheet: List all your brokers/platforms as rows (US, EU, or any other market). Update your credentials in n8n for Google Sheets and Telegram. Set Your Telegram Chat ID: Secure, so only you or your group receive updates. Customize Schedules: Change times for your local market hours or as you prefer. Send Commands in Telegram: /total for overall summary /Robinhood, /ETRADE, etc., for individual broker updates Who Is This For? Investors managing accounts across several brokers. Traders seeking real-time daily summaries. Portfolio managers wanting one consolidated, secure view. Users in any country, for any major market. Make It Yours! 🌏 Customize the sheet and workflow for your unique blend of accounts, currencies, and platforms—track mutual funds, stocks, ETFs, cryptos, or more. Get peace of mind with every notification, organized and delivered just for you! Start tracking smarter, not harder. Transform your finance workflow with n8n + Telegram today! 🚀
by Yar Malik (Asfandyar)
Intro This template is for teams, individuals, or businesses who want to automatically send daily email reminders (e.g., updates, status alerts, follow‑ups) using n8n + Gmail. How it works Cron Trigger fires every day at your specified time. Google Sheets node reads all rows from your sheet. If node filters rows matching your condition (e.g., Status = "Pending"). Send a message (Gmail) sends a customized email to each filtered row. Required Google Sheet Structure | Column Name | Type | Example | Notes | |-------------|--------|--------------------------|------------------------------------| | Email | string | user@example.com | Recipient email address | | Status | string | Pending | Filter criterion | | Subject | string | Daily Status Update | Email subject (supports variables) | | Body | string | “Please update your task”| Email body (text or HTML) | Detailed Setup Steps Google Sheets Build your sheet with the columns above. In n8n → Credentials, add Google Sheets API (avoid sensitive names). Gmail In n8n → Credentials → Gmail (OAuth2 or SMTP), connect your account. Do not include your real email in the credential name. Import & Configure Export the workflow JSON (three‑dot menu → Export). Paste it under Template Code in the Creator form. In each node, select your Google Sheets and Gmail credentials. Sticky Notes On the If node: “Defines which rows to email.” On the Gmail node: “Sends the email.” Customization Guidance Adjust schedule: change the Cron expression in **Cron Trigger. Modify filter: edit the condition in the **If node. Customize email**: use expressions like {{$node["Get row(s) in sheet"].json["Subject"]}}. Troubleshooting Verify the Google Sheet is shared with the connected service account. Check your Cron timezone and expression. Ensure Gmail credentials are valid and not rate‑limited. Security & Best Practices Remove** any real email addresses and sheet IDs. Use** n8n Credentials or environment variables—never hard‑code secrets. Add** sticky notes for any complex logic.
by Rosh Ragel
What it Does Automatically checks your Google Calendar to determine if you're officially off work for the rest of today. If so, it auto-sends a personalized out‑of‑office reply via Gmail, telling senders when you’ll be back—based on your next calendar entry within the next 2 weeks. Prerequisites To use this template, you'll need: Gmail credentials (for the trigger and reply nodes) Google Calendar credentials (for both calendar checks) A dedicated work calendar selected in the Calendar nodes Workflow Logic Gmail Trigger Monitors incoming emails every minute Can be filtered (e.g., labels or VIP senders) Calendar Check #1 Inspects if any events remain today Calendar Check #2 If no remaining events, scan the next 14 days for the next event Function Node Formats the return date as Weekday, Month D, YYYY (e.g., “Thursday, July 24, 2025”) Gmail Send Sends a customized out‑of‑office email, using the formatted date Optionally includes n8n attribution (editable) User Setup Instructions Gmail Trigger: Connect your Gmail account and add any desired filters (labels, senders). Google Calendar Nodes: Connect your calendar account and select your “work” calendar in both nodes. Function Node: No changes needed unless you prefer a different date format. Gmail Send Node: Edit the message template and toggle attribution as desired. Customization - Options Edit the final email content and tone in the Send node Adjust calendar lookahead in Calendar Check #2 (default is 14 days) Add Gmail filters to restrict auto-replies (e.g. only specific senders or labels) Why It's Useful Ideal for freelancers, consultants, or remote workers who don’t follow a strict 9–5, yet want automated responses aligned with their actual availability, not a static setting. It’s dynamic, real-time, and easy to tweak. Classification Use Case: Calendar-driven out-of-office automation Recommended audience: Business professionals, freelancers, remote employees
by Audun
Who Is This For? Web developers SEO specialists Digital marketers What Problem Is This Workflow Solving? Automates the extraction of internal links from a webpage Eliminates the manual and error-prone process of collecting links Facilitates analysis of website structure and optimization What This Workflow Does Uses HTTP request node to fetch HTML content from a specified webpage Parses the HTML to identify and extract internal links Compiles a list of URLs directing to pages within the same domain Setup Configure the Set Base URL node: Set the url field to the URL you want to analyze. How to Customize This Workflow to Your Needs Change the target URL in the Set Base URL node to analyze different webpages. Add nodes to: Filter or categorize the extracted links Export the list to a database or CSV Send links via email or integrate with other tools This workflow can be used as a base for workflows to manage the process of extracting internal links, aiding in website optimization and SEO efforts.
by Aitor | 1Node
Elevate your Stripe workflows with an AI agent that intelligently, securely, and interactively handles essential Stripe data operations. Leveraging the Kimi K2 model via OpenRouter, this n8n template enables safe data retrieval. From fetching summarized financial insights to managing customer discounts, while strictly enforcing privacy, concise outputs, and operational boundaries. 🧾 Requirements Stripe: Active Stripe account API key with read and write access. n8n: Deployed n8n instance (cloud or self-hosted) OpenRouter: Active OpenRouter account with credit API key from OpenRouter 🔗 Useful Links Stripe n8n Stripe Credentials Setup OpenRouter 🚦 Workflow Breakdown Trigger: User Request Workflow initiates when an authenticated user sends a message in the chat trigger. AI Agent (Kimi K2 OpenRouter): Intent Analysis Determines whether the user wants to: List customers, charges, or coupons Retrieve the account’s balance Create a new coupon in Stripe Filters unsupported or unclear requests, explaining permissions or terminology as needed. Stripe Data Retrieval For data queries: Only returns summarized, masked lists (e.g., last 10 transactions/customers) Sensitive details, such as card numbers, are automatically masked or truncated Never exposes or logs confidential information Coupon Creation When a coupon creation is requested: AI agent collects coupon parameters (discount, expiration, restrictions) Clearly summarizes the action and requires explicit user confirmation before proceeding Creates the coupon upon confirmation and replies with only the public-safe coupon details 🛡️ Privacy & Security No data storage:** All responses are ephemeral; sensitive Stripe data is never retained. Strict minimization:** Outputs are tightly scoped; only partial identifiers are shown and only when necessary. Retention rules enforced:** No logs, exports, or secondary storage of Stripe data. Confirmation required:** Actions modifying Stripe (like coupon creation) always require the user to approve before execution. Compliance-ready:** Aligned with Stripe and general data protection standards. ⏱️ Setup Steps Setup time: 10–15 minutes Add Stripe API credentials in n8n Add the OpenRouter API credentials in n8n and select your desired AI model to run the agent. In our template we selected Kimi K2 from Moonshot AI. ✅ Summary This workflow template connects a privacy-prioritized AI agent (Kimi K2 via OpenRouter) with your Stripe account to enable: Fast, summarized access to customer, transaction, coupon, and balance data Secure, confirmed creation of discounts/coupons Complete adherence to authorization, privacy, and operational best practices 🙋♂️ Need Help? Feel free to contact us at 1 Node Get instant access to a library of free resources we created.
by Nikhil Kuriakose
How it works Triggers on submitting an n8n form Uses the form details to prepare a message Sends the message to Slack Set up Steps Add in your team name Add in message tone Set up Open AI Set up Slack
by CustomJS
n8n Workflow: Invoice PDF Generator This n8n workflow captures invoice data and generates a PDF invoice, ready to be sent or saved. It uses a webhook to trigger the process, preprocesses the invoice data, and converts it to a PDF using HTML and custom styling. @custom-js/n8n-nodes-pdf-toolkit Features: Webhook Trigger**: Receives incoming data, including invoice details. Preprocessing**: Transforms the invoice data into HTML format. HTML to PDF Conversion**: Converts the preprocessed HTML into a styled PDF document. Response**: Sends the generated PDF back to the webhook response. Notice Community nodes can only be installed on self-hosted instances of n8n. Requirements Self-hosted** n8n instance A CustomJS API key for website screenshots. Invoice data** for PDF generation Workflow Steps: Webhook Trigger: Accepts incoming data (e.g., invoice number, recipient details, itemized list). This data is passed to the next node for processing. Set Data Node: Configures initial values for the invoice, including the recipient, sender, invoice number, and the items on the invoice. The invoice details include information like description, unit price, and quantity. Preprocess Node: Processes the raw data to format it correctly for HTML. This includes splitting addresses and converting the items into an HTML table format. HTML to PDF Conversion: Converts the generated HTML into a PDF document. The HTML includes a header, a detailed invoice table, and a footer with contact information. Respond to Webhook: Returns the generated PDF as a response to the initial webhook request. Setup Guide: 1. Configure CustomJS API Sign up at CustomJS. Retrieve your API key from the profile page. Add your API key as n8n credentials. 2. Design Workflow Create a Webhook: Set up a webhook to trigger the workflow when invoice data is received. Prepare Data: Ensure the incoming request contains fields like "Invoice No", "Bill To", "From", and "Details" (list of items with price and quantity). Customize the HTML: The HTML template for the invoice includes custom styling to give the invoice a professional look. Convert to PDF: The HTML to PDF node is configured with the data generated from the preprocessing step to convert the invoice HTML to a PDF format. Example Invoice Data: { "Invoice No": "1", "Bill To": "John Doe\n1234 Elm St, Apt 567\nCity, Country, 12345", "From": "ABC Corporation\n789 Business Ave\nCity, Country, 67890", "Details": [ { "description": "Web Hosting", "price": 150, "qty": 2 }, { "description": "Domain", "price": 15, "qty": 5 } ], "Email": "support@mycompany.com" } Result PDF File
by ivn
About: This workflow automates the transcription of YouTube videos by processing a video URL provided via a chat message. Designed for users who need quick access to video content in text form, this workflow ensures a seamless experience for transcribing videos on demand, regardless of the topic. Who is this for? This workflow is designed for individuals who need quick and accurate transcriptions of YouTube videos without watching them in full. It is particularly useful for: Students who need text-based notes from educational videos. Researchers looking to extract information from lectures or discussions. Professionals who prefer reading over watching videos. Casual users who want an efficient way to summarize video content. What problem is this workflow solving? Manually transcribing YouTube videos is time-consuming and prone to errors. Watching long videos just to extract key information is inefficient. This workflow automates transcription, allowing users to quickly convert video content into text. Use cases include: Summarizing lectures or webinars. Extracting insights from interviews and discussions. Creating searchable text from video content. Generating reference material without watching entire videos. What This Workflow Does? This workflow automates the transcription of YouTube videos by: Accepting Input: User provide a YouTube video URL through a chat message. Processing the Video: It utilizes an external transcription service to retrieve the full transcript of the YouTube video from the provided URL. Enhancing Output: An AI model (OpenAI) refines the transcription for accuracy and readability. Delivering Results: The final text transcript is returned to the user via the chat interface. Setup: Install n8n: Ensure you have n8n installed and running. Import the Workflow: Copy the JSON workflow file into your n8n instance. Configure API Keys: Set up your Supadata (Supadata) API key for transcription. Configure the OpenAI (OpenAI) API key for additional processing. Run the Workflow: Provide a YouTube video URL and receive a transcription in response. How to customize this workflow to your needs: The workflow is flexible and can be tailored to suit specific requirements. Here are some customization ideas: Language Support:** Adjust the transcription language in both the HTTP Request and OpenAI nodes to support transcriptions in different languages (e.g., French, German). Integrate with Other Services:** Store transcriptions in a database, send them via email, or connect with a document management system. Notification:** Add a notification node (e.g., email or Slack) to alert you when the transcription is complete, especially for long videos. Quality Check:** Integrate an additional AI step to summarize or highlight key points in the transcript for quicker insights. This workflow is designed to be scalable, efficient, and adaptable to various transcription needs. Limitations Video Length Limitation:** Very long videos may not have a complete transcription due to constraints in processing capacity or service limitations. Transcription Dependency:** The accuracy of the transcription relies entirely on the presence of video captions or subtitles. If a video lacks these, no transcription will be generated. Access Restrictions:** Private or restricted YouTube videos may not be accessible for transcription due to permission limitations. Processing Time:** The time required to process a video can vary significantly, especially for longer videos, depending on the transcription service and server resources. Regional Restrictions:** Some YouTube videos may have geographic or regional access limitations, which could prevent the workflow from retrieving the content for transcription.