by Angel Menendez
Video Demo: Click here to see a video of this workflow in action. Summary Description: The "IT Department Q&A Workflow" is designed to streamline and automate the process of handling IT-related inquiries from employees through Slack. When an employee sends a direct message (DM) to the IT department's Slack channel, the workflow is triggered. The initial step involves the "Receive DMs" node, which listens for new messages. Upon receiving a message, the workflow verifies the webhook by responding to Slack's challenge request, ensuring that the communication channel is active and secure. Once the webhook is verified, the workflow checks if the message sender is a bot using the "Check if Bot" node. If the sender is identified as a bot, the workflow terminates the process to avoid unnecessary actions. If the sender is a human, the workflow sends an acknowledgment message back to the user, confirming that their query is being processed. This is achieved through the "Send Initial Message" node, which posts a simple message like "On it!" to the user's Slack channel. The core functionality of the workflow is powered by the "AI Agent" node, which utilizes the OpenAI GPT-4 model to interpret and respond to the user's query. This AI-driven node processes the text of the received message, generating an appropriate response based on the context and information available. To maintain conversation context, the "Window Buffer Memory" node stores the last five messages from each user, ensuring that the AI agent can provide coherent and contextually relevant answers. Additionally, the workflow includes a custom Knowledge Base (KB) tool (see that tool template here) that integrates with the AI agent, allowing it to search the company's internal KB for relevant information. After generating the response, the workflow cleans up the initial acknowledgment message using the "Delete Initial Message" node to keep the conversation thread clean. Finally, the generated response is sent back to the user via the "Send Message" node, providing them with the information or assistance they requested. This workflow effectively automates the IT support process, reducing response times and improving efficiency. To quickly deploy the Knowledge Ninja app in Slack, use the app manifest below and don't forget to replace the two sample urls: { "display_information": { "name": "Knowledge Ninja", "description": "IT Department Q&A Workflow", "background_color": "#005e5e" }, "features": { "bot_user": { "display_name": "IT Ops AI SlackBot Workflow", "always_online": true } }, "oauth_config": { "redirect_urls": [ "Replace everything inside the double quotes with your slack redirect oauth url, for example: https://n8n.domain.com/rest/oauth2-credential/callback" ], "scopes": { "user": [ "search:read" ], "bot": [ "chat:write", "chat:write.customize", "groups:history", "groups:read", "groups:write", "groups:write.invites", "groups:write.topic", "im:history", "im:read", "im:write", "mpim:history", "mpim:read", "mpim:write", "mpim:write.topic", "usergroups:read", "usergroups:write", "users:write", "channels:history" ] } }, "settings": { "event_subscriptions": { "request_url": "Replace everything inside the double quotes with your workflow webhook url, for example: https://n8n.domain.com/webhook/99db3e73-57d8-4107-ab02-5b7e713894ad", "bot_events": [ "message.im" ] }, "org_deploy_enabled": false, "socket_mode_enabled": false, "token_rotation_enabled": false } }
by Joseph LePage
From PDF to Powerful Blog Posts: AI-Powered Content Transformation Turn complex documents into engaging digital content that drives results. This n8n Workflow uses AI to transforms lengthy PDFs into compelling blog posts that attract and retain readers while you focus on strategic initiatives. Time-Saving Innovation 🚅Lightning-Fast Processing Transform lengthy documents into polished blog content in under 1 minute, eliminating hours of manual work. Our system handles the heavy lifting, delivering up to a 95% reduction in content production time. 📱Intelligent Analysis The AI engine identifies and extracts key insights, organizing information for maximum impact. Each document undergoes comprehensive analysis to ensure no valuable content is overlooked. Advanced Content Optimization ✍️Dynamic Writing Styles Possible Adjust the prompt for multiple tone options: Professional for corporate communications Conversational for engaging blogs Thought leadership for industry authority 📊SEO-Ready Content Potential Adjust the prompt to automatically optimized for search engines, incorporating relevant keywords and semantic structure to improve visibility and drive organic traffic. Ideal Applications 🤼Content Marketing Teams Scale content production without sacrificing quality or consistency. Perfect for teams looking to maintain a robust publishing schedule while maximizing resource efficiency. 🏫Academic Communication Help researchers and institutions share complex findings with broader audiences through accessible, engaging content that maintains academic integrity. 🧑💻Digital Publishers Streamline the content transformation process while ensuring each piece meets modern digital standards and reader expectations. Transform your content strategy with an intelligent system that delivers consistent, high-quality results while dramatically reducing production time.
by Joseph LePage
-- Disclaimer: This workflow uses a community node and therefore only works for self-hosted n8n users -- Transform YouTube videos into comprehensive summaries and structured analysis instantly. This n8n workflow automatically extracts, processes, and analyzes video transcripts to deliver clear, organized insights without watching the entire video. Time-Saving Features 🚀 Instant Processing Simply provide a YouTube URL and receive a structured summary within seconds, eliminating the need to watch lengthy videos. Perfect for research, learning, or content analysis. 🤖 AI-Powered Analysis Leverages GPT-4o-mini to analyze video transcripts, organizing key concepts and insights into a clear, hierarchical structure with main topics and essential points. Smart Processing Pipeline 📝 Automated Transcript Extraction Supports public YouTube video Handles multiple URL formats Extracts complete video transcripts automatically 🧠 Intelligent Content Organization Breaks down content into main topics Highlights key concepts and terminology Maintains technical accuracy while improving clarity Structures information logically with markdown formatting Perfect For 📚 Researchers & Students Quick comprehension of educational content and lectures without watching entire videos. 💼 Business Professionals Efficient analysis of industry talks, presentations, and training materials. 🎯 Content Creators Rapid research and competitive analysis of video content in your niche. Technical Implementation 🔄 Workflow Components Webhook endpoint for URL submission YouTube API integration for video details Transcript extraction system GPT-4 powered analysis engine Telegram notification system (optional) Transform your video content consumption with an intelligent system that delivers structured, comprehensive summaries while saving hours of viewing time.
by Yaron Been
This workflow provides automated access to the Notdaniel Voxtral Small 24B 2507 AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for audio generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete audio generation process using the Notdaniel Voxtral Small 24B 2507 model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Voxtral Small is an enhancement of Mistral Small 3 that incorporates state-of-the-art audio input capabilities and excels at speech transcription, translation and audio understanding. Key Capabilities AI-driven audio generation and processing** High-quality sound synthesis** Advanced audio manipulation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Notdaniel/voxtral-small-24b-2507 AI model Notdaniel Voxtral Small 24B 2507**: The core AI model for audio generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Audio Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Music Production**: Generate background music and audio tracks Podcast Enhancement**: Create intro/outro music and sound effects Audio Content**: Produce voiceovers and audio narration Sound Design**: Generate custom audio for games and applications Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #audiogeneration #aiaudio #soundgeneration #musicai #audioautomation #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by getBible
Overview The Get Bible Query Workflow is a modular and self-standing workflow designed to retrieve scriptures dynamically based on structured input. It serves as an intermediary layer that extracts references, queries the GetBible API, and returns scriptures in a standardized JSON format. This workflow is fully prepared for integration—simply call it from another workflow with the required JSON input, and it will return the requested scripture data. Who Is This For? This workflow is ideal for developers, Bible study apps, research tools, and dynamic scripture-based projects that need seamless access to scriptural content without direct API interaction. ✅ Use Cases: Bible Study Apps** → Embed scripture retrieval functionality. Research & Theology Tools** → Fetch structured verse data. Dynamic Content Generation** → Integrate real-time scripture references. Sermon Preparation** → Automate scripture lookups. How It Works Trigger Workflow → This workflow is designed to be called from another workflow with a structured JSON input. Receive Input → Accepts a JSON object containing references, translation, and API version. Extract References → Parses single verses, comma-separated lists, and ranged passages. Query API → Sends structured requests to the GetBible API. Format Response → Returns structured JSON output, maintaining API response consistency. JSON Input Structure References** → Should include the book name, chapter, and verse(s). Multiple Verses** → Separated by commas (e.g., John 3:16,18). Verse Ranges** → Defined with a dash (e.g., John 3:16-18). Translation** → Choose from the supported translations. API Version** → Currently supports v2. Example JSON Input { "references": [ "1 John 3:16", "Jn 3:16", "James 3:16", "Rom 3:16" ], "translation": "kjv", "version": "v2" } Example API Response { "result": { "kjv_62_3": { "translation": "King James Version", "abbreviation": "kjv", "book_name": "1 John", "chapter": 3, "ref": ["1 John 3:16"], "verses": [ { "chapter": 3, "verse": 16, "name": "1 John 3:16", "text": "Hereby perceive we the love of God, because he laid down his life for us: and we ought to lay down our lives for the brethren." } ] } } } 💡 Fully structured and formatted response – ready for seamless integration. Integration and Usage The GetBible Query Workflow is designed for immediate use. Simply call it from another workflow and pass the appropriate JSON object as input, and it will return the requested scripture passages. ✔️ No additional configuration is required. ✔️ Designed for fast, reliable, and structured scripture retrieval. ✔️ Fully compatible with GetBible API responses. Why Use This Workflow? ✔️ Fast & Reliable → Direct API integration for efficient queries. ✔️ Flexible Queries → Supports single, multi-verse, and ranged requests. ✔️ Agent-Compatible → Easily integrates into automated workflows. ✔️ No Code Needed → Just configure the JSON input and run the workflow. Next Steps 🔗 API Support 📖 API Documentation 💬 Need help? Join the community for support! 🚀
by Davide
How It Works Form Submission: The workflow starts with the On form submission node, which triggers when a user submits a contact form. The form collects the user's name, email, and message. Text Classification: The Text Classifier node uses an AI model (GPT-4) to classify the submitted message into one of the predefined categories: Request Quote: For quote requests. Product info: For general product inquiries. General problem: For issues or problems related to products. Order: For questions about placed orders. Other: For any messages that don’t fit the above categories. Email Routing: Based on the classification, the workflow routes the message to the appropriate department via email: Prod. Dep.: For product-related inquiries. Quote Dep.: For quote requests. Gen. Dep.: For general problems. Order Dep.: For order-related questions. Other Dep.: For all other inquiries. Each email includes the user's name, email, message, and the classified category. Data Logging: The workflow logs the form submission and classification results into a Google Sheets document. Each department has its own sheet where the data is appended, including: User’s name, email, and message. Submission date and time. Assigned category. Email recipient details. AI Model Integration: The OpenAI node provides the AI model (GPT-4) used by the Text Classifier to classify the messages. The model is instructed to classify the text into one of the predefined categories without additional explanations. Set Up Steps Configure the Form Trigger: Set up the On form submission node to collect user inputs (name, email, and message) and trigger the workflow. Set Up the Text Classifier: Configure the Text Classifier node to use the OpenAI model (GPT-4) for text classification. Define the categories and their descriptions (e.g., "Request Quote", "Product info", etc.). Set the fallback category to "Other" for unclassifiable messages. Configure Email Sending: Set up the Email Send nodes for each department (Prod. Dep., Quote Dep., Gen. Dep., Order Dep., Other Dep.). Configure the email subject, body, and reply-to address using the form data and classification results. Ensure SMTP credentials are correctly configured for sending emails. Set Up Google Sheets Integration: Configure the Google Sheets nodes to append data to the appropriate sheets for each department. Map the form data (name, email, message, date, category, and recipient) to the corresponding columns in the Google Sheets document. Test the Workflow: Submit a test form to ensure the workflow correctly classifies the message, sends the email to the right department, and logs the data in Google Sheets. Verify that the OpenAI model is classifying messages accurately. Activate the Workflow: Once tested, activate the workflow to automate the process of handling contact form submissions. Key Features Automated Classification**: Uses AI to classify messages into relevant categories, reducing manual effort. Email Routing**: Sends emails to the appropriate department based on the classification. Data Logging**: Logs all form submissions and classification results in Google Sheets for tracking and analysis. Scalability**: Easily adaptable to additional categories or departments by modifying the workflow. This workflow is ideal for eCommerce businesses or customer support teams looking to automate and streamline the handling of contact form submissions. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by lin@davoy.tech
Are you looking to create a counseling chatbot that provides emotional support and mental health guidance through the LINE messaging platform ? This guide will walk you through connecting LINE with powerful AI language models like GPT-4 to build a chatbot that supports users in navigating their emotions, offering 24/7 conversational therapy and accessible mental health resources . By leveraging LINE's webhook integration and Azure OpenAI , this template allows you to design a chatbot that is both empathetic and efficient, ensuring users receive timely and professional responses. Whether you're a developer, counselor, or business owner, this guide will help you create a customizable counseling chatbot tailored to your audience's needs. Who Is This Template For? Developers who want to integrate AI-powered chatbots into the LINE platform for mental health applications. Counselors & Therapists looking to expand their reach and provide automated emotional support to clients outside of traditional sessions. Businesses & Organizations focused on improving mental health accessibility and offering innovative solutions to their users. Educators & Nonprofits seeking tools to provide free or low-cost counseling services to underserved communities. How this work? Line Webhook to receive new message Send loading animation in Line Check if the input is text or not Send the text as prompt in chat model (GPT 4o) Reply the message to user (you'll need 'edit field' to format it before reply) Pre-Requisites You have access to the LINE Developers Console. An Azure OpenAI account with necessary credentials. Set-up To receive messages from LINE, configure your webhook: Set up a webhook in LINE Developer Console. Copy the Webhook URL from the Line Chatbot node and paste it into the LINE Console. Ensure to remove any 'test' part when moving to production. The loading animation reassures users that the system is processing their request. Authorize using header authorization Message Handling Use the Check Message Type IsText? node to verify if the incoming message is text. If the message type is text, proceed with ChatGPT processing; otherwise, send a reply indicating non-text inputs are not supported. AI Agent Configuration Define the system message within the AI Agent node to guide the conversation based on desired interaction principles. Connect the Azure OpenAI Chat Model to the AI Agent. Formatting Responses Ensure responses are properly formatted before sending them back to the user. Reply Message Use the ReplyMessage - Line node to send the formatted response. Ensure proper header authorization using Bearer tokens.
by Udit Rawat
This workflow is for automating and centralizing your bookmarking process using AI-powered tagging and seamless integration between your Android device and a self-hosted Read Deck platform (https://readeck.org/en/). This workflow eliminates manual entry, organizes links with smart AI-generated tags, and ensures your bookmarks are always accessible, searchable, and secure. How It Works 📱 Android Shortcut Integration Use the HTTP Shortcuts app to create a 1-tap trigger that sends URLs and titles from your Android phone directly to n8n. 🤖 AI-Powered Tagging & Processing Leverage ChatGPT-4 to analyze content context and auto-generate relevant tags (e.g., “Tech Tutorials,” “Productivity Tools”). Extract clean titles and URLs from messy shared data (even from apps like Twitter or Reddit). 🔗 Readeck Integration Automatically save processed bookmarks to your self-hosted Readeck-like platform with structured metadata (title, URL, tags). ⚡ Silent Automation It runs in the background—no pop-ups or interruptions. 🔒 Pro Security Optional authentication (API tokens, headers) to protect your data. Use Case Perfect for researchers, content creators, or anyone drowning in tabs who wants to: Save articles, videos, or social posts in one click. Organize bookmarks with AI-generated tags. Build a personal knowledge base that’s always accessible. Tutorial 1️⃣ Set Up Android Shortcut Install "HTTP Shortcuts" and configure it to send data to your n8n webhook. Enable “Share Menu” to trigger bookmarks from any app. 2️⃣ Configure n8n Workflow Import the template and add your Read Deck API token (or similar service). 3️⃣ Test & Scale Share a link from your phone—watch it appear in Read Deck instantly! Add error handling or notifications for advanced use. Note: For self-hosted platforms, ensure your instance is publicly accessible (or use a VPN). Why Choose This Workflow? Zero Manual Entry: Save hours of copying/pasting. AI Organization: Say goodbye to chaotic bookmark folders. Privacy First: Host your data on your terms. Transform your bookmarking chaos into a streamlined system—try “Save: Bookmark” today! 🚀
by Don Jayamaha Jr
Track NFT market trends, collections, and trades in real time—directly from Telegram! This master workflow integrates the OpenSea API, GPT-4o-mini AI, and Telegram, allowing users to request natural-language NFT analytics and receive structured insights instantly. Whether you're an NFT trader, collector, or market analyst, this Telegram-native assistant brings you on-demand market intelligence—powered by OpenSea and AI. > ⚠️ Important: This workflow requires three sub-workflows to function properly. These must be downloaded and published in your n8n instance. 🧩 Required Sub-Workflows To activate this template, download and publish the following workflows: Analyze NFT Market Trends with AI-Powered OpenSea Analytics Agent Tool Get Real-time NFT Insights with OpenSea AI-Powered NFT Agent Tool Get Real-time NFT Marketplace Insights with OpenSea Marketplace Agent Tool 📌 You can also find these by visiting my Creator profile: 👉 https://n8n.io/creators/don-the-gem-dealer/ How It Works A Telegram bot receives a message (e.g., “Top sales for Azuki”). The AI router in this workflow determines which agent should process the request: Marketplace Agent → Listings, offers, and orders Analytics Agent → Sales volume, price trends, wallet behavior NFT Agent → Metadata, traits, ownership info The selected agent queries the OpenSea API using your API key. The response is processed using GPT-4o-mini, formatted, and sent back via Telegram. What You Can Do with This Agent 🔹 Discover undervalued NFTs based on trait rarity and price 🔹 Track market trends for any collection in real time 🔹 Compare collection performance by volume, sales, and listings 🔹 Analyze flipping trends and whale activity across wallets 🔹 Retrieve NFT ownership and metadata instantly 🔹 View trait-specific offers for insight into rarity-driven demand Example Queries You Can Use ✅ "What are the cheapest NFTs in the Pudgy Penguins collection?" ✅ "Get sales volume for Azuki and CloneX over the last 30 days." ✅ "Who owns Bored Ape #456?" ✅ "Show the best current offers for Moonbirds." Set Up Steps Create a Telegram Bot Use @BotFather to create your bot and get the API token. Get an OpenSea API Key Apply for your API key via the OpenSea Developer Portal. Configure n8n Credentials Add your Telegram Bot and OpenSea API Key under Credentials in n8n. Download Required Sub-Workflows Install and publish the following workflows: Analytics Agent Tool NFT Agent Tool Marketplace Agent Tool Deploy & Test Chat with your Telegram bot. Try: "Compare BAYC and Azuki volume" or "Show listings for Doodles." ✅ Final Notes > If your queries don’t respond correctly, make sure all three sub-workflows are installed and published, not just saved. 🚀 Dominate the NFT market with AI-powered OpenSea intelligence—right from your Telegram inbox!
by Samir Saci
Tags: Supply Chain, Logistics, Control Tower Context Hey! I’m Samir, a Supply Chain Engineer and Data Scientist from Paris, and the founder of LogiGreen Consulting. We design tools to help companies improve their logistics processes using data analytics, AI, and automation—to reduce costs and minimize environmental impact. > Let’s use N8N to build smarter and more sustainable supply chains! 📬 For business inquiries, you can add me on LinkedIn Who is this template for? This workflow template is designed for logistics operations that need a monitoring solution for their distribution chains. Connected to your Transportation Management Systems, this AI agent can answer any question about the shipments handled by your distribution teams. How does it work? The workflow is connected to a Google BigQuery table that stores outbound order data (customer deliveries). Here’s what the AI agent does: 🤔 Receives a user question via chat. 🧠 Understands the request and generates the correct SQL query. ✅ Executes the SQL query using a BigQuery node. 💬 Responds to the user in plain English. Thanks to the chat memory, users can ask follow-up questions to dive deeper into the data. What do I need to get started? This workflow requires no advanced programming skills. You’ll need: A Google BigQuery account with an SQL table storing transactional records. An OpenAI API key (GPT-4o) for the chat model. Next Steps Follow the sticky notes in the workflow to configure each node and start using AI to support your supply chain operations. 🎥 Watch My Tutorial 🚀 Curious how N8N can transform your logistics operations? Notes The chat trigger can easily be replaced with Teams, Telegram, or Slack for a better user experience. You can also connect this to a customer chat window using a webhook. This workflow was built using N8N version 1.82.1 Submitted: March 24, 2025
by Tom
This is the workflow powering the n8n demo shown at StrapiConf 2022. The workflow searches matching Tweets every 30 minutes using the Interval node and listens to Form submissions using the Webhook node. Sentiment analysis is handled by Google using the Google Cloud Natural Language node before the result is stored in Strapi using the Strapi node. (These were originally two separate workflows that have been combined into one to simplify sharing.)
by Friedemann Schuetz
This n8n workflow template uses community nodes and is only compatible with the self-hosted version of n8n. Welcome to my Wikipedia Podcast Telegram Bot Workflow! This workflow creates an intelligent Telegram bot that transforms Wikipedia articles into engaging 5-minute podcast episodes using natural language queries and voice messages. What this workflow does This workflow processes incoming Telegram messages (text or voice, e.g. "Berlin") and generates professional podcast content about any Wikipedia topic (e.g. "Berlin", "Shakespeare", etc.). The AI agent researches the requested subject, creates a structured podcast script, and delivers it as high-quality audio directly through Telegram. Key Features: Voice message support (speech-to-text and text-to-speech) Wikipedia research integration for accurate content Professional podcast structure (intro, main content, outro) Natural-sounding AI voice synthesis Conversational and educational tone optimized for audio consumption This workflow has the following sequence: Telegram Trigger - Receives incoming messages (text or voice) from users via Telegram bot Text or Voice Switch - Routes the message based on input type (text message vs. voice message) Voice Message Processing (if voice input): Retrieval of voice file from Telegram Transcription of voice message to text using OpenAI Whisper Text Message Preparation (if text input) - Prepares the text message for the AI agent Wikipedia Podcast Agent - Core AI agent that: Researches the requested topic using Wikipedia tool Creates a professional 5-minute podcast script (600-750 words) Follows structured format: intro, main content, outro Uses conversational, accessible, and enthusiastic tone ElevenLabs Text to Speech - Converts the podcast script into natural-sounding audio using AI voice synthesis Send Voice Response - Delivers the generated podcast audio back to the user via Telegram Requirements: Telegram Bot API**: Documentation Create a bot via @BotFather on Telegram Get bot token and configure webhook Anthropic API** (Claude 4 Sonnet): Documentation Used for AI agent processing and podcast script generation Provides Wikipedia research capabilities OpenAI API**: Documentation Used for speech transcription (Whisper model) ElevenLabs API**: Documentation Used for high-quality text-to-speech generation Provides natural-sounding voice synthesis Important: The workflow uses the Wikipedia tool integrated with Claude 4 Sonnet to ensure accurate and comprehensive research. The AI agent is specifically prompted to create engaging, educational podcast content suitable for audio consumption. Configuration Notes: Update the Telegram chat ID in the trigger for your specific bot Modify the voice selection in ElevenLabs for different narrator styles The system prompt can be customized for different podcast formats or target audiences Supports both individual users and can be extended for group chats Feel free to contact me via LinkedIn, if you have any questions!