by Dahiana
Monitor website performance with PageSpeed Insights and save to Google Sheets with alerts This n8n template automatically monitors website performance using Google's PageSpeed Insights API, compiles detailed reports, and tracks performance trends over time in Google Sheets. Use cases: Agency client monitoring, competitor analysis, performance regression detection, SEO reporting, site migration monitoring, A/B testing performance impact, and maintaining performance SLAs. Who's it for Digital agencies monitoring client websites SEO professionals tracking site performance DevOps teams maintaining performance SLAs Business owners wanting automated site monitoring How it works Automated Testing:** Scheduled audits of multiple websites using PageSpeed Insights API Core Web Vitals:** Tracks LCP, FID, CLS, and overall performance scores Historical Tracking:** Maintains performance history for trend analysis Alert System:** Sends notifications when performance drops below thresholds Detailed Reporting:** Captures specific recommendations and optimization opportunities Two Workflow Paths Scheduled Audit: Automatically tests all URLs from Google Sheet on schedule On-Demand Testing: Webhook endpoint for immediate single-URL testing How to set up Get a free PageSpeed Insights API key from Google Cloud Console Create Google Sheet with columns: URL, Site Name, Category, Alert Threshold, Last_Processed_Date and Device. Set up Google Sheets API credentials Configure notification preferences (Slack, email, etc.) Set audit schedule (daily, weekly, or custom) Define performance thresholds for alerts Requirements Google PageSpeed Insights API key (free) Google Sheets API access n8n instance (cloud or self-hosted) Optional: Slack/email for notifications Google Sheet Structure Input Sheet ("sites"): URL, Site_Name, Category, Alert_Threshold, Last_Processed_Date and Device. Results Sheet ("audit_results"): Date, URL, Site_Name, Device, Performance_Score, LCP, FID, CLS, Recommendations, Full_Report_URL API Usage (On-Demand) POST to webhook: { "url": "https://example.com", "site_name": "Example Site", "alert_threshold": 75 } How to customize Add custom performance thresholds per site Include additional metrics (accessibility, SEO, best practices) Connect to other dashboards (Data Studio, Grafana) Add competitor benchmarking Integrate with project management tools for issue tracking Set up different notification channels based on severity Sample Google Sheet Included
by Sateesh
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. AI-Powered LinkedIn Publishing via Telegram Workflow Transform your LinkedIn presence with this intelligent n8n workflow that converts simple Telegram messages into professional LinkedIn posts through AI-powered content generation and approval workflows. 🎯 Who Is This For? Content Creators & Influencers** seeking to maintain consistent LinkedIn presence Marketing Professionals** managing multiple client accounts Business Owners** wanting to automate thought leadership content Social Media Managers** streamlining content workflows Entrepreneurs** maximizing content efficiency while maintaining quality 🚀 Benefits Time Efficiency**: Reduces content creation time by 80-90% Quality Consistency**: Maintains professional standards across all posts Content Diversity**: Leverages multiple sources for rich, varied content Real-Time Relevance**: Incorporates latest industry trends and news Approval Control**: Human oversight ensures brand alignment Scalability**: Handles multiple users and high-volume content creation 🔧 Core Features Smart Content Classification Multi-Input Processing**: Handles URLs, topics, direct content, or combinations Intelligent Routing**: Automatically determines whether to scrape, search, or generate directly Context Preservation**: Maintains original user intent throughout the process Advanced Content Gathering Web Scraping**: Firecrawl integration for extracting article content from URLs Real-Time Search**: Brave Search API for latest industry trends and news Content Synthesis**: Merges multiple sources into coherent, valuable insights AI-Powered Content Generation Google Gemini Integration**: Creates professional, LinkedIn-optimized posts Platform-Specific Formatting**: Mobile-friendly paragraphs, engaging hooks, strategic CTAs SEO Optimization**: Relevant hashtags and keyword integration Character Management**: Ensures posts stay within LinkedIn's 2800 character limit Interactive Approval System Telegram Preview**: Rich preview with post analytics and formatting Action Buttons**: Approve, Edit, or Reject with single-click convenience Edit Workflow**: AI-powered rewriting based on user feedback Real-Time Updates**: Instant feedback and status notifications Comprehensive Content Tracking Google Sheets Integration**: Complete audit trail of all posts and content metrics Content Analytics**: Character counts, hashtag usage, source attribution User Authorization**: Secure access control with authorized user validation Post Management**: Unique ID generation for tracking and reference 🔄 How It Works Message Reception: Secure Telegram trigger with user validation Content Classification: AI analyzes input type and extracts actionable elements Dynamic Routing: Intelligent branching based on content requirements: URL Path: Web scraping → content extraction → processing Topic Path: Web search → latest information gathering → synthesis Direct Path: Immediate processing for ready-to-post content Content Synthesis: Merges all gathered information into comprehensive context AI Generation: Creates LinkedIn-optimized post with professional formatting Interactive Approval: Telegram preview with approval workflow Publishing: Direct LinkedIn posting upon approval Content Logging: Complete tracking in Google Sheets 📈 Use Cases Daily Industry Updates: Transform news URLs into thought leadership posts Content Repurposing: Convert articles and research into LinkedIn insights Trend Commentary: Generate posts about trending topics with real-time data Educational Content: Create informative posts from technical documentation Personal Branding: Maintain consistent professional presence with minimal effort 🛠️ Technical Requirements Required Community Nodes Install these community nodes in your n8n instance: Brave Search Integration @brave/n8n-nodes-brave-search Firecrawl Web Scraping @mendable/n8n-nodes-firecrawl LangChain AI Integration @n8n/n8n-nodes-langchain APIs & Services Required Google Gemini (Content generation and classification) Firecrawl API (Web scraping) Brave Search API (Real-time search) Telegram Bot API (Interface and notifications) LinkedIn API (Content publishing) Google Sheets API (Content tracking and logging) 🔑 Setup Guide 1. Telegram Bot Setup Search for @BotFather on Telegram Send /newbot and follow prompts Copy the bot token Send /setprivacy to BotFather and set to Disable 2. Google Gemini API Visit Google AI Studio Sign in and click "Get API Key" → "Create API Key" Copy your API key Free tier: 60 requests per minute 3. Firecrawl API Visit Firecrawl.dev Sign up and go to Dashboard → API Keys Copy your API key Free tier: 500 pages/month 4. Brave Search API Visit Brave Search API Sign up and create application Copy subscription key Free tier: 1,000 queries/month 5. LinkedIn API Visit LinkedIn Developers Create app with required details Request "Share on LinkedIn" product Copy Client ID and Client Secret Add redirect URL: https://your-n8n-domain.com/rest/oauth2-credential/callback 6. Google Sheets API Visit Google Cloud Console Enable Google Sheets API Create OAuth 2.0 Client ID Copy Client ID and Client Secret 🛠️ Installation Steps Phase 1: Preparation Install required community nodes Restart n8n after installation Create Google Sheet for logging Set up Telegram Bot Phase 2: Import and Configure Import workflow JSON in n8n Configure all API credentials Test each connection Phase 3: Customization Update authorized user ID in "Authorized Telegram Users" node Configure Google Sheets document ID Test Telegram connection Phase 4: Testing Test with different input types: URL only: https://example.com/article Topic only: artificial intelligence trends Mixed: AI trends https://example.com/ai-news 🎨 Customization Options Content Personalization Modify AI prompts to match your brand voice Adjust content length and formatting preferences Customize hashtag strategies and CTA approaches Configure approval workflow steps Source Integration Add additional search engines or content sources Integrate with RSS feeds or news APIs Connect to internal knowledge bases Customize web scraping parameters 🔒 Security Features User Authorization**: Whitelist-based access control Secure Token Management**: Encrypted API key handling Data Privacy**: Secure processing of scraped content Audit Trail**: Complete logging of all user interactions 🔮 Future Expansion Possibilities This workflow serves as a foundation for: Performance Analytics Module**: LinkedIn engagement tracking Content Optimization Engine**: A/B testing and refinement Multi-Platform Publishing**: Expand to Twitter, Facebook, Instagram Advanced Scheduling**: Time-optimized posting Content Series Management**: Automated follow-ups 💡 Why Choose This Workflow This represents a complete LinkedIn content automation solution that maintains quality and personal touch while dramatically reducing time and effort. Perfect for professionals who want to maximize LinkedIn impact without sacrificing content quality or spending hours on manual creation. Ready to transform your LinkedIn presence? Install this workflow and start automating your professional content creation today!
by ConnectSafely
Rewrite viral LinkedIn posts in your voice with AI and Telegram approval using Google Gemini Who's it for This workflow is designed for LinkedIn creators, personal brand builders, thought leaders, and content marketers who want to consistently create engaging content without starting from scratch. Perfect for professionals who admire viral posts from others but want to adapt those ideas to their own unique voice and style. If you're struggling to maintain a consistent posting schedule, looking for content inspiration, or want to repurpose trending ideas while keeping your authentic voice, this automation handles the creative heavy lifting while giving you full control over what gets published. How it works The workflow transforms viral LinkedIn posts into personalized content that matches your writing style, complete with AI-generated images, all controlled through Telegram. The process flow: Send any LinkedIn post URL to your Telegram bot Security check validates your Telegram user ID ConnectSafely.ai scrapes the original post content and engagement metrics Your custom persona profile is loaded (tone, phrases, formatting preferences) Google Gemini AI rewrites the post to match YOUR voice Gemini generates a professional, on-brand image for the post Preview is sent to Telegram for your review Approve or reject with a simple reply On approval, the post goes live on LinkedIn automatically Setup steps Step 1: Create a Telegram Bot Open Telegram and search for @BotFather Send /newbot and follow the prompts to create your bot Save the API token provided by BotFather Get your Telegram User ID by messaging @userinfobot Step 2: Configure Telegram Credentials in n8n Go to Credentials → Add Credential → Telegram API Paste your bot token from BotFather Save the credential Update all Telegram nodes to use this credential Step 3: Set Up Security Check Open the 🔒 Security Check node Replace YOUR_TELEGRAM_USER_ID with your actual Telegram user ID This ensures only YOU can trigger the workflow Step 4: Configure ConnectSafely.ai API Sign up at ConnectSafely.ai Navigate to Settings → API Keys in your dashboard Generate a new API key In n8n, go to Credentials → Add Credential → ConnectSafely API Paste your API key and save Connect this credential to the 🔍 Scrape LinkedIn Post node Step 5: Configure Google Gemini API Go to Google AI Studio Create or select a project Generate an API key In n8n, go to Credentials → Add Credential → Google Gemini (PaLM) API Paste your API key and save Connect this credential to both: Google Gemini Chat Model node Generate an image node Step 6: Connect Your LinkedIn Account In n8n, go to Credentials → Add Credential → LinkedIn OAuth2 API Follow the OAuth flow to connect your LinkedIn account Connect this credential to the Create LinkedIn Post node Update the person parameter with your LinkedIn Person ID (URN) Step 7: Customize Your Persona Open the 👤 Load Your Persona node Edit the PERSONA object to match YOUR writing style: Update name with your name Modify expertiseAreas with your topics Adjust commonPhrases with phrases you actually use Set preferredEmojis to your favorites Customize styleNotes to capture your unique voice Step 8: Activate the Workflow Save your workflow Toggle the workflow to Active Your Telegram bot is now ready to receive LinkedIn URLs Customization Persona Customization The 👤 Load Your Persona node is where you define your unique voice. Key areas to customize: | Field | Description | Example | |-------|-------------|---------| | tone | Overall communication style | "Professional yet approachable, data-driven" | | voice | Perspective and personality | "First-person, authentic, vulnerable" | | formatting | Structure preferences | "Short paragraphs, emoji bullets, line breaks" | | hooks | Opening style | "Start with contrarian takes or personal stories" | | expertiseAreas | Your niche topics | ["SaaS growth", "Leadership", "Remote work"] | | commonPhrases | Signature expressions | ["Here's the truth:", "I learned this the hard way:"] | Image Generation The 📝 Create Image Prompt node generates the image prompt. Modify the style parameters to match your brand: Current style**: Modern, clean, corporate, vector art Customize**: Change to photography, illustrations, or abstract visuals Post Length In the persona configuration, adjust postLength: "short" - Quick insights (under 500 characters) "medium" - Standard posts (500-1500 characters) "long" - Deep dives (1500-3000 characters) AI Model Selection The workflow uses gemini-2.5-pro for text. You can switch to other models in the Google Gemini Chat Model node based on your needs. Requirements | Requirement | Details | |-------------|---------| | n8n Version | 1.0+ recommended | | Telegram Bot | Created via @BotFather | | ConnectSafely.ai Account | API key required | | Google AI Studio Account | Gemini API key required | | LinkedIn Account | OAuth2 connected in n8n | | Community Node | n8n-nodes-connectsafely-ai (self-hosted only) | ⚠️ Note: This workflow uses the ConnectSafely community node, which requires a self-hosted n8n instance. Use cases Content Repurposing**: Transform competitor or industry leader posts into your own perspective Consistent Posting**: Maintain a regular posting schedule without content creation burnout Style Consistency**: Ensure every post matches your established personal brand Trend Riding**: Quickly create content around viral topics while they're still relevant A/B Testing**: Test different approaches by approving or rejecting variations Troubleshooting Common Issues & Solutions Issue: Bot not responding to messages Solution**: Verify the Telegram webhook is active; check the Telegram Trigger node is properly configured Issue: "Profile not found" from ConnectSafely.ai Solution**: Ensure the LinkedIn URL is complete and public. Some posts on private profiles can't be scraped Issue: Image generation fails Solution**: Verify your Gemini API key has access to image generation models. Check quota limits in Google AI Studio Issue: LinkedIn post fails to publish Solution**: Confirm your LinkedIn OAuth2 credentials are valid and haven't expired. Re-authorize if needed Issue: AI generates posts that don't match your style Solution**: Be more specific in your persona configuration. Add more example phrases and detailed style notes Issue: Security check blocks your messages Solution**: Double-check your Telegram User ID is correctly entered (must be a number, not username) Documentation & Resources Official Documentation ConnectSafely.ai Docs**: https://connectsafely.ai/docs Google Gemini API**: https://ai.google.dev/docs Telegram Bot API**: https://core.telegram.org/bots/api LinkedIn API**: https://docs.microsoft.com/linkedin/ Support ConnectSafely Support**: support@connectsafely.ai n8n Community**: https://community.n8n.io Connect With Us Stay updated with the latest automation tips, LinkedIn strategies, and platform updates: LinkedIn**: linkedin.com/company/connectsafelyai YouTube**: youtube.com/@ConnectSafelyAI-v2x Instagram**: instagram.com/connectsafely.ai Facebook**: facebook.com/connectsafelyai X (Twitter)**: x.com/AiConnectsafely Bluesky**: connectsafelyai.bsky.social Mastodon**: mastodon.social/@connectsafely Need Custom Workflows? Looking to build sophisticated LinkedIn automation workflows tailored to your business needs? Contact our team for custom automation development, strategy consulting, and enterprise solutions. We specialize in: Multi-channel engagement workflows AI-powered personalization at scale Lead scoring and qualification automation CRM integration and data synchronization Custom reporting and analytics pipelines
by Angel Menendez
Workflow Description Who is this for? This workflow is designed for sales and revenue teams using Gong and Salesforce to track and analyze sales calls. It helps automate the extraction, filtering, and preprocessing of Gong call data for further AI analysis. What problem is this solving? Sales teams often generate large amounts of call data, but not all calls are relevant for deeper analysis. This workflow filters calls based on predefined criteria, extracts relevant metadata, and formats the data before passing it to an AI processing pipeline. What this workflow does Triggers on new Gong calls synced to Salesforce** every hour. Filters calls based on opportunity stage** (Discovery or Meeting Booked). Retrieves Gong call details** via API. Formats call data into a structured JSON object** for AI processing. Passes the structured data to a Gong Call Preprocessor workflow** for further insights. Setup Ensure that you have connected Salesforce and Gong APIs with valid credentials. Modify the Salesforce query in Get all custom Salesforce Gong Objects to match your organization’s requirements. Set the schedule trigger interval in the Run Hourly node if needed. Connect this workflow to an AI processing workflow to analyze call transcripts. Workflow Templates: CallForge - 01 - Filter Gong Calls Synced to Salesforce by Opportunity Stage CallForge - 02 - Prep Gong Calls with Sheets & Notion for AI Summarization CallForge - 03 - Gong Transcript Processor and Salesforce Enricher CallForge - 04 - AI Workflow for Gong.io Sales Calls CallForge - 05 - Gong.io Call Analysis with Azure AI & CRM Sync CallForge - 06 - Automate Sales Insights with Gong.io, Notion & AI CallForge - 07 - AI Marketing Data Processing with Gong & Notion CallForge - 08 - AI Product Insights from Sales Calls with Notion How to customize Change filtering logic: Adjust the **opportunity stage filter (Check if Opportunity Stage is Meeting Booked or Discovery) to match your sales process. Modify data formatting**: Add or remove fields in the Format call into correct JSON Object node to customize the output. Adjust trigger frequency**: Change the Run Hourly node to run at a different interval if required.
by Danielle Gomes
Automatically classify incoming leads based on the sentiment of their message using Google Gemini, store them in Supabase by category, and send tailored WhatsApp messages via the official WhatsApp Cloud API. ✅ Use Case: This workflow is ideal for sales, onboarding, and customer support teams who want to: Understand the tone and urgency of each lead Prioritize hot leads instantly Send smart, automatic WhatsApp replies based on user sentiment 🧠 How it works: Capture lead via a Typeform webhook Clean and structure the data (name, email, message, etc.) Run sentiment analysis using Google Gemini to classify the message as: Positive → Hot Lead Neutral → Warm Lead Negative → Cold Lead Store lead data in Supabase under the corresponding category Merge data to unify flow paths Send WhatsApp message using the official WhatsApp Cloud API, with a custom reply for each sentiment result 🔧 Tools used: Typeform (incoming data) Google Gemini (AI-based sentiment classification) Supabase (database) WhatsApp Cloud API (response automation) 🏷 Tags: AI, Sentiment Analysis, Lead Qualification, Supabase, WhatsApp, Gemini, Typeform, CRM, Automation, Customer Engagement
by Piotr Sobolewski
How it works This advanced workflow transforms your long-form audio content (like podcast episodes or webinar recordings) into digestible, ready-to-use marketing assets. It's designed for podcasters, content creators, and marketers who want to maximize their content's reach. It automatically: Takes a full transcript of your audio/video content as input. Generates a concise, comprehensive summary of the episode using advanced AI. Extracts a list of key topics and keywords from the transcript, perfect for SEO, tagging, and content categorization. Delivers the summary and keywords directly to your inbox or a connected tool for easy access. Streamline your content repurposing pipeline and unlock new value from your audio and video assets with intelligent automation! Set up steps Setting up this powerful workflow typically takes around 20-30 minutes, as it involves multiple AI steps. You'll need to: Obtain API keys for your preferred AI service (e.g., OpenAI, Google AI). Have access to a method for generating transcripts from your audio/video (e.g., manually pasting, or using a separate transcription service like AssemblyAI, Whisper, etc.). Connect your preferred email service (e.g., Gmail) to receive the output. All detailed setup instructions and specific configuration guidance are provided within the workflow itself using sticky notes.
by Solomon
This n8n template demonstrates how to obtain token usage from AI Agents and places the data into a spreadsheet that calculates the estimated cost of the execution. Obtaining the token usage from AI Agents is tricky, because it doesn't provide all the data from tool calls. This workflow taps into the workflow execution metadata to extract token usage information. Works well with OpenAI, Google and Anthropic. Other LLM providers might need small tweaks. How it works The AI Agent executes and then calls a subworkflow to calculate the token usage. The data is stored in Google Sheets The spreadsheet has formulas to calculate the estimated cost of the execution. How to use The AI Agent is used as an example. Feel free to replace this with other agents you have. Call the subworkflow AFTER all the other branches have finished executing. Requirements LLM account (OpenAI, Gemini...) for API usage. Google Drive and Sheets credentials n8n API key of your instance
by Danger
How it Works This meta-workflow is designed to intelligently scan all your active workflows in n8n, identify those that contain Webhook nodes, and automatically generate a Swagger (OpenAPI) specification based on them. The output Swagger document reflects all accessible endpoints from your Webhook nodes, making it easier to: Visualize your API structure Share your endpoints Integrate with tools like Postman or Swagger UI Enhanced Parameter Support If you want the Swagger to reflect request parameters (e.g., query or body fields), you can annotate your Webhook nodes using the Note section. When configured properly, these annotations enrich your Swagger documentation with parameter names, types, and descriptions. Setup Steps Add the WebhookDocs to n8n Import the WebhookDocs JSON file into your n8n instance. Activate the WebhookDocs (you can also use the test-endpoint) Annotate Webhook Nodes (Optional but Recommended) To enable parameter documentation, open the Note section of each Webhook node and add annotations in the following format: //@body field_name string description //@query field_name string description Open the page https://n8n.youristance.com/webhook/swagger
by Jimleuk
This n8n template demonstrates how to use OpenAI's Responses API with existing LLM and AI Agent nodes. Though I would recommend just waiting for official support, if you're impatient and would like a round-about way to integrate OpenAI's responses API into your existing AI workflows then this template is sure to satisfy! This approach implements a simple API wrapper for the Responses API using n8n's builtin webhooks. When the base url is pointed to these webhooks using a custom OpenAI credential, it's possible to intercept the request and remap for compatibility. How it works An OpenAI subnode is attached to our agent but has a special custom credential where the base_url is changed to point at this template's webhooks. When executing a query, the agent's request is forwarded to our mini chat completion workflow. Here, we take the default request and remap the values to use with a HTTP node which is set to query the Responses API. Once a response is received, we'll need to remap the output for Langchain compatibility. This just means the LLM or Agent node can parse it and respond to the user. There are two response formats, one for streaming and one for non-streaming responses. How to use You must activate this workflow to be able to use the webhooks. Create the custom OpenAI credential as instructed. Go to your existing AI workflows and replace the LLM node with the custom OpenAI credential. You do not need to copy anything else over to the existing template. Requirements OpenAI account for Responses API Customising this workflow Feel free to experiment with other LLMs using this same technique! Keep up to date with the Responses API announcements and make modifications as required.
by Aadarsh Jain
Who is this for? This workflow is designed for DevOps engineers, platform engineers, and Kubernetes administrators who want to interact with their Kubernetes clusters through natural language queries in n8n. It's perfect for teams who need quick cluster insights without memorizing complex kubectl commands or switching between multiple cluster contexts manually. How it works? The workflow operates in three intelligent stages: Cluster Discovery & Context Switching - Automatically lists available clusters from your kubeconfig and switches to the appropriate cluster based on your natural language query Command Generation - Uses GPT-4o to analyze your request and generate the correct kubectl command with proper flags, selectors, and output formatting Command Execution - Executes the generated kubectl command against your selected cluster and returns the results The workflow supports multi-cluster environments and can handle queries like: "Show me all pods in production cluster" "List failing deployments in production" "Get pod details in kube-system namespace" Setup Clone the MCP Server git clone https://github.com/aadarshjain/kubectl-mcp-server cd kubectl-mcp-server Configure your kubeconfig - Ensure your ~/.kube/config contains all the clusters you want to access Set up MCP STDIO credentials in n8n Command: /full/path/to/python-package Arguments: /full/path/to/kubectl-mcp-server/server.py Import the workflow into your n8n instance Configure OpenAI credentials for the GPT-4o models Test the workflow using the chat interface with queries like "show pods in [cluster-name]"
by Mary Newhauser
RAG over a PDF with Weaviate This workflow allows you to upload a PDF file and ask questions about it using the Question and Answer Chain and the Weaviate Vector Store nodes. Who it's for This workflow is the simplest possible implementation of RAG with Weaviate in n8n. It's intended to act as an extendable template for RAG over your own documents. Prerequisites An existing Weaviate cluster. You can view instructions for setting up a local cluster with Docker here or a Weaviate Cloud cluster here. API keys to generate embeddings and power chat models. We use OpenAI, but feel free to switch out the models as you like. Self-hosted n8n instance. See this video for how to get set up in just three minutes. How it works Part 1: Manually upload data In this example, we manually upload a 100+ page article from arXiv called "A Survey of Large Language Models". But you can replace this with your own more advanced data pipeline, if you wish. Part 2: Embed and load data into Weaviate collection Here, we generate embeddings for the full-text of the article and store them in Weaviate. Part 3: Perform RAG over PDF file with Weaviate In this part of the workflow, you can enter your query by running the Chat Node and get a RAG response grounded in context via the Question and Answer Chain node. How to run the workflow Go through the prerequisites, creating a Weaviate cluster (can be local or cloud), downloading self-hosted n8n, and adding your API keys and other credentials. Select the embedding and chat models you'd like to use. Upload a PDF file you want to ask questions about. Execute the rest of the workflow.
by AI/ML API | D1m7asis
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. n8n Workflow Template: AI‑Powered Mental Health Support Bot Overview: This template enables you to build a Telegram bot that delivers real‑time, empathetic mental health support. Incoming messages tagged with #vent, #insight, or #cope are routed to GPT‑4o via the AI/ML API, which returns tailored, compassionate responses. How it works: Telegram Trigger listens for new chat messages or voice notes. Show Typing Indicator immediately signals “typing…” in the chat. Switch Node examines the text prefix and routes to one of four branches (Vent, Insight, Cope, or default). Set Prompt nodes build a JSON payload with a specific role‑play prompt for each branch. AI/ML API node (model gpt-4o) generates the response. Telegram node sends the AI’s answer back to the user. Setup Steps: Connect your Telegram bot token in the Telegram credentials. Add your AI/ML API key (GPT‑4o) in n8n’s credential settings. Activate the workflow and deploy your n8n instance webhook URL to BotFather. Test by sending #vent I’m stressed, #insight Why do I feel…, or any tag in your Telegram chat. This plug‑and‑play workflow brings AI‑driven emotional support directly into Telegram.