by Khairul Muhtadin
The Page Speed Insight workflow automates website performance analysis by integrating Google PageSpeed Insights API with Discord messaging and Gemini. This n8n workflow provides expert-level performance audits and comparisons, delivering actionable insights for website owners, SEO professionals, and developers. Disclaimer: this workflow using community nodes Google PageSpeed Insights Community Node 💡 Why Use Page Speed Insight? Save Time:** Instantly analyze and compare website speeds without manual tool usage Eliminate Guesswork:** Receive expert audit reports that translate technical data into clear, actionable insights Improve Website Outcomes:** Identify critical bottlenecks and enhancements prioritized by AI-driven analysis Seamless Integration:** Pull URLs and deliver reports directly via Discord for team collaboration and immediate response ⚡ Who Is This For? Webmasters and website owners seeking fast, automated performance checks SEO analysts who need consistent, data-backed website comparisons Developers requiring clear, prioritized action points from performance audits Digital agencies managing multiple client sites with ongoing monitoring needs 🔧 What This Workflow Does ⏱ Trigger:** Discord message containing URLs or scheduled execution 📎 Parse:** Extracts URLs and determines analysis type (single/comparison) 🔍 Analyze:** Calls Google PageSpeed API for performance data 🤖 Process:** AI generates user-friendly reports from raw Lighthouse JSON 💌 Deliver:** Sends chunked reports to Discord channels 🗂 Log:** Stores execution data for review and improvement 🔐 Setup Instructions Import the provided JSON workflow into your n8n instance Set up credentials for: Google PageSpeed API (ensure you have a valid API key — get yours here) Discord Bot API with permissions to read messages and send messages in your chosen guild/channel Customize the workflow by adjusting: Discord guild and channel IDs where messages are monitored and results posted Scheduled trigger interval if needed Any prompt text or AI model parameters to tailor report tone and detail level Test thoroughly with real URLs and Discord interaction to confirm smooth data flow and output quality 🧩 Pre-Requirements Active n8n instance (Cloud or self-hosted) n8n Google PageSpeed community node Google PageSpeed Insights API key Discord Bot credentials with channel access Google Gemini AI credentials (recommended) 🛠️ Customize It Further Extend to analyze desktop performance or other device types easily by modifying the PageSpeed API call Integrate with Slack, email, or other team tools alongside Discord for broader notification Enhance report depth by adding more AI-driven insights like competitor site recommendations or historical trend tracking 🧠 Nodes Used Google PageSpeed Insights Community Node Discord (getAllMessages, sendMessage) Code (URL parsing, message chunking) AI Language Model (Google Gemini) Schedule Trigger Switch (message type handling) Sticky Notes (workflow guidance) 📞 Support Made by: khaisa Studio Tag: automation, performance, SEO, google-pagespeed, discord Category: Monitoring & Reporting Need a custom solution? Contact Me
by Budi SJ
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 🎯 Purpose This workflow helps you automatically monitor stock related news, extract the main content, summarize it using a LLM (via OpenRouter), and send real time alerts to Telegram and store them in Google Sheets. ⚙️ How It Works Trigger A Cron node triggers the workflow every 15 minutes (adjustable). RSS Feed node checks latest articles from Google Alerts RSS. The workflow filters duplicates using Google Sheets as a log. The article URL is sent to Jina AI Readability API to extract the main body text. The content is summarized using a model from OpenRouter (e.g., Gemini, Claude, GPT-4). You can customize the prompt to suit your tone and analysis needs. The result is appended to a Google Sheets file. Sends the title, summary, and reccomendation to Telegram chat. 🧾 Google Sheets Template Create a Google Sheet using this template: Stock Alert 🧰 Requirements Telegram Bot + your Chat ID OpenRouter account and API key Jina AI account for content extraction Google Account with access to Google Sheets Google Alerts RSS feed 🛠 Setup Instructions Install required credentials: Add OpenRouter API key to n8n credentials. Add Telegram Bot Token and Chat ID. Add Google Sheets credentials. Add Jina AI credentials. Create or copy the Google Sheet using the link above. Go to Google Alerts, create alerts, and copy the RSS feed URL. Replace placeholder API keys and URLs. Adjust Telegram Chat ID. 🔐 Security Note All sensitive credentials (e.g., API keys, personal chat IDs) have been removed from this template. Please replace them using the n8n credentials manager before activating the workflow.
by Khairul Muhtadin
The blogblizt: polylang workflow streamlines the creation and publication of high-quality blog content using powerful automation with n8n, OpenAI’s GPT and the WordPress API. It enables effortlessly generate SEO-friendly articles complete with metadata and optimized featured images, improving content freshness and search engine visibility. 💡 Why Use blogblizt? Automate content creation** to keep your blog fresh and engaging Generate SEO-optimized posts** with expert-crafted titles, meta descriptions, and focus keyphrases Save hours** of manual writing, image sourcing, and SEO configuration Leverage AI** for topic ideation and high-quality writing tailored to international student audiences Seamlessly publish and manage drafts** directly on your WordPress site via API Produce captivating, relevant featured images** without external tools Support multilingual content creation** with randomized language selection for diversity ⚡ Who Is This For? Content strategists managing WordPress blogs needing efficient topic generation SEO specialists wanting automated post creation with optimized metadata Website owners aiming to maintain active, multilingual content Marketers who want to leverage AI for high-quality, consistent article production ❓ What Problem Does It Solve? This workflow automates the entire editorial cycle—from generating engaging topics with AI, drafting full-length articles, producing featured images automatically, to posting drafts configured for SEO on WordPress—dramatically reducing editor workload and improving content output. 🔧 What This Workflow Does ⏱ Trigger Runs on manual trigger or a weekly schedule to ensure consistent content flow 📎 Fetch Site Context Retrieves recent posts, taxonomies, and WordPress API schema to understand site structure 🔍 Generate Topic Uses OpenAI GPT-4.1-mini to roll a random language and craft a targeted blog post topic + SEO metadata 🤖 Draft Article Composes a comprehensive, SEO-friendly article tailored to the generated topic 💌 Create Draft Posts the draft on WordPress with Yoast SEO fields populated 🖼 Generate Image Creates a high-quality, cinematic featured image via AI 📤 Upload & Attach Uploads the image to the WordPress media library and sets it as the post’s featured image 🔐 Setup Instructions Import the workflow file into n8n: Add credentials: WordPress API (with create-post & media permissions) OpenAI API key (for GPT and image models) Customize categories, languages, and schedule in the relevant nodes Adjust the Schedule Trigger timing as desired (e.g. every Monday at 9 AM) Test end-to-end on a staging WordPress site to verify drafts and images publish correctly 🧩 Pre-Requirements An operational n8n instance (Cloud or self-hosted) (self-hosted or n8n cloud) WordPress site with REST API access & proper authentication OpenAI account with API access for both language and image models (Optional) Yoast SEO plugin installed for metadata recognition 🛠️ Customize It Further Tweak OpenAI prompts for niche topics or additional languages Add social-media nodes to auto-share new posts Insert an editorial review step before publishing Refine image prompts for different visual styles (e.g., “modern infographic” vs. “cinematic portrait”) 🧠 Nodes Used Manual Trigger** Schedule Trigger** (weekly) HTTP Request** (fetch posts, taxonomies, schema; upload media) Code** (JavaScript analyzers for API schema & taxonomy parsing) OpenAI Chat** (GPT-4.1-mini for topics & articles) OpenAI Image Generation** (for featured images) WordPress** (create draft post) Sticky Notes** (in-flow documentation) 📞 Support Built by: Khaisa Studio Tags: wordpress, marketing, polylang Category: Content Creation Need a custom? contact me on LinkedIn or Web
by Lucas Peyrin
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. How it works This template is your personal launchpad into the world of AI-powered automation. It provides a fully functional, interactive AI chatbot that you can set up in minutes, designed specifically for those new to AI Agents. What is an AI Agent? Think of it as a smart assistant that doesn't just talk—it acts. You give it a set of "tools" (like other n8n tool nodes), and it intelligently decides which tool to use to answer your questions or complete your tasks. This starter kit comes with a pre-built "toolbox" of superpowers, allowing your agent to: Get the Weather:** Ask for the forecast anywhere in the world. Get the News:** Fetch the latest headlines from n8n, CNN, and others. The workflow is designed to be a hands-on learning experience, with detailed sticky notes explaining every component, from the chat interface to the agent's "brain" and "memory." Set up steps Setup time: ~2-3 minutes This workflow is designed to be incredibly easy to start. You only need one free API key to get it working. Add Your AI Key: The workflow uses Google's Gemini model by default. You will need a free Gemini API key. Find the Gemini node on the canvas. The sticky note right below it (How to Get Google Gemini Credentials) provides a link and simple instructions to get your key. In the Gemini node, click the Credential dropdown and select + Create New Credential to add your key. Activate the Workflow: At the top-right of the screen, click the "Inactive" toggle switch. It will turn green and say "Active". Your agent is now live! Start Chatting: Open the Example Chat Window node (it has a 💬 icon). In its parameter panel, you will see a Chat URL. Click the link to copy it. Paste the URL into a new browser tab and start asking your agent questions! Optional: The template also includes disabled OpenAI chat model node and tools for Google Calendar, and Gmail. You can enable and configure these later to change the underlying AI model or give your agent even more superpowers!
by RedOne
🎙️ AI Audio Assistant with Voice-to-Voice Response Who is this for? Businesses, customer service teams, content creators, and organizations who want to provide intelligent voice-based interactions through Telegram. Perfect for accessibility-focused services, multilingual support, or hands-free customer assistance. What problem does this solve? Enables natural voice conversations with AI Breaks down language and accessibility barriers Provides instant voice responses to customer queries Reduces typing requirements for users Offers 24/7 voice-based customer support Maintains conversation context across voice interactions What this workflow does: Receives voice messages via Telegram bot Transcribes audio using Deepgram's advanced speech-to-text Processes transcribed text through AI agent with knowledge base access Generates intelligent responses based on conversation context Converts AI response to natural-sounding speech using Deepgram TTS Sends audio response back to user via Telegram Maintains conversation memory for contextual interactions 🔧 Technical Architecture Core Components: Telegram Bot**: Receives and sends voice messages Deepgram STT**: Transcribes voice to text with high accuracy OpenAI GPT**: Processes queries and generates responses Supabase Knowledge Base**: Stores and retrieves business information Memory Management**: Maintains conversation context Deepgram TTS**: Converts text responses to natural speech Data Flow: Voice Message → Telegram API → File Download Audio File → Deepgram STT → Transcript Transcript → AI Agent → Response Generation Response → Deepgram TTS → Audio File Audio Response → Telegram → User 🛠️ Setup Instructions Prerequisites Telegram Bot Token Create bot via @BotFather Get bot token and configure webhook Deepgram API Key Sign up at deepgram.com Get API key for STT and TTS services Note: Currently hardcoded in workflow OpenAI API Key OpenAI account with API access Configure in OpenAI Chat Model node Supabase Database Create Supabase project Set up knowledge_base table Configure API credentials Step-by-Step Setup Configure Telegram Bot Update telegramToken in "Prepare Voice Message Data" node Set correct bot token in Telegram nodes Test bot connectivity Set Up Deepgram Integration Replace API key in "Transcribe with Deepgram" node Update TTS endpoint in "HTTP Request" node Test voice transcription accuracy Configure Knowledge Base -- Create knowledge_base table in Supabase CREATE TABLE knowledge_base ( id UUID DEFAULT gen_random_uuid() PRIMARY KEY, question TEXT NOT NULL, answer TEXT NOT NULL, category VARCHAR(100), keywords TEXT[], created_at TIMESTAMP WITH TIME ZONE DEFAULT NOW() ); Customize AI Prompts Update system message in "Telegram AI Agent" node Adjust temperature and max tokens in OpenAI model Configure memory session keys Test End-to-End Flow Send test voice message to bot Verify transcription accuracy Check AI response quality Validate audio output clarity 🎛️ Configuration Options Voice Recognition Settings Model**: nova-2 (Deepgram's latest model) Language**: English (en) - can be changed Smart Format**: Enabled for better punctuation AI Response Settings Temperature**: 0.3 (conservative responses) Max Tokens**: 100 (adjust based on needs) Memory**: Session-based conversation context Text-to-Speech Settings Model**: aura-2-thalia-en (natural female voice) Alternative voices**: Available in Deepgram TTS API Audio Format**: Optimized for Telegram 🔒 Security Considerations API Key Management // Current implementation has hardcoded tokens // Recommended: Use environment variables const telegramToken = process.env.TELEGRAM_BOT_TOKEN; const deepgramKey = process.env.DEEPGRAM_API_KEY; Data Privacy Voice messages are processed by external APIs Consider data retention policies Implement user consent mechanisms Ensure GDPR compliance if applicable 📊 Monitoring & Analytics Key Metrics to Track Voice message processing time Transcription accuracy rates AI response quality scores User engagement metrics Error rates and failure points Recommended Logging // Add to workflow for monitoring console.log({ timestamp: new Date().toISOString(), user_id: userData.user_id, transcript_confidence: transcriptData.confidence, response_length: aiResponse.length, processing_time: processingTime }); 🚀 Customization Ideas Enhanced Features Multi-language Support Add language detection Support multiple TTS voices Translate responses Voice Commands Implement wake words Add voice shortcuts Create voice menus Advanced AI Features Sentiment analysis Intent classification Escalation triggers Integration Expansions Connect to CRM systems Add calendar scheduling Integrate with help desk tools Performance Optimizations Implement audio preprocessing Add response caching Optimize API call sequences Implement retry mechanisms 🐛 Troubleshooting Common Issues Voice Not Transcribing Check Deepgram API key validity Verify audio format compatibility Test with shorter voice messages Poor Audio Quality Adjust TTS model settings Check network connectivity Verify Telegram audio limits AI Responses Too Generic Improve knowledge base content Adjust system prompts Increase context window Memory Not Working Check session key configuration Verify user ID extraction Test conversation continuity 💡 Best Practices Voice Interface Design Keep responses concise and clear Use natural speech patterns Avoid technical jargon Provide clear next steps Knowledge Base Management Regular content updates Clear categorization Keyword optimization Quality assurance testing User Experience Fast response times (<5 seconds) Consistent voice personality Graceful error handling Clear capability communication 📈 Success Metrics Technical KPIs Response time: <3 seconds average Transcription accuracy: >95% User satisfaction: >4.5/5 Uptime: >99.5% Business KPIs Customer query resolution rate Support ticket reduction User engagement increase Cost per interaction decrease 🔄 Maintenance Schedule Daily Monitor error logs Check API rate limits Verify service uptime Weekly Review conversation quality Update knowledge base Analyze usage patterns Monthly Performance optimization Security audit Feature updates User feedback review 📚 Additional Resources Documentation Links Deepgram STT API Deepgram TTS API Telegram Bot API OpenAI API Supabase Documentation Community Support n8n Community Forum Telegram Bot Developers Group Deepgram Developer Discord OpenAI Developer Community Note: This template requires active API subscriptions for Deepgram and OpenAI services. Costs may apply based on usage volume.
by Mind-Front
Workflow Description This workflow is a powerful, fully automated web query and semantic reranking system that allows users to perform precise, detailed searches, intelligently rank search results and provide high-quality, structured output. Built with AI-powered components, the workflow leverages semantic query generation, result re-ranking, and real-time reporting to deliver actionable insights. It is particularly well-suited for real-time data retrieval, market research, and any domain requiring automated yet customizable search result processing. How It Works Webhook Integration for Input: The workflow begins with a Webhook Node that captures the user's search query as input, enabling seamless integration with other systems. Step 1: Semantic Query Generation (Powered by "Semantic Search - Query Maker"): Using AI (Google Gemini), the initial query is refined and transformed into a context-aware, expert-level search query. The process ensures that the search engine retrieves the most relevant and precise results. Step 2: Web Search Execution: A free Brave Search API processes the refined query to fetch search results, ensuring speed and cost efficiency. Step 3: Semantic Re-Ranking of Results (Powered by "Semantic Search - Result Re-Ranker"): The workflow reranks the search results based on relevance to the original question, prioritizing the most relevant URLs dynamically. Results are passed through AI-powered intelligent reranking to ensure the final output reflects optimal relevance and quality. Step 4: Structured Output Generation: Results are converted into a well-structured, organized JSON format, ranking the top 10 search results with their titles, links, and descriptions. Missing ranks (if fewer than 10 results) are handled gracefully with placeholders, ensuring consistency. Step 5: Real-Time Reporting: The reranked search results are sent back to the user or integrated system via the Webhook Node in a JSON-formatted response. Reports are highly structured and ready for downstream processing or consumption. Key Features AI-Powered Query Refinement: Transforms basic queries into detailed, expert-level search terms for optimal results. Dual-Stage Semantic Search: Combines query generation and result reranking for precise, high-relevance outputs. Top 10 Result Reranking: Dynamically ranks and organizes the top 10 results based on semantic relevance to the query. Customizable Integration: Fully modifiable for alternative APIs or integrations, such as other search engines or custom ranking logic. JSON-Formatted Structured Results: Outputs reranked results in a standardized format, ideal for integration into systems requiring machine-readable data. Webhook-Based Flexibility: Works seamlessly with Webhook inputs for easy deployment in diverse workflows. Cost-Effective API Usage: Pre-integrated with the free Brave Search API, minimizing operational costs while delivering accurate search results. Instructions for API Setup Brave Search API: Visit api.search.brave.com to obtain a free-tier API key for web search. AI Integration (Google Gemini): Visit Google AI Studio and generate an API key for semantic query generation and reranking. Webhook Configuration: Set up the input Webhook to capture search queries and the output Webhook to deliver reranked results. Why Choose This Workflow? Precision and Relevance**: Combines AI-based query generation with advanced reranking for accurate results. Fully Customizable**: Easily adapt the workflow to alternative APIs, search engines, or ranking logic. Real-Time Insights**: Provides structured, real-time output ready for immediate use. Scalable and Modular**: Ideal for businesses, researchers, and data analysts needing a robust, repeatable solution. Tags AI Workflow, Semantic Search, Query Refinement, Search Result Reranking, Real-Time Search, Web Search Automation, Google Search, Brave Search, News Search, API Integration, Market Research, Competitive Intelligence, Business Intelligence,Google Gemini, Anthropic Claude, OpenAI, GPT, LLM
by Miha
Combine Tech News in a Personalized Weekly Newsletter This n8n template automates the collection, storage, and summarization of technology news from top sites, turning it into a concise, personalized weekly newsletter. If you like staying informed but want to reduce daily distractions, this workflow is perfect for you. It leverages RSS feeds, vector databases, and LLMs to read and curate tech content on your behalf—so you only receive what truly matters. How it works A daily scheduled trigger fetches articles from multiple popular tech RSS feeds like Wired, TechCrunch, and The Verge. Fetched articles are: Normalized to extract titles, summaries, and publish dates. Converted to vector embeddings via OpenAI and stored in memory for fast semantic querying. A weekly scheduled trigger activates the AI summarization flow: The AI is provided with your interests (e.g., AI, games, gadgets) and the desired number of items (e.g., 15). It queries the vector store to retrieve relevant articles and summarizes the most newsworthy stories. The summary is converted into a clean, email-friendly format and sent to your inbox. How to use Connect your OpenAI and Gmail accounts to n8n. Customize the list of RSS feeds in the “Set Tech News RSS Feeds” node. Update your interests and number of desired news items in the “Your Topics of Interest” node. Activate the workflow and let the automation run on schedule. Requirements OpenAI** credentials for embeddings and summarization Gmail** (or another email service) for sending the newsletter Customizing this workflow Want to use different sources? Swap in your own RSS feeds, or use an API-based news aggregator. Replace the in-memory vector store with Pinecone, Weaviate, or another persistent vector DB for longer-term storage. Adjust the agent's summarization style to suit internal updates, industry-specific briefings, or even entertainment recaps. Prefer chat over email? Replace the email node with a Telegram bot to receive your personalized tech newsletter directly in a Telegram chat.
by Kumar Shivam
This workflow automates the restaurant POS (Point of Sale) data management process, facilitating seamless order handling, customer tracking, inventory management, and sales reporting. It retrieves order details, processes payment information, updates inventory, and generates real-time sales reports, all integrated into a centralized system that improves restaurant operations. The workflow integrates various systems, including a POS terminal to gather order data, payment gateways to process transactions, inventory management tools to update stock, and reporting tools like Google Sheets or an internal database for generating sales and performance reports. Who Needs Restaurant POS Automation? This POS automation workflow is ideal for restaurant owners, managers, and staff looking to streamline their operations: Restaurant Owners – Automate order processing, track sales, and monitor inventory to ensure smooth operations. Managers – Access real-time sales data and performance reports to make informed decisions. Staff – Reduce manual work, focusing on providing better customer service while the system handles orders and payments. Inventory Teams – Automatically update inventory levels based on orders and ingredient usage. If you need a reliable and automated POS solution to manage restaurant orders, payments, inventory, and reporting, this workflow minimizes human error, boosts efficiency, and saves valuable time. Why Use This Workflow? End-to-End Automation – Automates everything from order input to inventory updates and sales reporting. Seamless Integration – Connects POS, payment systems, inventory management, and reporting tools for smooth data flow.(if needed) Real-Time Data – Provides up-to-the-minute reports on sales, stock levels, and order statuses. Scalable & Efficient – Supports multiple locations, multiple users, and high order volumes. Step-by-Step: How This Workflow Manages POS Data Collect Orders – Retrieves order details from the POS system, including customer information, ordered items, and payment details. Update Inventory – Decreases inventory levels based on sold items, ensuring stock counts are always accurate. Generate Reports – Compiles sales, revenue, and inventory data into real-time reports and stores them in Google Sheets or an internal database. Track Customer Data – Keeps a log of customer details and order history for better service and marketing insights. Customization: Tailor to Your Needs Multiple POS Systems – Adapt the workflow to work with different POS systems or terminals based on your restaurant setup. Custom Reporting – Modify the reporting format or include specific sales metrics (e.g., daily totals, best-selling items, employee performance). Inventory Management – Adjust inventory updates to include alerts when stock reaches critical levels or needs reordering. Integration with Accounting Software – Connect with platforms like QuickBooks for automated financial tracking. 🔑 Prerequisites POS System Integration – Ensure the POS system can export order data in a compatible format. Payment Gateway API – Set up the necessary API keys for payment processing (e.g., Stripe, PayPal). Inventory Management Tools – Use inventory software or databases that can automatically update stock levels. Reporting Tools – Use Google Sheets or an internal database to store and generate sales and inventory reports. 🚀 Installation & Setup Configure Credentials Set up API credentials for payment gateways and inventory management tools. Import Workflow Import the workflow into your automation platform (e.g., n8n, Zapier). Link POS system, payment gateway, and inventory management systems. Test & Run Process a test order to ensure that data flows correctly through each step. Verify that inventory updates and reports are generated as expected. ⚠ Important Data Privacy – Ensure compliance with data protection regulations (e.g., GDPR, PCI DSS) when handling customer payment and order data. System Downtime – Monitor system performance to ensure that the workflow runs without disruptions during peak hours. Summary This restaurant POS automation workflow integrates order management, payment processing, inventory updates, and real-time reporting, enabling efficient restaurant operations. Whether you are running a single location or a chain of restaurants, this solution streamlines daily tasks, reduces errors, and provides valuable insights, saving time and improving customer satisfaction. 🚀
by Ranjan Dailata
Who this is for The Real Estate Intelligence Tracker is a powerful automated workflow designed for real estate analysts, investors, proptech startups, and market researchers who need to collect and analyze structured data from real estate listings across the web at scale. This workflow is tailored for: Real Estate Analysts** - Tracking property prices, locations, and market trends Investment Firms** - Sourcing high-opportunity listings for portfolio decisions PropTech Developers** - Automating listing insights for SaaS platforms Market Researchers** - Extracting insights from competitive housing data Growth Teams** - Monitoring geographic property trends and pricing fluctuations What problem is this workflow solving? Collecting structured real estate listing data from property websites is difficult due to bot protections and unstructured HTML content. Manual data collection is slow and error-prone, and traditional scrapers often get blocked or miss context. This workflow solves: Automated bypass of anti-bot protection using Bright Data Web Unlocker Conversion of unstructured HTML content into clean text using a Markdown-to-text LLM pipeline Structured extraction of key listing data like price, location, property type, and features using OpenAI Aggregation and delivery of insights to Google Sheets, local storage, and webhook-based alerts What this workflow does Convert to Text: Transforms scraped HTML/markdown into clean text using a Basic LLM Chain Structured Data Extraction: Uses OpenAI GPT-4o with the Information Extractor node to parse property attributes (price, address, area, type, etc.) Aggregate & Merge: Combines data from multiple pages or listings into a cohesive structure Outbound Data Handling: Google Sheets** – Appends the structured real estate data for further analysis Save to Disk** – Persists structured JSON/text data locally Webhook Notification** – Sends data alerts or summaries to any third-party platform Pre-conditions You need to have a Bright Data account and do the necessary setup as mentioned in the "Setup" section below. You need to have an OpenAI Account. Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. In n8n, Configure the Google Sheet Credentials with your own account. Follow this documentation - Set Google Sheet Credential In n8n, configure the OpenAi account credentials. Ensure the URL and Bright Data zone name are correctly set in the Set URL, Filename and Bright Data Zone node. Set the desired local path in the Write a file to disk node to save the responses. How to customize this workflow to your needs Target Multiple Sites or Locations Update the Bright Data URL node dynamically with a list of regional real estate websites Loop through different city/state filter URLs Customize Extracted Fields Modify the Information Extractor prompt to extract fields like: Property size, number of bedrooms/bathrooms Days on market Nearby amenities or schools Agent contact details Integrate with More Destinations Add nodes to export data to Notion, Airtable, HubSpot, or your custom database Generate automated reports using PDF generators and email them Data Quality and Logging Add validation checks (e.g., missing price or address) Save intermediate files (markdown, raw HTML, JSON output) to disk for audit purposes
by Jimleuk
This n8n template monitors active support issues in Linear.app to track the mood of their ongoing conversation between reporter and assignee using Sentiment Analysis. When sentiment dips into the negative, a notification is sent via Slack to alert the team. How it works A scheduled trigger is used to fetch recently updated issues in Linear using the GraphQL node. Each issue's comments thread is passed into a simple Information Extractor node to identify the overall sentiment. The resulting sentiment analysis combined with the some issue details are uploaded to Airtable for review. When the template is re-run at a later date, each issue is re-analysed for sentiment Each issue's new sentiment state is saved to the airtable whilst its previous state is moved to the "previous sentiment" column. An Airtable trigger is used to watch for recently updated rows Each matching Airtable row is filtered to check if it has a previous non-negative state but now has a negative state in its current sentiment. The results are sent via notification to a team slack channel for priority. Check out the sample Airtable here: https://airtable.com/appViDaeaFw4qv9La/shrq6HgeYzpW6uwXL How to use Modify the GraphQL filter to fetch issues to a relevant issue type, team or person. Update the Slack channel to ensure messages are sent to the correct location or persons. The Airtable also serves to give a snapshot of Sentiment across support tickets for a given period. It's possible to use this to assess the daily operations. Requirements Linear for issue tracking (but feel free to use another system if preferred) Airtable for Database OpenAI for LLM and Sentiment Analysis Customising the workflow Add more granular levels of sentiment to reduce the number of alerts. Explore different types of sentiment based on issue types and customer types. This may help prioritise alerts and response. Run across teams or categories of issues to get an overview of sentiment across the support organisation.
by Davide
This automated workflow takes a static image and a textual prompt and transforms them into an animated video using the MiniMax Hailuo 02 model. It then uploads the generated video to YouTube and TikTok, and updates a Google Sheet with relevant links and metadata. Benefits of This Workflow Fully Automated Pipeline**: From prompt to video to social media publication — all without manual intervention. Scalable Content Creation**: Generate and distribute dozens of videos per hour with minimal human input. Cross-Platform Posting: Automatically pushes content to **YouTube and TikTok simultaneously. SEO Optimization**: Uses AI to generate catchy, keyword-rich video titles that improve visibility. Easy Integration**: Based on Google Sheets for input/output, making it accessible to non-technical users. Time-Efficient**: Batch-processing enabled with scheduled runs every few minutes. Customizable Duration**: Video duration can be adjusted (default is 6 seconds). How It Works Trigger & Data Fetching: The workflow starts either manually or via a scheduled trigger (e.g., every 5 minutes). It checks a Google Sheet for new entries where the "VIDEO" column is empty, indicating pending video generation tasks. Video Creation: For each entry, the workflow extracts the image URL and prompt from the Google Sheet. It sends these inputs to the MiniMax Hailuo 02 to generate a video. The API processes the image and prompt, optimizes the prompt, and creates a short video (default: 6 seconds). Status Monitoring: The workflow polls the API every 60 seconds to check if the video is COMPLETED. Once ready, it retrieves the video URL and uploads the file to Google Drive. YouTube & TikTok Upload: The video is sent to YouTube and TikTok via the Upload-Post.com API (The free plan allows uploads to all platforms except TikTok. To enable, upgrade to a paid plan.). A GPT-generated SEO-optimized title is created for the video. The Google Sheet is updated with the video URL and YouTube link. Set Up Steps Google Sheet Setup: Create a Google Sheet with columns: IMAGE (input image URL), PROMPT (video description), VIDEO (auto-filled), and YOUTUBE_URL (auto-filled). Link the sheet to the workflow using the Google Sheets node. API Keys: Obtain a fal.run API key (for MiniMax Hailuo) and configure the "Authorization" header in the "Create video" node. Get an Upload-Post.com API key (10 free uploads/month) and set it in the "Upload on YouTube/TikTok" nodes. Workflow Configuration: Replace YOUR_USERNAME in the Upload-Post nodes with your profile name (e.g., "test1"). Adjust the video duration (6 or 10 seconds) in the "Create video" node. Set the Schedule Trigger interval (e.g., 5 minutes) to automate checks for new tasks. Execution: Run the workflow manually or let the scheduler process new rows automatically. The system handles video generation, uploads, and Google Sheet updates end-to-end. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Niklas Hatje
Use Case This workflow is beneficial when you're automatically adding new leads to your Pipedrive CRM. Usually, you'd have to manually review each lead to determine if they're a good fit. This process is time-consuming and increases the chances of missing important leads. This workflow ensures every new lead is promptly evaluated upon addition. What this workflow does The workflow runs every 5 minutes. On every run, it checks your new Pipedrive leads and enriches them with Clearbit. It then marks items as enriched and checks if the company of the new lead matches certain criteria (in this case if they are B2B and have more than 100 employees) and sends a Slack alert to a channel for every match. Pre Conditions You must have Pipedrive, Clearbit, and Slack accounts. You also need to set up the custom fields Domain and Enriched at in Pipedrive. Setup Go to Company Settings -> Data fields -> Organization and add Domain as a custom field Go to Company Settings -> Data fields -> Leads and add Enriched at as a custom date field Add your Pipedrive, Clearbit and Slack credentials. Fill the setup node below. To get the ID of your custom domain fields, simply run the Show only custom organization fields and Show only custom lead fields nodes below and copy the keys of your domain, and enriched at fields. How to adjust this workflow to your needs Modify the criteria to suit your definition of an interesting lead. If you only want to focus on interesting leads in Pipedrive, add a node that archives all others. This workflow was built using n8n version 1.29.1