by Oneclick AI Squad
This guide walks you through setting up an AI-driven workflow to automate flight and hotel reservation processes using a conversational travel booking system. The workflow accepts booking requests, processes them via APIs, and sends confirmations, enabling a seamless travel booking experience. What’s the Goal? Automatically accept and process booking requests for flights and hotels via HTTP POST. Use AI to understand natural language requests and route them to appropriate data processors. Search for flights and hotels using external APIs and process booking confirmations. Send confirmation emails and return structured booking data to users. Enable an automated system for efficient travel reservations. By the end, you’ll have a self-running system that handles travel bookings effortlessly. Why Does It Matter? Manual booking processes are time-consuming and prone to errors. This workflow offers: Zero Human Error**: AI ensures accurate request parsing and booking processing. Time-Saving Automation**: Automates the entire booking lifecycle, boosting efficiency. Seamless Confirmation**: Sends automated emails and responses without manual intervention. Enhanced User Experience**: Provides a conversational interface for bookings. Think of it as your reliable travel booking assistant that keeps the process smooth and efficient. How It Works Here’s the step-by-step flow of the automation: Step 1: Trigger the Workflow Webhook Trigger**: Accepts incoming booking requests via HTTP POST, initiating the workflow. Step 2: Parse the Request AI Request Parser**: Uses AI to understand natural language booking requests (e.g., flight or hotel) and extracts relevant details. Step 3: Route Booking Type Booking Type Router**: Determines whether the request is for a flight or hotel and routes it to the respective data processor. Step 4: Process Flight Data Flight Data Processor**: Handles flight-specific data and prepares it for the search API. Step 5: Search Flight API Flight Search API**: Searches for available flights based on parameters (e.g., https://api.aviationstack.com) and returns results. Step 6: Process Hotel Data Hotel Data Processor**: Handles hotel-specific data and prepares it for the search API. Step 7: Search Hotel API Hotel Search API**: Searches for available hotels based on parameters (e.g., https://api.booking.com) and returns results. Step 8: Process Flight Booking Flight Booking Processor**: Processes flight bookings and generates confirmation details. Step 9: Process Hotel Booking Hotel Booking Processor**: Processes hotel bookings and generates confirmation details. Step 10: Generate Confirmation Message Confirmation Message Generator**: Creates structured confirmation messages for the user. Step 11: Send Confirmation Email Send Confirmation Email**: Sends booking confirmation via email to the user. Step 12: Send Response Send Response**: Returns structured booking data to the user, completing the workflow. How to Use the Workflow? Importing the workflow in n8n is a straightforward process. Follow these steps to import the Conversational Travel Booker workflow: Download the Workflow: Obtain the workflow file (e.g., JSON export from n8n). Open n8n: Log in to your n8n instance. Import Workflow: Navigate to the workflows section, click "Import," and upload the workflow file. Configure Nodes: Adjust settings (e.g., API keys, webhook URLs) as needed. Execute Workflow: Test and activate the workflow to start processing bookings. Requirements n8n account and instance setup. Access to flight and hotel search APIs (e.g., Aviationstack, Booking.com). Email service integration for sending confirmations. Webhook URL for receiving booking requests. Customizing this Workflow Modify the AI Request Parser to handle additional languages or booking types. Update API endpoints in Flight Search API and Hotel Search API nodes to match your preferred providers. Adjust the Send Confirmation Email node to include custom email templates or additional recipients. Schedule the Webhook Trigger to align with your business hours or demand peaks.
by Cristian Tala Sánchez
✨ SEO Blog Post Automation with Perplexity, GPT, Leonardo AI & WordPress This workflow automates the creation and publishing of weekly SEO-optimized blog posts using AI and publishes them directly to WordPress — with featured images and tracking in Google Sheets. 🧠 Who is this for This automation is ideal for: Startup platforms and tech blogs Content creators and marketers Solopreneurs who want consistent blog output Spanish-speaking audiences focused on startup trends ⚙️ What it does ⏰ Runs every Monday at 6:00 AM via CRON 📡 Uses Perplexity AI to research trending startup topics 📝 Generates a 1000–1500 word article with GPT in structured HTML 🎨 Creates a cinematic blog image using Leonardo AI 🖼️ Uploads the image to WordPress with alt text and SEO-friendly filename 📰 Publishes the post in a pre-defined category 📊 Logs the post in Google Sheets for tracking 🚀 How to set it up Connect your credentials: Perplexity API OpenAI (GPT-4.1 Mini or similar) Leonardo AI (Bearer token) WordPress (Basic Auth) Google Sheets (OAuth2) Customize your content: Adjust the prompt inside the HTTP node to fit your tone or focus Change the WordPress category ID Update scheduling if you want a different publishing day Test the workflow manually to ensure all steps function correctly 💡 Pro tips Add Slack or email nodes to get notified when a post goes live Use multiple categories or RSS feeds for content diversification Adjust GPT prompt to support different languages or tones Add post-validation rules if needed before publishing 🎯 Why this matters This workflow gives you a full editorial process on autopilot: research, writing, design, publishing, and tracking — all powered by AI. No more blank pages or manual posting. Use it to scale your content strategy, boost your SEO, and stay relevant — 100% hands-free.
by Rahi
Workflow 1: Domain and Email Health 🩺 This part of the workflow is triggered every 5 hours by the Schedule Trigger1 node. Its purpose is to pull health metrics for both email domains and individual email addresses. How it Works: ++Schedule Trigger:++ The Schedule Trigger1 node initiates the workflow every 5 hours. ++API Requests:++ Two separate HTTP Request nodes, HTTP Request5 and HTTP Request6, make API calls to Smartlead. ++HTTP Request5 calls++ the endpoint for domain-wise health metrics. ++HTTP Request6 calls++ the endpoint for email-wise health metrics. Both requests use the same api_key and a date range from 2025-07-04 to the current day. ++Data Splitting:++ The Split Out5 and Split Out6 nodes take the JSON response from the API calls and split the data into individual items. This is necessary so each row of data can be processed and added to Google Sheets separately. ++Google Sheets Integration:++ Finally, the Append or update row in sheet5 and Append or update row in sheet6 nodes update two different Google Sheets: ++Append or update row in sheet5 adds++ or updates rows in the DomainHealth sheet, matching on the domain column. ++Append or update row in sheet6 adds++ or updates rows in the EmailHealth sheet, matching on the from_email column. Workflow 2: Global and Campaign-Specific Analytics 📊 This second part of the workflow is triggered every 2 hours by the Schedule Trigger node. Its goal is to get a day-by-day overview of email engagement and campaign-specific performance. How it Works: Schedule Trigger: The Schedule Trigger node starts this workflow every 2 hours. ++API Requests:++ Two HTTP Request nodes, HTTP Request and HTTP Request1, call different Smartlead API endpoints. ++HTTP Request++ retrieves day-wise overall stats for email engagement. ++HTTP Request1 ++retrieves overall stats for each campaign. ++Data Splitting:++ The Split Out and Split Out1 nodes separate the JSON responses into individual data items for processing. ++Google Sheets Integration:++ The Append or update row in sheet and Append or update row in sheet1 nodes then write the data to Google Sheets. ++Append or update row in sheet++ updates the Sheet1 sheet with day-wise metrics, using the date as a matching column. ++Append or update row in sheet1++ updates the CampaignWise sheet with campaign performance metrics, using the campaign id to match rows.
by Yaron Been
Description This workflow automatically monitors online forums for specific keywords, topics, or competitor mentions. It helps you stay on top of relevant discussions without manually checking multiple forums throughout the day. Overview This workflow automatically monitors selected forums for new posts, keywords, or competitor activity. It uses Bright Data to scrape forum content and can notify you or save results to a spreadsheet or other service. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping forum content without getting blocked. (Optional) Google Sheets/Discord/Telegram:** For notifications or data storage. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Notifications (Optional): Configure notification/storage nodes as needed. Customize: Specify the forums and keywords to monitor. Use Cases Brand Monitoring:** Stay updated on mentions of your brand or product. Competitor Tracking:** Monitor competitor activity in your industry forums. Community Managers:** Get alerts for new threads or topics. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #forums #brightdata #webscraping #monitoring #forummonitoring #brandmentions #competitortracking #keywordtracking #n8nworkflow #workflow #nocode #communitymanagement #onlineforums #discussionboards #brandawareness #marketresearch #socialmediamonitoring #contentmonitoring #reputationmanagement #digitalmarketing #businessintelligence #onlinediscussions #competitoranalysis
by DIGITAL BIZ TECH
Knowledge RAG & AI Chat Agent: Google Drive to Qdrant Description This workflow transforms a Google Drive folder into an intelligent, searchable knowledge base and provides a chat agent to query it. It’s composed of two distinct flows: An ingestion pipeline to process documents. A live chat agent that uses RAG (Retrieval-Augmented Generation) and optional web search to answer user questions. This system fully automates the creation of a “Chat with your docs” solution and enhances it with external web-searching capabilities. Quick Implementation Steps Import the workflow JSON into your n8n instance. Set up credentials for Google Drive, Mistral AI, OpenAI, and Qdrant. Open the Web Search node and add your Tavily AI API key to the Authorization header. In the Google Drive (List Files) node, set the Folder ID you want to ingest. Run the workflow manually once to populate your Qdrant database (Flow 1). Activate the workflow to enable the chat trigger (Flow 2). Copy the public webhook URL from the When chat message received node and open it in a new tab to start chatting. What It Does The workflow is divided into two primary functions: 1. Knowledge Base Ingestion (Manual Trigger) This flow populates your vector database. Scans Google Drive:** Lists all files from a specified folder. Processes Files Individually:** Downloads each file. Extracts Text via OCR:* Uses *Mistral AI OCR API** for text extraction from PDFs, images, etc. Generates Smart Metadata:** A Mistral LLM assigns metadata like document_type, project, and assigned_to. Chunks & Embeds:* Text is cleaned, chunked, and embedded via *OpenAI’s text-embedding-3-small** model. Stores in Qdrant:** Text chunks, embeddings, and metadata are stored in a Qdrant collection (docaiauto). 2. AI Chat Agent (Chat Trigger) This flow powers the conversational interface. Handles User Queries:** Triggered when a user sends a chat message. Internal RAG Retrieval:* Searches *Qdrant Vector Store** first for answers. Web Search Fallback:* If unavailable internally, the agent offers to perform a *Tavily AI web search**. Contextual Responses:** Combines internal and external info for comprehensive answers. Who's It For Ideal for: Teams building internal AI knowledge bases from Google Drive. Developers creating AI-powered support, research, or onboarding bots. Organizations implementing RAG pipelines. Anyone making unstructured Google Drive documents searchable via chat. Requirements n8n instance** (self-hosted or cloud). Google Drive Credentials** (to list and download files). Mistral AI API Key** (for OCR & metadata extraction). OpenAI API Key** (for embeddings and chat LLM). Qdrant instance** (cloud or self-hosted). Tavily AI API Key** (for web search). How It Works The workflow runs two independent flows in parallel: Flow 1: Ingestion Pipeline (Manual Trigger) List Files: Fetch files from Google Drive using the Folder ID. Loop & Download: Each file is processed one by one. OCR Processing: Upload file to Mistral Retrieve signed URL Extract text using Mistral DOC OCR Metadata Extraction: Analyze text using a Mistral LLM. Text Cleaning & Chunking: Split into 1000-character chunks. Embeddings Creation: Use OpenAI embeddings. Vector Insertion: Push chunks + metadata into Qdrant. Flow 2: AI Chat Agent (Chat Trigger) Chat Trigger: Starts when a chat message is received. AI Agent: Uses OpenAI + Simple Memory to process context. RAG Retrieval: Queries Qdrant for related data. Decision Logic: Found → Form answer. Not found → Ask if user wants web search. Web Search: Performs Tavily web lookup. Final Response: Synthesizes internal + external info. How To Set Up 1. Import the Workflow Upload the provided JSON into your n8n instance. 2. Configure Credentials Create and assign: Google Drive** → Google Drive nodes Mistral AI** → Upload, Signed URL, DOC OCR, Cloud Chat Model OpenAI** → Embeddings + Chat Model nodes Qdrant** → Vector Store nodes 3. Add Tavily API Key Open Web Search node → Parameters → Headers Add your key under Authorization (e.g., tvly-xxxx). 4. Node Configuration Google Drive (List Files):** Set Folder ID. Qdrant Nodes:** Ensure same collection name (docaiauto). 5. Run Ingestion (Flow 1) Click Test workflow to populate Qdrant with your Drive documents. 6. Activate Chat (Flow 2) Toggle the workflow ON to enable real-time chat. 7. Test Open the webhook URL and start chatting! How To Customize Change LLMs:** Swap models in OpenAI or Mistral nodes (e.g., GPT-4o, Claude 3). Modify Prompts:** Edit the system message in ai chat agent to alter tone or logic. Chunking Strategy:** Adjust chunkSize and chunkOverlap in the Code node. Different Sources:** Replace Google Drive with AWS S3, Local Folder, etc. Automate Updates:* Add a *Cron** node for scheduled ingestion. Validation:** Add post-processing steps after metadata extraction. Expand Tools:** Add more functional nodes like Google Calendar or Calculator. Use Case Examples Internal HR Bot:** Answer HR-related queries from stored policy docs. Tech Support Assistant:** Retrieve troubleshooting steps for products. Research Assistant:** Summarize and compare market reports. Project Management Bot:** Query document ownership or project status. Troubleshooting Guide | Issue | Possible Solution | |------------|------------------------| | Chat agent doesn’t respond | Check OpenAI API key and model availability (e.g., gpt-4.1-mini). | | Known documents not found | Ensure ingestion flow ran and both Qdrant nodes use same collection name. | | OCR node fails | Verify Mistral API key and input file integrity. | | Web search not triggered | Re-check Tavily API key in Web Search node headers. | | Incorrect metadata | Tune Information Extractor prompt or use a stronger Mistral model. | Need Help or More Workflows? Want to customize this workflow for your business or integrate it with your existing tools? Our team at Digital Biz Tech can tailor it precisely to your use case from automation logic to AI-powered enhancements. 💡 We can help you set it up for free — from connecting credentials to deploying it live. Contact: shilpa.raju@digitalbiz.tech Website: https://www.digitalbiz.tech LinkedIn: https://www.linkedin.com/company/digital-biz-tech/ You can also DM us on LinkedIn for any help.
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically monitors keyword rankings across search engines to track SEO performance and identify optimization opportunities. It saves you time by eliminating the need to manually check keyword positions and provides comprehensive ranking data for strategic SEO decision making. Overview This workflow automatically scrapes search engine results pages (SERPs) to track keyword rankings, competitor positions, and search features. It uses Bright Data to access search results without restrictions and AI to intelligently parse ranking data, track changes, and identify SEO opportunities. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping search engine results without being blocked OpenAI**: AI agent for intelligent ranking analysis and SEO insights Google Sheets**: For storing keyword ranking data and tracking changes How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your ranking tracking spreadsheet Customize: Define target keywords and ranking monitoring parameters Use Cases SEO Teams**: Track keyword performance and identify ranking improvements Content Marketing**: Monitor content ranking success and optimization needs Competitive Analysis**: Track competitor keyword rankings and strategies Digital Marketing**: Measure organic search performance and ROI Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #keywordrankings #seo #searchrankings #brightdata #webscraping #seotools #n8nworkflow #workflow #nocode #ranktracking #keywordmonitoring #seoautomation #searchmarketing #organicseo #seoresearch #rankinganalysis #keywordanalysis #searchengines #seomonitoring #digitalmarketing #serp #keywordtracking #seoanalytics #searchoptimization #rankingreports #keywordresearch #seoinsights #searchperformance
by Nazmy
Bearer Token Validation This n8n template helps you manage and validate tokens easily using: n8n as your backend workflow engine Airtable as your lightweight token store 🚀 What It Does Stores user tokens securely in Airtable with expiry or usage metadata. Validates incoming tokens in your workflows (e.g., webhook APIs). Rejects invalid or expired tokens automatically for security. Can be extended to generate, rotate, or revoke tokens for user management. How It Works Webhook node receives requests with a Bearer header. Airtable Query looks up the provided token. Validation Logic (Code node): Checks if the token exists. Verifies expiry or usage limits if configured. Returns success if valid, or error if error with describing the issue. Note: This is the simplest way to do auth, just for simplification Why Use This No need for a full backend to manage secure token validation. Clean, modular, and ready for your SaaS workflows. Enjoy building secure automations with n8n + Airtable! 🚀 Built by: Nazmy
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically monitors competitor pricing changes and website updates to keep you informed of market movements. It saves you time by eliminating the need to manually check competitor websites and provides alerts only when actual changes occur, preventing information overload. Overview This workflow automatically scrapes competitor pricing pages (like ClickUp) and compares current pricing with previously stored data. It uses Bright Data to access competitor websites without being blocked and AI to intelligently extract pricing information, updating your tracking spreadsheet only when changes are detected. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping competitor websites without being blocked OpenAI**: AI agent for intelligent pricing data extraction and parsing Google Sheets**: For storing and comparing historical pricing data How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your pricing tracking spreadsheet Customize: Set your competitor URLs and pricing monitoring schedule Use Cases Product Teams**: Monitor competitor feature and pricing changes for strategic planning Sales Teams**: Stay informed of competitor pricing to adjust sales strategies Marketing Teams**: Track competitor messaging and positioning changes Business Intelligence**: Build comprehensive competitor analysis databases Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #competitoranalysis #pricingmonitoring #brightdata #webscraping #competitortracking #marketintelligence #n8nworkflow #workflow #nocode #pricetracking #businessintelligence #competitiveanalysis #marketresearch #competitormonitoring #pricingdata #websitemonitoring #competitorpricing #marketanalysis #competitorwatch #pricingalerts #businessautomation #competitorinsights #markettrends #pricingchanges #competitorupdates #strategicanalysis #marketposition #competitiveintelligence
by Grigory Frolov
WordPress Blog to Google Sheets Sync Posts • Categories • Tags • Media 🧩 Overview This n8n workflow automatically syncs your WordPress website content — including posts, categories, tags, and media — into Google Sheets. It helps automate content reporting, SEO analysis, and data backups. The workflow can run on schedule or on demand via a webhook. 💡 Use cases Maintain a live database of blog posts in Google Sheets. Create dashboards in Google Data Studio or Looker Studio. Track new articles for newsletters or social media scheduling. Backup all WordPress content and media outside of your CMS. ⚙️ Prerequisites Before importing the workflow, ensure you have: A WordPress website with the REST API enabled (default in WP 4.7+). Authentication: either Application Passwords or Basic Auth credentials. A Google Sheet with the following tabs: Posts Categories Tags Media The following credentials configured in n8n: HTTP Basic Auth (for WordPress) Google Sheets OAuth2 🚀 Setup instructions Import the workflow into your n8n instance. Replace all example WordPress API URLs with your domain, for example: https://yourdomain.com/wp-json/wp/v2/ Connect your HTTP Basic Auth credentials (WordPress username + Application Password). Connect your Google Sheets OAuth2 account. Update the spreadsheet ID in each Google Sheets node with your own. Adjust the Schedule Trigger (e.g. run daily at 2:00 AM). Run once manually to verify data sync. 🧠 Workflow structure | Section | Description | |----------|--------------| | Schedule / Webhook Trigger | Starts the workflow manually or automatically | | Variables & Loop Vars | Initialize pagination for REST API requests | | Get Posts → Split Out → Update Posts | Fetch and update all WordPress posts | | Get Categories → Update Categories | Sync WordPress categories | | Get Tags → Update Tags | Sync WordPress tags | | Get Media → Split Out → Update Media | Sync media library (images, videos, etc.) | | IF Loops | Handles pagination logic until all items are retrieved | ⚠️ Notes & Limitations Works with standard WordPress REST API endpoints only. Custom post types require editing endpoint URLs. The per_page value defaults to 10; increase for faster syncs. For large sites, consider increasing n8n memory or adding execution logs. Avoid running the workflow too frequently to prevent API rate limits. 🎥 Video Tutorial A step-by-step setup guide is available here: 👉 https://www.youtube.com/watch?v=czSMWyD6f-0 Please subscribe to my YouTube channel to support me: 👉 https://www.youtube.com/@gregfrolovpersonal 👨💻 Author Created by: Grigory Frolov SEO & Automation Specialist — helping businesses integrate WordPress, AI, and data tools with n8n. 🧾 License This workflow is provided under the MIT License. Feel free to use, modify, and share improvements with the community.
by Markhah
Overview This n8n workflow is a modular AI analyst system that provides real-time insights from CoinMarketCap’s centralized and decentralized data sources. Using GPT-based AI, the system interprets natural language questions about the crypto market and delegates them to specialized agent workflows. It supports Telegram chat input and returns structured results such as coin quotes, DEX liquidity, exchange info, and community sentiment—all integrated from the CoinMarketCap API ecosystem. Prerequisites a. OpenAI or Gemini account (via GPT-4o-mini or equivalent LLM). b. Telegram Bot API token (for message input/output). c. Valid CoinMarketCap API key. 📦 Required subflows: CoinMarketCap_Crypto_Agent_Tool CoinMarketCap_Exchange_and_Community_Agent_Tool CoinMarketCap_DEXScan_Agent_Tool d. All tools must be installed and configured before use. Each one acts as a specialized endpoint wrapper for CoinMarketCap APIs. How It Works Telegram Input Users send a query to the bot (e.g. “Top DEX pairs on Ethereum”). Session Memory & Agent Brain Session is tracked via chat.id GPT-4o-mini interprets the query, routes to sub-agents Sub-Agent Workflows Crypto Agent: prices, rankings, conversions Exchange Agent: community sentiment, exchange token holdings DEX Agent: OHLCV data, liquidity pools, trades Multi-Agent Coordination AI can combine queries across tools (e.g., get token ID → fetch quote → analyze liquidity) Ensures valid parameters and avoids API errors Telegram Output Final analysis is sent back to the user as a formatted message. Troubleshooting Tips Error Code Meaning Fix 400 Bad request Check symbol/slug/ID validity 401 Unauthorized Verify CoinMarketCap API key 429 Rate limit exceeded Throttle or upgrade API tier 500 Server error Retry with backoff or report to CMC Example Telegram Queries “Show me top 5 coins by market cap” “Get price of ETH on Uniswap and Binance” “How much liquidity is in SOL-USDC pair?” “Fear & Greed Index and trending tokens” SEO Tags (ẩn hoặc ghi chú riêng): coinmarketcap, n8n crypto analyst, crypto ai telegram bot, dex liquidity, CMC price tracker, gpt-4o crypto market, token sentiment dashboard, fear and greed index
by Ranjan Dailata
Description This workflow automates the process of scraping the latest discussions from HackerNews, transforming raw threads into human readable content using Google Gemini, and exporting the final content into a well-formatted Google Doc. Overview This n8n workflow is responsible for extracting trending posts from the HackerNews API. It loops through each item, performs HTTP data extraction, utilizes Google Gemini to generate human-readable insights, and then exports the enriched content into Google Docs for distribution, archiving, or content creation. Who this workflow is for Tech Newsletter Writers**: Automate the collection and summarization of trending HackerNews posts for inclusion in weekly or daily newsletters. Content Creators & Bloggers**: Quickly generate structured summaries and insights from HackerNews threads to use as inspiration or supporting content for blog posts, videos, or social media. Startup Founders & Product Builders**: Monitor HackerNews for discussions relevant to your niche or competitors, and keep a pulse on the community’s opinions. Investors & Analysts**: Surface early signals from the tech ecosystem by identifying what’s trending and how the community is reacting. Researchers & Students**: Analyze popular discussions and emerging trends in technology, programming, and startups—enriched with AI-generated insights. Digital Agencies & Consultants**: Offer HackerNews monitoring and insight reports as a value-added service to clients interested in the tech space. Tools Used n8n**: The core automation engine that manages the trigger, transformation, and export. HackerNews API**: Provides access to trending or new HN posts. Google Gemini**: Enriches HackerNews content with structured insights and human-like summaries. Google Docs**: Automatically creates and updates a document with the enriched content, ready for sharing or publishing. How to Install Import the Workflow**: Download the .json file and import it into your n8n instance. Set Up HackerNews Source**: Choose whether to use the HN API (via HTTP Request node) or RSS Feed node. Configure Gemini API**: Add your Google Gemini API key and design the prompt to extract pros/cons, key themes, or insights. Set Up Google Docs Integration**: Connect your Google account and configure the Google Docs node to create/update a document. Test and Deploy**: Run a test job to ensure data flows correctly and outputs are formatted as expected. Use Cases Tech Newsletter Authors**: Generate ready-to-use summaries of trending HackerNews threads. Startup Founders**: Stay informed on key discussions, product launches, and community feedback. Investors & Analysts**: Spot early trends, technical insights, and startup momentum directly from HN. Researchers**: Track community reactions to new technologies or frameworks. Content Creators**: Use the enriched data to spark blog posts, YouTube scripts, or LinkedIn updates. Connect with Me Email: ranjancse@gmail.com LinkedIn: https://www.linkedin.com/in/ranjan-dailata/ Get Bright Data: Bright Data (Supports free workflows with a small commission) #n8n #automation #hackernews #contentcuration #aiwriting #geminiapi #googlegemini #techtrends #newsletterautomation #googleworkspace #rssautomation #nocode #structureddata #webscraping #contentautomation #hninsights #aiworkflow #googleintegration #webmonitoring #hnnews #aiassistant #gdocs #automationtools #gptlike #geminiwriter
by Lucas Perret
This workflow monitor G2 reviews URLS. When a new review is published, it will: trigger a Slack notification record the review in Google Sheets To install it, you'll need: access to Slack, Google Sheets and ScrapingBee Full guide here: https://lempire.notion.site/Scrape-G2-reviews-with-n8n-3f46e280e8f24a68b3797f98d2fba433?pvs=4