by Ranjan Dailata
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Description This workflow automates the creation of structured eBooks by generating chapters, table of contents, and content using Google Gemini Flash 2.0. Overview This n8n workflow allows users to input a topic or outline, which is then processed by Google Gemini Flash 2.0 to generate chapter titles, structured table of contents, and detailed section-wise content. The final output is formatted and exported into a Google Document, ready for review and further publishing. Who This Workflow Is For Authors & Writers** Save time by auto-generating chapter ideas, summaries, and full-length content based on a topic or outline—great for fiction and nonfiction alike. Content Marketers** Rapidly create downloadable eBooks, whitepapers, or lead magnets for campaigns without relying on long production cycles. Educators & Course Creators** Convert your syllabus, course modules, or learning outcomes into structured, well-formatted educational eBooks. Agencies & Freelancers** Offer AI-powered eBook creation as a value-added service to clients in need of fast, professional content. Entrepreneurs & Coaches** Turn your knowledge, frameworks, or training material into publish-ready books to promote your brand or monetize content. Technical Writers & Documentarians** Generate structured documentation or guides from outlines, simplifying the technical writing process with the help of AI. Tools Used n8n**: Orchestrates input handling, AI processing, formatting, and export. Google Gemini Flash 2.0**: Generates high-quality, structured content, including chapters, summaries, and body text. Google Docs**: Used to compile and format the full eBook in a collaborative document. Google Drive / Email**: Optional nodes for storing or delivering the final output. How to Install Import the Workflow**: Download and import the .json file into your n8n instance. Configure Gemini Flash 2.0**: Add your API key and set the desired creativity, length, and tone options. Provide Input**: Use a webhook or manual trigger to define the eBook topic or structure. Customize Format**: Modify prompts or Gemini instructions to match your eBook format, voice, or domain (e.g., fiction, business, technical). Export to Google Docs**: Authenticate and configure the Google Docs node to write the output chapter-wise into a new or existing document. Optional Distribution**: Connect to Google Drive or Gmail to store or send the final eBook. Use Cases Writers & Authors**: Quickly draft entire eBooks based on minimal input. Marketers**: Generate lead magnets, guides, and product documentation at scale. Educators**: Produce structured learning materials or course eBooks. Agencies**: Offer eBook creation services powered by AI. Entrepreneurs**: Turn knowledge into content assets without hiring ghostwriters. Connect with Me Email: ranjancse@gmail.com LinkedIn: https://www.linkedin.com/in/ranjan-dailata/ Get Bright Data: Bright Data (Supports free workflows with a small commission) #n8n #automation #ebookcreation #googleai #geminiflash #aiwriting #gdocs #contentautomation #ebookworkflow #nocode #contentmarketing #gemini #aiwriter #automatedpublishing #aicontent #bookcreation #geminiworkflow #ebookgenerator #gptalternative #flash20 #geminiflash2 #authorautomation #educationalcontent #aiinmarketing #n8nworkflow
by Raz Hadas
Description Transform your investment strategy with a fully automated, AI-driven trading bot. This workflow bridges the gap between AI-powered market insights and real-world trading by executing buy and sell orders directly through the Alpaca paper trading API. Designed to work in tandem with the Automated Stock Sentiment Analysis workflow, this solution takes the top-performing stocks based on daily news sentiment and automatically rebalances your portfolio. It's perfect for algorithmic traders, data-driven investors, and n8n enthusiasts who want to see their AI analysis translate into tangible actions, all while maintaining a comprehensive log of every transaction in Google Sheets. Key Features & Benefits Automated Trading Execution:** Automatically places buy and sell orders on the Alpaca paper trading platform without manual intervention. Sentiment-Driven Decisions:** Leverages the output from the sentiment analysis workflow to make informed decisions, selling positions with waning sentiment and buying into those with high positive sentiment. Dynamic Portfolio Rebalancing:** Intelligently calculates which positions to close and how to allocate the resulting funds into new, high-potential stocks. Paper Trading Ready:** Safely test and refine your trading strategies in a risk-free environment using Alpaca's paper trading API. Daily Performance Tracking:** Automatically logs your account equity and daily percentage change to a Google Sheet, giving you a clear view of your portfolio's performance. Detailed Trade Logging:** Every buy and sell order is meticulously recorded in a Google Sheet for easy review and historical analysis. Scheduled and Autonomous:** The entire process runs on a daily schedule, making it a "set and forget" solution for systematic trading. How It Works This workflow executes a sophisticated, automated trading strategy in a few key stages: Daily Kick-off & Snapshot: The workflow triggers on a daily schedule, first fetching your current Alpaca account balance and logging it to a Google Sheet to track daily performance. Strategy Formulation: It then reads the daily sentiment scores produced by the accompanying "Stock Sentiment Analysis" workflow. A Code node filters these results to identify the top four stocks with the highest positive sentiment. The Decision Engine: The core of the workflow is a custom Code node that acts as the trading brain. It: Retrieves your currently open positions from Alpaca. Compares your holdings against the day's top four sentiment stocks. Generates a "sell list" of positions you hold that are no longer in the top four. Generates a "buy list" of top-sentiment stocks that you don't yet own. Calculates the total cash value from the "sell list" and determines the exact notional value to invest in each stock on the "buy list." Trade Execution: The workflow first iterates through the "sell list" and executes a DELETE request to Alpaca for each, closing the positions. A Wait node pauses the workflow for two minutes to ensure the sell orders are filled and the account balance is updated. It then iterates through the "buy list," executing POST requests to Alpaca to purchase the new assets with the calculated funds. Record Keeping: All executed orders (both buys and sells) are merged and logged in a dedicated Google Sheet, giving you a permanent and detailed transaction history. Nodes Used Schedule Trigger HttpRequest (Alpaca API) Google Sheets Code (JavaScript) SplitOut Wait Merge This workflow is the perfect next step for anyone looking to take their AI analysis to the next level. Take the emotion out of your trading and let this bot systematically execute your data-driven strategy.
by Humble Turtle
Github Deployer Agent Overview The Github Deployer Agent is an intelligent automation tool that integrates with Slack to streamline code deployment workflows. Powered by Anthropic's Claude 3.5 and Tavily for web search, it enables seamless, context-aware file pushes to a GitHub repository with minimal user input. Capabilities Accepts natural language via Slack Automatically pushes code to a default GitHub repository Uses Claude 3.5 for code generation and decision-making Leverages Tavily for real-time web search to enhance context Supports folder structure hints to ensure clean and organized repositories Required Connections To operate correctly, the following integrations must be in place: Slack API Token with permission to read messages and post responses GitHub Personal Access Token with repo write permissions Tavily API Key for external search functionality Claude 3.5 API Access via Anthropic Detailed configuration instructions are provided in the workflow Example Input From Slack, you can send messages like: "Generate a basic README.md for my Python project and store it in the root directory." Customising This Workflow You can tailor the workflow by: Modifying default folder paths or repository settings Integrate Jira node to use issue keys as default folder naming Add slack file upload option
by kenandrewmiranda
An automated n8n workflow that analyzes stocks using RSI and MACD, summarizes insights with OpenAI, and sends a Slack-ready market update every hour. This workflow: Runs hourly from 6:30 AM to 2:30 PM PT, Mon–Fri Checks if the U.S. stock market is open using Alpaca’s /clock API Pulls daily stock bars for a list of tickers via Alpaca’s /v2/stocks/bars Calculates RSI and MACD using a Python code node Categorizes each stock as Buy / Hold / Sell Uses OpenAI Assistant to summarize the results in Slack markdown Sends the message to a specific Slack user or channel
by Cooper
Chat with thing This n8n template lets you build a smart AI chat assistant that can handle text, images, and PDFs — using OpenAI's GPT-4o multimodal model. It supports dynamic conversations and file analysis, making it great for AI-driven support bots, personal assistants, or embedded chat widgets. 🔍 How it Works The chat trigger node kicks off a session using n8n's hosted chat UI. Users can send text or upload images or PDFs — the workflow checks if a file was included. If an image is uploaded, the file is converted to base64 and analyzed using GPT-4o's vision capabilities. GPT-4o generates a natural language description of the image and responds to the user's question in context. A memory buffer keeps track of the conversation thread, so follow-up questions are handled intelligently. OpenAI’s chat model handles both text-only and mixed media input seamlessly. 🧪 How to Use You can embed this in a website or use it with your own webhook/chat interface. The logic is modular — just swap out the chatTrigger node for another input (e.g. form or API). To use with documents, you can modify the logic to pass PDF content to GPT-4 directly. You can extend it with action nodes, e.g. saving results to Notion, Airtable, or sending replies via email or Slack. 🔐 Requirements Your OpenAI GPT-4o API key Set File Upload on the chat 🚀 Use Cases PDF explainer bot Internal knowledge chat with media support Personal assistant for mixed content
by Milan Vasarhelyi - SmoothWork
Video Introduction Want to automate your inbox or need a custom workflow? 📞 Book a Call | 💬 DM me on Linkedin Get Alerted When Your Website Is Down – n8n as an Uptime Robot Alternative If you manage a website (or client sites), getting notified when your site goes down is critical. But you’ve probably experienced alert fatigue if you’ve ever received a flood of downtime emails for every tiny outage. These short hiccups are common, and most monitoring tools don’t let you filter them out. Here’s how you can set up website downtime notifications with n8n - no extra subscriptions, and no more false positives spamming your inbox. For your most important sites, you can even get direct text alerts to your phone. We’ll use n8n, a powerful automation tool with thousands of templates and integrations. You can run it in the cloud or even self-host it for free. Why Use n8n for Website Monitoring? Uptime Robot and similar services have limitations or get expensive fast if you monitor several sites. n8n gives you full control - you choose when and how you get notified, set your own timing and thresholds, and you aren’t stuck with default alert logic. Plus, n8n can automate much more than just uptime monitoring: use it to handle other business workflows too. Quick Start: Free n8n Website Monitoring Template Use the template to get started fast. Log in or sign up for n8n Cloud, or self-host for free if you prefer. Configure your schedule (default: hourly) and list the sites you want to monitor. Key setting: wait seconds. Recommended: 300 seconds (5 minutes). If your site goes down, the workflow waits before sending an alert—no notification for a 1-2 minute outage, only real, persistent problems. How to Test & Use Activate the workflow. Toggle it on in n8n and it’ll check your sites automatically. Test instantly: Add a non-existent URL and run the workflow. After the wait time, you’ll get an email alert for that URL. Notifications stay organized: Alerts go straight to your inbox (see my AI email labelling template if you want color-coded organization). Get Critical Alerts on Your Phone (Telegram) Email is great, but for your most important sites you want instant notifications on your phone. Best option: Use Telegram. n8n can send you messages via a Telegram bot—easy to set up, free, and works in seconds. How: Create a Telegram bot via BotFather (instructions in the template). Add your bot token and chat ID to n8n. Now, you’ll get downtime alerts right on Telegram—no missed notifications, no extra costs. FAQ Can I monitor unlimited sites?** Yes, just list them all in the config. What if downtime is just a few seconds?** The default 5-minute wait filters these out. Do I need a paid n8n plan?** No - self-hosting is free, and the workflow works on free plans. Why not SMS or WhatsApp?** Telegram bots are fastest, easiest, and don’t require extra API setup or subscriptions.
by Jonathan | NEX
Are you drowning in a sea of security notifications? Do your analysts spend more time sifting through low-level logs than investigating real threats? This workflow transforms n8n into an autonomous SOC (Security Operations Center) Analyst, tackling alert fatigue head-on. Leveraging the NixGuard Security RAG connector, this workflow automates the entire alert triage process. It ingests raw security events (from sources like Wazuh, your SIEM, or EDR), uses AI to analyze and assign a priority, and then intelligently routes the alert to the correct Slack channel. How It Works: Ingest & Filter: The workflow runs on a schedule, fetching all recent security alerts. It first performs a basic filtering to isolate events that meet a minimum severity threshold (e.g., level 7+). AI Analysis & Prioritization: The aggregated high-severity alerts are then sent to the AI with a specific prompt, asking it to analyze the situation and return a structured JSON object containing a single, overall priority (Critical, High, Info) and a concise summary. Intelligent Routing: A Switch node reads the AI-assigned priority and routes the notification to the appropriate destination. Critical alerts go to your #security-incident-response channel, high-priority alerts to #security-investigations, and informational ones to #security-logs. Key Features & Benefits: Eliminate Alert Fatigue:** Drastically reduce the noise by having AI pre-process and categorize alerts before they hit your team. Automate SOC Tier 1 Triage:** Free up your human analysts from repetitive triage tasks so they can focus on high-value investigation and threat hunting. Faster Incident Response:** Route critical alerts to the right people in real-time, cutting down on crucial response time. Consistent Prioritization:** Use AI to ensure a consistent, unbiased approach to alert prioritization, 24/7. Smart Routing Logic:** Go beyond simple keyword matching. The Switch node ensures alerts are delivered to the team best equipped to handle them based on AI-assessed severity. Who is this for? SOC Analysts & Security Engineers** looking to automate alert triage and incident response workflows. SecOps and DevOps Teams** who want to build a more efficient, automated security operations pipeline. IT Managers and Directors** aiming to improve their team's efficiency and reduce the risk of missing critical alerts. Anyone using Wazuh, a SIEM, or other security tools that generate a high volume of alerts. Stop manually triaging alerts. Install this workflow to build your own AI-powered security automation platform and let your team focus on what matters most. Don't have the main workflow yet? Get it HERE! 🔗 Learn more about NixGuard: thenex.world 🔗 Get started with a free security subscription: thenex.world/security/subscribe Tags / Keywords: AI, Security, SOC, Automation, Triage, Alerting, Cybersecurity, Wazuh, SIEM, Slack, Incident Response, Alert Fatigue, SecOps, Generative AI, LLM, NixGuard, Routing
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically analyzes competitor backlink profiles to understand their link building strategies and identify opportunities for your own SEO efforts. It saves you time by eliminating the need to manually research competitor links and provides detailed insights into their most valuable linking relationships. Overview This workflow automatically scrapes backlink analysis tools and competitor websites to extract comprehensive backlink data including referring domains, anchor text, link quality metrics, and link acquisition patterns. It uses Bright Data to access backlink databases and AI to intelligently analyze competitor link strategies. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping backlink analysis platforms without being blocked OpenAI**: AI agent for intelligent backlink strategy analysis Google Sheets**: For storing competitor backlink data and insights How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your backlink analysis spreadsheet Customize: Define competitor domains and backlink analysis parameters Use Cases SEO Strategy**: Learn from competitor link building success and replicate strategies Link Prospecting**: Identify websites that link to competitors but not to you Competitive Intelligence**: Understand competitor SEO strategies and authority sources Link Building**: Find high-quality link opportunities in your industry Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #competitorbacklinks #backlinkanalysis #seo #linkbuilding #brightdata #webscraping #competitoranalysis #n8nworkflow #workflow #nocode #linkanalysis #backlinkresearch #seoanalysis #competitiveintelligence #linkresearch #seostrategy #backlinkmonitoring #linkprospecting #domainanalysis #seotools #backlinkaudit #linkbuilding #organicseo #searchmarketing #competitorresearch #linkstrategy #seocompetitor #backlinkinsights
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically analyzes online data sources to uncover unmet customer needs and product opportunities. Instead of spending days on manual research, you receive an AI-driven report of market gaps directly in your inbox. Overview The automation aggregates product reviews, forum discussions, and competitor feature lists via Bright Data scraping. OpenAI then clusters complaints and feature requests to highlight where demand is high but supply is low. A summary table is generated in Google Sheets and a concise report is emailed to stakeholders. Tools Used n8n** – Automates the entire pipeline Bright Data** – Scrapes review sites and competitor pages reliably OpenAI** – Performs NLP clustering and gap analysis Google Sheets** – Stores structured opportunity data Gmail** – Delivers the weekly insight report How to Install Import the Workflow: Add the .json to your n8n instance. Configure Bright Data: Insert your credentials in the MCP Client node. Set Up OpenAI: Provide your OpenAI API key. Connect Google Sheets & Gmail: Authorize both services. Customize Keywords & Sources: Edit the Set node to focus on your industry and competitors. Use Cases Product Management**: Prioritize roadmap features that address true pain points. Startup Ideation**: Validate gaps before building an MVP. Competitive Strategy**: Identify areas where rivals are weak. Investor Research**: Spot underserved niches for portfolio companies. Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #marketanalysis #productmanagement #brightdata #openai #marketgaps #n8nworkflow #nocode #competitiveanalysis
by Mario
Template to get your public IP address and push it to Namecheaps Dynamic DNS per subdomain. Open "yourdomain.com" Insert your domain and your Namecheap DDNS password Open "subdomains" Replaces and insert your subdomains Execute Workflow Have fun!
by Yaron Been
This workflow automatically monitors competitor pricing across multiple products and services to track market positioning and pricing strategies. It saves you time by eliminating the need to manually check competitor prices and provides real-time insights into pricing changes and market trends. Overview This workflow automatically scrapes competitor websites and pricing pages to extract current pricing information, product details, and promotional offers. It uses Bright Data to access pricing data without restrictions and AI to intelligently parse pricing information and detect changes over time. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping competitor pricing pages without being blocked OpenAI**: AI agent for intelligent pricing data extraction and analysis Google Sheets**: For storing pricing data and tracking changes over time How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your pricing tracking spreadsheet Customize: Define competitor URLs and pricing monitoring parameters Use Cases Pricing Strategy**: Stay competitive by monitoring market pricing trends Product Management**: Track competitor feature and pricing changes Sales Teams**: Provide up-to-date competitive pricing information Market Research**: Analyze pricing patterns and market positioning Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #pricemonitoring #competitorpricing #brightdata #webscraping #pricinganalysis #n8nworkflow #workflow #nocode #competitoranalysis #pricingdata #marketresearch #pricingtrends #competitiveintelligence #pricingtracking #marketanalysis #pricecomparison #competitormonitoring #businessintelligence #pricingstrategy #marketpositioning #pricinginsights #competitorresearch #pricingautomation #markettrends #pricealerts #dynamicpricing #pricingoptimization #competitivepricing
by Itamar
🕵️ Company Research Agent (n8n + Explorium + LLM) This n8n workflow automates company research by combining Explorium’s MCP server, web scraping tools, and an AI agent. Results are written to a Google Sheet for easy use in GTM, product analysis, or competitive research. 🚀 What It Does Given a list of company domains or names, this workflow will: Look up company information using: 🧠 LLM Agent to guide the research 🔎 Explorium MCP Server for firmographic & tech signals 🌐 Website content and SerpAPI scraping (optional) Extract key commercial details (see below) Format the output in a consistent JSON structure Update a connected Google Sheet with the enriched results 🧩 Extracted Fields Each company is enriched with: domain linkedinUrl has_free_trial cheapest_plan has_enterprise_plan last_case_study_link market (e.g., B2B or B2C) integrations (e.g., Slack, Hubspot, MySQL) enrichment_status 📥 Input Sheet Format | input | |-------------| | Explorium | | n8n | | Apple | | ... | 📤 Output Sheet Format | domain | linkedinUrl | has_free_trial | cheapest_plan | has_enterprise_plan | last_case_study_link | market | integrations | enrichment_status | |--------------|----------------------------------|----------------|----------------|----------------------|-----------------------------|--------|---------------------------------------------------|-------------------| | Explorium.ai | https://linkedin.com/company/... | TRUE | 69 | TRUE | https://www.explorium.com | B2B | ["HubSpot", "Zapier", "Salesforce", ...] | done | | n8n.io | https://linkedin.com/company/... | TRUE | 20 | TRUE | https://n8n.io/case-studies | B2B | ["Slack", "Gmail", "MySQL", "Google Sheets", ...] | done | 🛠️ Tools Used n8n** (Automation platform) Explorium MCP Server** – rich company enrichment via API Anthropic Claude or OpenAI** – used by the AI researcher Google Sheets** – stores output data Structured Output Parser** – ensures clean, predictable JSON formatting 📦 How to Set It Up Add your company domains or names to the input sheet Configure your MCP and SerpAPI credentials in n8n Run the workflow using the Test Workflow trigger Watch the sheet populate with results You can adapt the system to output different formats or fields depending on your team's research goals. 📌 Use Cases Competitive landscape analysis Lead intelligence for outbound campaigns Feature benchmarking (e.g., who offers enterprise or free trial) VC/investment research 🧠 Notes This agent is easily customizable. Adjust the LLM prompt or Output Parser to extract different properties. Explorium MCP is leveraged as the core enrichment engine, ensuring signal accuracy and freshness.