by DUBCOM
Workflow: Snapshot Contabo How it Works This workflow automates daily backups (snapshots) of VPS instances hosted on Contabo. Each day at midnight, it checks for existing snapshots and ensures that only the latest backups are retained by removing older ones. It provides a seamless, hands-off backup process to keep your data secure. Setup Steps Setting up this workflow is quick, typically taking about 10-15 minutes. The essential part of the setup is providing the necessary credentials, which you can easily retrieve from your Contabo control panel. Import the Workflow: Download and upload the workflow JSON into n8n. Configure Credentials: Add CLIENT_ID, CLIENT_SECRET, API_USER, and API_PASSWORD in the credential node. Activate the Workflow: Enable it to run automatically at midnight every day. Flow Overview Schedule Trigger (00:00 daily):** Automatically initiates the workflow. Formatted Date:** Prepares a timestamp for naming the snapshot. List Snapshots:** Verifies if an existing snapshot is available for each VPS. Conditional Logic:** No Snapshot? Proceeds to create a new one. Snapshot Found? Deletes the old snapshot before creating a new one. Key Points Snapshot Retention:** Old snapshots are deleted to ensure only the latest backups are stored. Unique Identifiers:** UUIDs are used to track and guarantee unique operations.
by Luciano Gutierrez
Healthcare Clinic Assistant with WhatsApp and Telegram Integration Version: 1.1.0 n8n Version: 1.88.0+ License: MIT 📋 Description A comprehensive and modular automation workflow designed for healthcare clinics. It manages patient communication, appointment scheduling, confirmations, rescheduling, internal tasks, and media processing by integrating WhatsApp, Telegram, Google Calendar, and Google Tasks, combined with AI-powered agents for maximum efficiency. This system guarantees proactive communication with patients, streamlined internal clinic management, and consistent data synchronization across platforms. 🌟 Key Features 🤖 AI-Powered Specialized Agents: Distinct agents handle WhatsApp patient support, appointment confirmations, and internal rescheduling tasks. 📱 Omnichannel Communication: Handles patient interactions via WhatsApp and staff commands via Telegram. 📅 Google Calendar Appointment Management: Full synchronization for creating, updating, canceling, and confirming appointments. 📋 Task Management with Google Tasks: Manages shopping lists and administrative tasks efficiently through staff Telegram requests. 🔔 Automated Appointment Reminders: Daily-triggered system proactively sends WhatsApp confirmations to patients for next-day appointments. 🖼️ Intelligent Media Processing: Transcribes audios, extracts text from images, and processes documents using OpenAI and OpenRouter AI models. 🛡️ Escalation to Human Support: Automatically detects sensitive or urgent cases and escalates them to a human agent when needed. 🏥 Use Cases Patient Communication:** Respond to inquiries, schedule, reschedule, and confirm appointments seamlessly via WhatsApp. Internal Clinic Operations:** Allow staff to modify appointments or add shopping list reminders directly from Telegram. Appointment Confirmation System:** Automatically contacts patients one day prior to appointments for confirmation or rescheduling. Task and Reminder Management:** Keeps clinic operations organized through automatic task management with Google Tasks. 🛠️ Technical Implementation WhatsApp Patient Interaction Flow Webhook Reception:** Incoming WhatsApp messages captured via Evolution API webhook. Message Classification:** Intelligent routing of messages based on content type (text, image, audio, document). Media Content Processing:** Audios: Download, convert, and transcribe via OpenAI Whisper. Images: Analyze and extract text/descriptions with OpenAI Vision model. Patient Request Handling:** Specialized WhatsApp assistant responds appropriately using AI prompts. Outbound Message Formatting:** Ensures messages comply with WhatsApp format standards. Message Delivery:** Sends responses back via Evolution API. Telegram Staff Management Flow Telegram Webhook Reception:** Captures messages from authorized staff accounts. Internal Assistant Processing:** Appointment Rescheduling: Identifies and updates appointments through MCP Google Calendar. Task Creation: Adds new entries to the clinic's shopping list using Google Tasks. Notifications and Confirmations:** Sends confirmations back to staff through Telegram. Appointment Reminder System Daily Trigger Activation:** Fires every weekday at 08:00 AM. Calendar Scraping:** Lists next day's appointments from Google Calendar. Patient Contact:** Sends WhatsApp confirmation messages for each appointment. Response Management:** Redirects confirmation or rescheduling replies to appropriate agents. ⚙️ Setup Instructions Import the Workflow n8n → Workflows → Import from File → Upload this JSON file. Configure Credentials Evolution API (WhatsApp Communication) Telegram Bot API (Staff Communication) Google Calendar OAuth2 (Appointment Management) Google Tasks OAuth2 (Task Management) OpenAI and OpenRouter APIs (AI Agents) PostgreSQL Database (Chat Memory) Set Sensitive Variables Replace placeholder values: {sua instância aqui} → Evolution API instance name {número_whatsapp} → WhatsApp numbers {url_do_servidor} → Server URLs {a sua apikey aqui} → API keys {seu_calendario} → Google Calendar ID Customize AI Prompts Adjust system prompts to fit your clinic’s tone, service style, and patient communication guidelines. Set clinic operating hours, escalation rules, and cancellation procedures in AI prompts. Activate and Test Simulate patient messages via WhatsApp. Test Telegram commands from staff members. Validate daily appointment reminders using the scheduled trigger. 🏷️ Tags Healthcare Clinic Management WhatsApp Integration Telegram Bot Appointment Scheduling Google Calendar Google Tasks AI Agents n8n Automation 📚 Technical Notes PostgreSQL is used for persistent chat memory across sessions. Multiple AI Models Used: OpenAI GPT-4.1-nano OpenAI GPT-4.1-mini Google Gemini 2.0 and 2.5 Full media content processing supported (audio, image, text). Compliant escalation workflows ensure patient safety and proper handoff to human staff when necessary. All sensitive patient data are securely stored inside calendar event descriptions for easy retrieval by agents. 📜 License This workflow is provided under the MIT License. Feel free to adapt and customize it for your clinic’s specific needs.
by Jez
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Uncover new business leads with this AI-Powered Prospect Discovery Agent! This n8n workflow acts as a specialized intelligent assistant that, given a business type and location, uses multiple search strategies to identify a list of potential prospect companies and their websites. Stop manually trawling through search results! This agent automates the initial phase of lead generation by: Understanding your target business profile (type, location, keywords). Strategically using web search tools (Brave Search, Google Gemini Search) to find relevant businesses. Performing quick validations to confirm relevance. Returning a clean, structured JSON list of prospect names and their website URLs. How it Works: The workflow is built around an AI agent powered by Google Gemini. This agent is equipped with tools like: Brave Web Search:** For broad initial sourcing of potential business candidates. Google Gemini Search:** For advanced, context-aware discovery and finding businesses mentioned in various online sources. Brave Local Search (Selective):** For quick verification of local presence or finding website URLs for identified names. Jina AI Web Page Scraper (Very Selective):** For extremely rapid relevance checks on uncertain websites by scanning page content for keywords. The agent's system prompt guides it to use these tools efficiently to build a list of prospects without getting bogged down in deep research on any single one at this discovery stage. Use Cases: Lead Generation:** Automatically generate lists of potential clients based on industry and location. Market Research:** Identify key players or types of businesses in a specific geographical area. Sales Development:** Provide SDRs with initial lists of companies to research further. Called as a Sub-Workflow:** Designed to be easily integrated as a "tool" into more complex orchestrating AI agents (e.g., a BNI Pitch Planner that first needs to identify who to target). Setup: Import the workflow. Configure Credentials: You'll need n8n credentials for: Google Gemini (for the Chat model and the Gemini Search/Vertex AI Search tool). Brave Search (e.g., via Smithery MCP, or adapt if you have direct API access). Jina AI (for the web scraper). Assign these to the respective nodes. Review System Prompt: The prospect_discovery_agent node contains a detailed system prompt. You can fine-tune this to adjust its search strategies or the strictness of its matching. Inputs: This workflow is triggered by an "Execute Workflow Trigger" node (prospect_discovery_workflow). It expects the following inputs: business_type (string): e.g., "artisan bakery" location_query (string): e.g., "Portland, Oregon" desired_num_prospects (number): e.g., 5 additional_keywords (string, optional): e.g., "organic, gluten-free" To Use (as a Sub-Workflow/Tool): This workflow is typically called by another n8n workflow (e.g., using a "Tool Workflow" node from the Langchain nodes). The calling workflow would provide the inputs listed above. The "Prospect Discovery" workflow will then execute and its final node (the prospect_discovery_agent) will output a JSON array of found prospects, like: [ { "business_name": "Rose Petal Bakery", "website_url": "https://rosepetalbakerypdx.com" }, { "business_name": "The Daily Bread Artisans", "website_url": "https://dailybreadpdx.com" } ] If no prospects are found, it returns an empty array []. This template provides a powerful and focused tool for automating the initial stages of prospect identification.
by David Ashby
Complete MCP server exposing 1 IP2Proxy Proxy Detection API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add IP2Proxy Proxy Detection credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the IP2Proxy Proxy Detection API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to https://api.ip2proxy.com • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (1 total) 🔧 General (1 endpoints) • GET /: Check Proxy IP 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native IP2Proxy Proxy Detection API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by Cyril Nicko Gaspar
🔍 Email Lookup with Google Search from Postgres Database This N8N workflow is designed to enrich seller data stored in a Postgres database by performing automated Google search lookups. It uses Bright Data's Web Unlocker to bypass search result restrictions and the HTML Extract node to parse and extract relevant information from webpages. The main purpose of this workflow is to discover missing contact details, company domains, and secondary emails for businesses or sellers based on existing database entries. 🎯 Problem This Workflow Solves Manually searching for missing seller or business details—like secondary emails, websites, or domain names—can be time-consuming and inefficient, especially for large datasets. This workflow automates the search and data enrichment process, significantly reducing manual effort while improving the quality and completeness of your seller database. ✅ Prerequisites Before using this template, make sure the following requirements are met: ✔️ A Bright Data account with access to the Web Unlocker or Amazon Scraper API ✔️ A valid Bright Data API key ✔️ An active PostgreSQL database with seller data ✔️ N8N self-hosted instance (recommended for using community nodes like n8n-nodes-brightdata) ✔️ Installed n8n-nodes-brightdata package (custom node for Bright Data integration) ⚙️ Setup Instructions Step 1: Prepare Your Postgres Table Create a table in Postgres with the following structure (you can adjust field names if needed): CREATE TABLE sellers ( seller_id SERIAL PRIMARY KEY, seller_name TEXT, primary_email TEXT, company_info TEXT, trade_name TEXT, business_address TEXT, coc_number TEXT, vat_number TEXT, commercial_register TEXT, secondary_email TEXT, domain TEXT, seller_slug TEXT, source TEXT ); Step 2: Setup Web Unlocker on Bright Data Go to your Bright Data dashboard. Navigate to Proxies & Scraping → Web Unlocker. Create a new zone, selecting Web Unlocker API under Scraping Solutions. Whitelist your server IP if required. Step 3: Generate API Key In the Bright Data dashboard, go to the API section. Generate a new API key. In N8N, create HTTP Request Credentials using Bearer Authentication with the API key. Step 4: Install the Bright Data Node in N8N In your N8N self-hosted instance, go to Settings → Community Nodes. Search and install n8n-nodes-brightdata. 🔄 Workflow Functionality 🔁 Trigger: Can be set to run on a schedule (e.g., daily) or manually. 📥 Read: Fetches seller records from the Postgres table. 🌐 Search: Uses Bright Data to perform a Google search based on seller_name, company_info, or trade_name. 🧾 Extract: Parses the HTML content using the HTML Extract node to identify potential websites and email addresses. 📝 Update: Writes enriched data (like domain or secondary_email) back to the Postgres table. 💡 Use Cases Lead enrichment for e-commerce sellers Domain and contact info discovery for B2B databases Email and web domain verification for CRM systems Market research automation 🛠️ Customization Tips You can enhance the parsing logic in the HTML Extract node to look for phone numbers, LinkedIn profiles, or social media links. Modify the search query logic to include additional parameters like location or industry for more refined results. Integrate additional APIs (e.g., Hunter.io, Clearbit) for email validation or social profile enrichment. Add filtering to skip entries that already have domain or secondary_email.
by Jimleuk
This template attempts to replicate OpenAI's DeepResearch feature which, at time of writing, is only available to their pro subscribers. > An agent that uses reasoning to synthesize large amount of online information and complete multi-step research tasks for you. Source Though the inner workings of DeepResearch have not been made public, it is presumed the feature relies on the ability to deep search the web, scrape web content and invoking reasoning models to generate reports. All of which n8n is really good at! Using this workflow, n8n users can enjoy a variation of the Deep Research experience for themselves and their teams at a fraction of the cost. Better yet, learn and customise this Deep Research template for their businesses and/or organisations. Check out the generated reports here: https://jimleuk.notion.site/19486dd60c0c80da9cb7eb1468ea9afd?v=19486dd60c0c805c8e0c000ce8c87acf How it works A form is used to first capture the user's research query and how deep they'd like the researcher to go. Once submitted, a blank Notion page is created which will later hold the final report and the researcher gets to work. The user's query goes through a recursive series of web serches and web scraping to collect data on the research topic to generate partial learnings. Once complete, all learnings are combined and given to a reasoning LLM to generate the final report. The report is then written to the placeholder Notion page created earlier. How to use Duplicate this Notion database template and make sure all Notion related nodes point to it. Sign-up for APIFY.com API Key for web search and scraping services. Ensure you have access to OpenAI's o3-mini model. Alternatively, switch this out for o1 series. You must publish this workflow and ensure the form url is publically accessible. On depth & breadth configuration For more detailed reports, increase depth and breadth but be warned the workflow will take exponentially longer and cost more to complete. The recommended defaults are usually good enough. Depth=1 & Breadth=2 - will take about 5 - 10mins. Depth=1 & Breadth=3 - will take about 15 - 20mins. Dpeth=3 & Breadth=5 - will take about 2+ hours! Customising this workflow I deliberately chose not to use AI-powered scrapers like Firecrawl as I felt these were quite costly and quotas would be quickly exhausted. However, feel free to switch web search and scraping services which suit your environment. Maybe you don't decide to source the web and instead, data collection comes from internal documents instead. This template gives you freedom to change this. Experiment with different Reasoning/Thinking models such as Deepseek and Google's Gemini 2.0. Finally, the LLM prompts could definitely be improved. Refine them to fit your use-case. Credits This template is largely based off the work by David Zhang (dzhng) and his open source implementation of Deep Research: https://github.com/dzhng/deep-research
by Ferenc Erb
Overview Transform your Bitrix24 Open Line channels with an intelligent chatbot that leverages Retrieval-Augmented Generation (RAG) technology to provide accurate, document-based responses to customer inquiries in real-time. Use Case This workflow is designed for organizations that want to enhance their customer support capabilities in Bitrix24 by providing automated, knowledge-based responses to customer inquiries. It's particularly useful for: Customer service teams handling repetitive questions Support departments with extensive documentation Sales teams needing quick access to product information Organizations looking to provide 24/7 customer support What This Workflow Does Smart Document Processing Automatically processes uploaded PDF documents Splits documents into manageable chunks Generates vector embeddings for semantic understanding Indexes content for efficient retrieval AI-Powered Responses Utilizes Google Gemini AI to generate natural language responses Constructs answers based on relevant document content Maintains conversation context for coherent interactions Provides fallback responses when information is not available Vector Database Integration Stores document embeddings in Qdrant vector database Enables semantic search beyond simple keyword matching Retrieves the most relevant information for each query Maintains a persistent knowledge base that grows over time Webhook Handler Processes incoming messages from Bitrix24 Open Line channels Handles authentication and security validation Routes different types of events to appropriate handlers Manages session and conversation state Event Routing Intelligently routes different event types: ONIMBOTMESSAGEADD: Processes new user messages ONIMBOTJOINCHAT: Handles bot joining a conversation ONAPPINSTALL: Manages application installation ONIMBOTDELETE: Handles bot deletion Document Management Organizes processed documents in designated folders Tracks document processing status Moves indexed documents to appropriate locations Maintains document metadata for reference Interactive Menu Provides menu-based options for common user requests Customizable menu items and responses Easy navigation for users seeking specific information Fallback to operator option when needed Technical Architecture Components Webhook Handler: Receives and validates incoming requests from Bitrix24 Credential Manager: Securely manages authentication tokens and API keys Event Router: Directs events to appropriate processing functions Document Processor: Handles document loading, chunking, and embedding Vector Store: Qdrant database for storing and retrieving document embeddings Retrieval System: Searches for relevant document chunks based on user queries LLM Integration: Google Gemini model for generating natural language responses Response Manager: Formats and sends responses back to Bitrix24 Integration Points Bitrix24 API**: For bot registration, message handling, and user interaction Ollama API**: For generating document embeddings Qdrant API**: For vector storage and retrieval Google Gemini API**: For AI-powered response generation Setup Instructions Prerequisites Active Bitrix24 account with Open Line channels enabled Access to n8n workflow system Ollama API credentials Qdrant vector database access Google Gemini API key Configuration Steps Initial Setup Import the workflow into your n8n instance Configure credentials for all services Set up webhook endpoints Bitrix24 Configuration Create a new Bitrix24 application Configure webhook URLs Set appropriate permissions Install the application to your Bitrix24 account Document Storage Create a designated folder in Bitrix24 for knowledge base documents Configure folder paths in the workflow settings Upload initial documents to be processed Bot Configuration Customize bot name, avatar, and description Configure welcome messages and menu options Set up fallback responses Testing Verify successful installation Test document processing pipeline Send test queries to evaluate response qu
by Hunyao
Note: This workflow assumes you already have your product’s Amazon reviews saved in a Google Sheet. If you still need those reviews, run my Amazon Reviews Scraper workflow first, then plug the resulting spreadsheet into this template. What it does Transforms any draft Google Doc into multiple high-converting sales pages. It blends Alex Hormozi’s value-stacking tactics with persona targeting based on Maslow’s Hierarchy of Needs, using your own customer reviews for proof and voice of customer (VOC). Perfect for • Growth and creative strategists • Freelance copywriters and agencies • Founders sharpening offers and funnels Apps used Google Sheets, Google Docs, LangChain OpenRouter LLM How it works Form Trigger collects Drive folder IDs, base copy URL and options. Workflow fetches the draft copy and product feature doc. It samples reviews, extracts VOC insights and maps them to Maslow needs. LLM drafts headlines and hooks following Hormozi’s \$100M Offers principles. Personas drive tone, objections and urgency in each copy variant. Loop writes one Google Doc per variant in your chosen folder. Customer analysis docs are saved to a second folder for reuse. Setup Share two Drive folders, copy the IDs (text after folders/). Paste each ID into Customer Analysis Folder ID and Advertorial Copy Folder ID. Provide File Name, Base copy (Google Docs URL) and Product Feature/USPs Doc. Optional: Reviews Sheet URL, Number of reviews to use, Target City. Set Number of Copies you need (1–20). Add Google Docs OAuth2 and Google Sheets OAuth2 credentials in n8n. If you have any questions in running the workflow, feel free to reach out to me at my youtube channel: https://www.youtube.com/@lifeofhunyao
by isaWOW
Description An AI-powered workflow that analyzes any website to identify missing pages that would improve user experience and business performance. Submit a URL, and the system detects existing pages, researches competitors using Perplexity, and generates a professional gap analysis report with prioritized recommendations—saved directly to Google Docs. What this workflow does This automation delivers a complete website page gap analysis: Smart page detection:** Automatically scans website HTML and identifies 15 common page types (About, Contact, Services, Blog, Portfolio, Pricing, FAQ, Testimonials, Team, Careers, Privacy, Terms, etc.) Business type classification:** Determines if the site is ecommerce, portfolio/agency, blog, SaaS, or service-based to tailor recommendations Competitor research:** Uses Perplexity to research 5-7 top competitors in the same industry and identify their page structures AI gap analysis:** GPT-4.1-mini with web search compares the website against industry standards and competitor best practices Prioritized recommendations:** Generates High/Medium/Low priority suggestions with business value explanations and actionable content ideas Google Docs report:** Saves a professional gap analysis report ready to share with clients or stakeholders Setup requirements Tools you'll need: Active n8n instance (self-hosted or n8n Cloud) Google Docs with OAuth access OpenAI API key (GPT-4.1-mini access with web search) Perplexity API key (for competitor research) Estimated setup time: 15–20 minutes Step-by-step setup 1. Connect Google Docs In n8n: Credentials → Add credential → Google Docs OAuth2 API Complete OAuth authentication Create a new Google Doc for storing reports (or use an existing one) Open "Save Report to Google Docs" node Paste the Google Doc URL in the documentURL field 2. Add OpenAI API credentials Get API key: https://platform.openai.com/api-keys In n8n: Credentials → Add credential → OpenAI API Paste your API key Open "OpenAI GPT-4.1 Mini with Web Search" node Select your OpenAI credential Ensure model is set to gpt-4.1-mini Verify Web Search is enabled in the built-in tools section 3. Add Perplexity API credentials Get API key: https://www.perplexity.ai/settings/api In n8n: Credentials → Add credential → Perplexity API Paste your API key Open "Perplexity Competitor Research Tool" node Select your Perplexity credential 4. Share the form URL Open "Submit Website URL for Analysis" node Copy the Form URL from the node settings Share this URL with anyone who needs to run website audits The form accepts a single field: Website URL 5. Test the workflow Open the Form URL in your browser Enter a test website: https://example.com Submit the form Wait 30-60 seconds for the analysis to complete Check your Google Docs—the gap analysis report should appear Verify that: Existing pages are correctly detected Business type is identified Competitor research is included Recommendations are prioritized and actionable 6. Activate the workflow Toggle the workflow to Active at the top The form will now accept submissions 24/7 Each submission generates a new report appended to your Google Doc How it works 1. URL submission via form User opens the form link and submits a website URL they want to analyze. The form triggers the workflow immediately. 2. HTML fetch and extraction The workflow sends an HTTP request to the submitted URL and retrieves the complete HTML source code of the website's homepage. 3. Automated page detection A code node analyzes the HTML to detect 15 common page types: Navigation pages:** Home, About, Contact, Services, Products Content pages:** Blog, Portfolio, Pricing, FAQ Social proof pages:** Testimonials, Team, Careers Legal pages:** Privacy Policy, Terms of Service The detection works by: Scanning all internal links in the HTML Matching URL patterns (e.g., /about, /contact-us, /services) Searching for navigation keywords in anchor text Normalizing URLs (removing query params, anchors, trailing slashes) 4. Business type classification The code also identifies the website's business type based on HTML content patterns: Ecommerce:** Detects shopping cart, checkout, product pages Portfolio/Agency:** Identifies portfolio, case studies, creative work Blog/Content site:** Finds blog posts, articles, news sections SaaS/Software:** Detects subscription, cloud platform, software keywords Service/Agency:** Identifies consulting, marketing, agency services General:** Default if no specific patterns match 5. Deep competitor research with Perplexity The AI Agent instructs Perplexity to: Crawl the target website thoroughly (navigation, footer, hidden menus) Research 5-7 top competitors in the identified business type Document each competitor's page structure Identify industry-standard pages that successful sites consistently have Perplexity focuses only on user-facing pages and ignores technical files (sitemap.xml, robots.txt, admin pages). 6. AI gap analysis and recommendations GPT-4.1-mini with web search: Compares the website's existing pages against competitor structures Identifies genuinely missing pages (cross-references with detected existing pages) Prioritizes recommendations based on business impact (High/Medium/Low) Provides specific business value explanations for each recommendation Suggests actionable content for each missing page Includes competitor examples showing how others use these pages Critical filtering rules: Never recommends pages already detected Excludes technical files and admin pages Only suggests pages with clear business value Each recommendation must be justified 7. Professional report generation The AI Agent outputs a structured report containing: Website Overview:** URL, business type, analysis date Pages Confirmed as Existing:** Formatted list of detected pages Gap Analysis Summary:** Total pages analyzed, existing vs. missing count Recommended Pages to Add:** Detailed recommendations with priority, business value, content suggestions, and competitor examples Implementation Priority Summary:** Quick-reference lists by priority level Next Steps:** 2-3 actionable implementation steps 8. Save to Google Docs The complete report is inserted into your Google Docs document. Each new analysis appends to the same document, creating a running archive of all audits performed. Key features ✅ Automatic page detection: Scans HTML and identifies 15 common page types without manual input—no need to know the site structure beforehand ✅ Business type intelligence: Classifies websites into 6 business categories to provide industry-relevant recommendations ✅ Perplexity-powered research: Deep competitor analysis covering 5-7 top sites in the same niche with complete page inventories ✅ Smart filtering: Never recommends pages that already exist, even if they use different naming conventions or URL structures ✅ Priority-based recommendations: Every suggestion labeled High/Medium/Low with clear business justification—know what to build first ✅ Actionable content ideas: Not just "add a blog"—specific suggestions for what content should go on each recommended page ✅ Competitor examples: See how successful competitors use each recommended page type in their own sites ✅ Google Docs integration: Professional reports saved automatically—no downloads, no formatting, ready to share ✅ Form-based workflow: Single URL submission—anyone can run audits without touching n8n Troubleshooting HTML fetch fails SSL certificate issues:** Some websites block automated requests. Try adding "rejectUnauthorized": false in HTTP Request options. Cloudflare/bot protection:** Sites with aggressive protection may block the request. Test with a simple site first (like your own). Timeout errors:** Increase timeout setting in HTTP Request node to 30-60 seconds for slow-loading sites. Pages not detected correctly Non-standard URL structure:** The workflow detects pages using common patterns (/about, /contact, etc.). If a site uses unusual URLs like /company-info instead of /about, manual review may be needed. Single-page websites:** Sites built as SPAs (single-page applications) with JavaScript routing may not have distinct page URLs in HTML. The workflow works best with traditional multi-page sites. Check detection details:** The code node outputs detection_details showing exactly what was found. Review this to debug false negatives. Perplexity API errors Rate limits:** Perplexity has usage limits. If you hit them, wait or upgrade your plan. API key invalid:** Verify the key is correct at https://www.perplexity.ai/settings/api Quota exceeded:** Check your Perplexity dashboard for remaining credits. OpenAI web search not working Web search disabled:** Ensure "Web Search" is enabled in the OpenAI Chat Model node under "Built-in Tools" Search context size:** Set to "Medium" for balanced performance Model compatibility:** Web search only works with GPT-4 models, not GPT-3.5 Report not saving to Google Docs Re-authenticate OAuth:** Go to Credentials → Google Docs OAuth2 API → Reconnect Document URL format:** Ensure the URL is a valid Google Docs link (not a folder or Sheets link) Permissions:** Verify the connected Google account has edit access to the document Document locked:** Check if the document is open in another tab with unsaved changes AI recommendations are too generic Perplexity research quality:** The AI relies on Perplexity's research. If competitors have similar structures, recommendations may overlap. Try analyzing a more unique website. Business type misclassification:** Check the detected business_type in the code output. If it's wrong, the recommendations will be off-target. Improve prompt:** Edit the AI Agent system prompt to emphasize more specific or creative recommendations for your niche. Use cases Web design agencies: Audit client websites before proposals. Show exactly which pages are missing compared to competitors, with business justification for each recommendation. Win more projects by demonstrating data-driven insights. SEO consultants: Include page gap analysis in site audits. Identify missing pages that competitors rank for (FAQ, pricing, testimonials). Provide clients with actionable roadmaps for site expansion. In-house marketing teams: Analyze competitor websites quarterly to spot new page types or content strategies. Keep your site competitive by identifying gaps before leadership asks about them. Freelance developers: Offer value-add services to existing clients. Run audits on their sites, identify quick wins (missing FAQ, testimonials, pricing pages), and sell additional development work. Startup founders: Before hiring a design agency, understand what pages you actually need. Get a competitor-researched report showing industry standards for your business type—save time and budget. Content strategists: Identify content opportunities beyond blog posts. See which informational pages (FAQ, case studies, resource libraries) competitors have that you're missing. Expected results Time savings:** 2-3 hours saved per website audit (manual competitor research eliminated) Analysis speed:** Complete gap analysis in 30-60 seconds vs. hours of manual work Competitor coverage:** Research 5-7 competitors automatically vs. 1-2 manual comparisons Report quality:** Professional, shareable reports vs. rough notes or spreadsheets Actionability:** Prioritized recommendations with business value vs. generic "add more pages" advice Scalability:** Run 50+ audits per day without additional effort Support Need help or custom development? 📧 Email: info@isawow.com 🌐 Website: https://isawow.com/
by Zacharia Kimotho
N8n recently introduced folders and it has been a big improvement on workflow management on top of the tags. This means the current workflows need to be moved manually to the folders. The simplest idea to try is to convert the current tags into folders and move all the current workflows within the respective tags into the folders This assumes the tag name will be used as the folder name. To Note For workflows that use more than 1 tag, the workflow will be assigned the last tag that runs as the folder. How does it work I took the liberty of simplifying the setup of this workflow that will be needed on your part and also be beginner-friendly Copy and paste this workflow into your n8n canvas. You must have existing workflows and tags before you can run this Set your n8n login details on the node set Credentials with the n8n URL, username, and password. Setup your n8n API credentials on the n8n node get workflows Run the workflow. This opens up a form where you can select the number of tags to move and click on submit The workflow responds with the successful number of workflows that were imported Read more about the template Built by Zacharia Kimotho - Imperol
by Yaron Been
Description This workflow automatically monitors companies across courts, regulators, and jurisdictions to detect legal risk signals early. It helps legal, compliance, and risk teams stay ahead of litigation threats without manually scanning dozens of public sources. Overview This workflow scrapes court records, regulatory filings, and legal news using Bright Data, then uses AI to classify, score, and cluster legal events by jurisdiction and topic. It filters noise, identifies high-risk cases, and generates executive-ready intelligence — including High-Risk Escalation Alerts and Litigation Monitoring Briefs — logged directly into Google Sheets dashboards. Tools Used n8n: The automation platform that orchestrates the workflow. Bright Data: For scraping court records, regulatory sources, and legal news without getting blocked. OpenRouter: For AI-powered legal case classification, risk scoring, and report generation. Google Sheets: For logging alerts, monitoring summaries, and error tracking. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data API credentials to the Bright Data node. Configure OpenRouter: Add your OpenRouter API key for AI classification and report generation. Set Up Google Sheets: Create a spreadsheet following the "Google Sheets Setup" sticky note inside the workflow, then connect each Google Sheets node to your document. Customize: Edit the configuration node to set your target companies, jurisdictions, courts, and regulators. Use Cases Corporate Legal Teams: Get early warnings on litigation involving your company or partners. M&A Due Diligence: Screen acquisition targets for hidden legal exposure before closing deals. Compliance Officers: Monitor regulatory actions across multiple jurisdictions in one place. Risk Analysts: Track litigation density patterns and jurisdiction concentration risks. Investor Relations: Surface legal threats that could impact portfolio companies or stock price. Connect with Me Website: https://www.nofluff.online YouTube: https://www.youtube.com/@YaronBeen/videos LinkedIn: https://www.linkedin.com/in/yaronbeen/ Get Bright Data: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #brightdata #webscraping #legalrisk #litigation #legaltech #riskmanagement #compliance #duediligence #legalmonitoring #courtrecords #regulatorymonitoring #corporatelegal #riskanalysis #litigationtracking #jurisdictionrisk #legalintelligence #n8nworkflow #workflow #nocode #businessintelligence #earlywarning #legalcompliance #riskassessment
by Suleman Hasib
Template Overview This template is designed for individuals and businesses who want to maintain a consistent presence on the Fediverse while also posting on Threads or managing multiple Fediverse profiles. By automating the process of resharing statuses or posts, this workflow saves time and ensures regular engagement across accounts. Use Case The template addresses the challenge of managing activity across Fediverse accounts by automatically boosting or resharing posts from a specific account to your own. It is especially helpful for users who want to consolidate engagement without manually reposting content across multiple platforms or profiles. How It Works The workflow runs on a scheduled trigger and retrieves recent posts from a specified Fediverse account, such as your Threads.net account. It uses a JavaScript filter to identify posts from the current day and then automatically boosts or reshares them to your selected Mastodon profile. Preconditions You need a Mastodon account with developer access. Identify a Threads.net or other Fediverse account which you want to boost. Basic familiarity with APIs and setting up credentials in n8n. Setup Steps Step 1: Create a Developer Application on Mastodon Log in to your Mastodon account and navigate to Preferences > Development > New Application. Fill out the required information and create your application. Set Scopes to atleast read, profile, write:statuses. Click Submit. Note down the access token generated for this application. Step 2: Get the Account ID Use the following command to retrieve the account ID for the profile you want to boost: curl -s "https://mastodon.social/api/v1/accounts/lookup?acct=<ACCOUNTNAME>" Alternatively, paste the URL into a GET node on n8n. From the returned JSON, copy the "id" field value (e.g., {"id":"110564198672505618", ...}). Step 3: Update the "Get Statuses" Node Replace <ACCOUNTID> in the URL field with the ID you retrieved in Step 2: https://mastodon.social/api/v1/accounts/<ACCOUNTID>/statuses Step 4: Configure the "Boost Statuses" Node Authentication type will already be set to Header Auth. Grab the access token from Step 1. In the Credential for Header Auth field, create a new credential. Click the pencil icon in the top-left corner to name your credential. In the Name field, enter Authorization. In the Value field, enter Bearer <YOUR_MASTODON_ACCESS_TOKEN>. (Note: there is a space after "Bearer.") Save the credential, and it should automatically be selected as your Header Auth. Step 5: Test the Workflow Run the workflow to ensure everything is set up correctly. Adjust filters or parameters as needed for your specific use case. Customization Guidance Replace mastodon.social with your own Mastodon domain if you're using a self-hosted instance. Adjust the JavaScript filter logic to meet your specific needs (e.g., filtering by hashtags or keywords). For enhanced security, store the access token as an n8n credential. Embedding it directly in the URL is ++not recommended++. Notes This workflow is designed to work with any Mastodon domain. Ensure your Mastodon account has appropriate permissions for boosting posts. By following these steps, you can automate your Fediverse engagement and focus on creating meaningful content while the workflow handles the rest!