by Dustin
Are you a cord-cutter? Do you find yourself looking through the many titles of videos uploaded to Youtube, just to find the ones you want to watch? Even when you subscribe to the channels you like, do you find that you want to watch the news now and my tech/n8n videos later? Well, now you can have n8n grab the last 8 videos, posted in the last 24 hours, and put them in a playlist for the day; and, each day the old playlist is deleted. Are you tired of a channel filling your subscriptions with tons of videos a day; this workflow can be used for any channel, whether you are subscribed to the channel or not. It's a YouTube playlist automation. How it works: Create your list of prefered Youtube Channels in a Google Sheet and it will create you a daily playlist; and, it will delete the playlist created yesterday. Instructions To set this up, you need to create a Google Sheet with the following headings in line 1: Channel User Name Channel Name Channel Link Channel ID Copy the 'Create your Channel List' into it's own workflow and link the Sheets links to your new sheet. To get the 'Create your Channel List' to work, you need to visit each channel's page that you want included in your playlist; you need to get the "@" name of the channel and add it to the 'Channel User Name' column of your Google Sheet. For example: if you wanted to include this channel: Recruit Training Videos - Corporal Stock, you would search for the name, to add to the next available row of the 'Channel User Name' column: @CorporalStock Once you add all Channel User Names, run the 'Create your Channel list workflow, and it will fill in the remaining details. Now the 'YT Playlist Creator' can be run. Note: The first time the workflow us run, disconnect the 'Delete Yesterday's Playlist' leg, or the workflow will error and stop (because there is no 'Yesterday's Playlist'. Note: this was made to create a playlist every day, delete yesterday's playlist, and only get the last 8 videos posted within the last 24 hours. I choose to put the date (YYMMDD format) in front of the playlist, to ensure that it doesn't conflict with another playlist. Also, I have it notifying me in Telegram, so I know that the new playlist is posted.
by explorium
Explorium Event-Triggered Outreach This n8n and agent-based workflow automates outbound prospecting by monitoring Explorium event data (e.g. product launches, new office opening, new investment and more), researching companies, identifying key contacts, and generating tailored sales emails leveraging the Explorium MCP server. Template Workflow Overview Node 1: Webhook Trigger Purpose: Listens for real-time product launch events pushed from Explorium's webhook system. How it works: Explorium sends HTTP POST requests containing event data The webhook payload includes company name, business ID, domain, product name, and event type Pay attention: Product launch is just one example. You can easily enroll to many more meaningful events. to learn about events and how to enroll to events, visit the events documentation. Node 2: Company Research Agent Agent Type: Tools Agent Purpose: Enrich company data after an event occurs. How it works: Uses Explorium MCP via the MCP Client tool to gather additional company data Uses Anthropic Claude (Chat Model) to process and interpret company information for downstream personalization Node 3: Employee Data Retrieval Purpose: Retrieve prospect-level data for targeting. How it works: Uses HTTP Request node to call Explorium's fetch_prospects endpoint Filters prospects by: Company business_id Departments: Product, R&D, etc... Seniority levels: owner, cxo, vp, director, senior, manager, partner, etc... Pay Attention: Follow our fetch prospect documentation for the full list of filter and best practice. Limits results to top 5 relevant employees Code nodes handle: Filtering logic Cleaning API response Formatting data for downstream agents Node 4: Conditional Branch - Prospect Data Check If Node: Checks whether prospect data was successfully retrieved Logic: If prospects found → personalized emails per person If no prospects → fallback to company-level general email Node 5A: Email Writer #1 (No Prospect Data) Agent Type: Tools Agent Purpose: Write generic outbound email using only company-level research and event info. Powered by: Anthropic Chat Model Node 5B: Loop Over Prospects → Email Writer #2 (Personalized) Agent Type: Tools Agent Purpose: Write highly personalized email for each identified employee. How it works: Loops through each individual prospect Passes company research + employee data to LLM agent Generates customized emails referencing: Prospect's title & department Product launch Role-relevant Explorium value proposition Node 6: Slack Notifications Purpose: Posts completed emails to internal Slack channel for review or testing before final deployment. Future State: Can be swapped with an email sequencing platform in production. Setup Requirements Explorium API Access MCP Client credentials for company enrichment and prospect fetching Registered webhook for event listening Get explorium api key n8n Configuration Secure environment variables for API keys & webhook secret Code nodes configured for JSON transformation, filtering & signature validation Customization Options Personalization Logic Update LLM prompt instructions to reflect ICP priorities Modify email templates based on role, department, or tenure logic Adjust fallback behavior when prospect data is unavailable API Request Tuning Adjust page_size for number of prospects retrieved Fine-tune seniority and department filters to match evolving targeting Future Expansion Swap Slack notifications for outbound email automation Integrate call task assignment directly into CRM Introduce engagement scoring feedback loop (opens, clicks, replies) Troubleshooting Tips Validate webhook signature matching to prevent unauthorized requests Ensure correct business_id is passed to prospect fetching endpoint Confirm business enrichment returns sufficient data for company researcher agents Review agent LLM responses for correct output structure and parsing consistency
by Mihai Farcas
How it works: The workflow starts by sending a request to a website to retrieve its HTML content. It then parses the HTML extracting the relevant information The extracted data is storted and converted into a CSV file. The CSV file is attached to an email and sent to your specified address. The data is simultaneously saved to both Google Sheets and Microsoft Excel for further analysis or use. Set-up steps: Change the website to scrape in the "Fetch website content" node Configure Microsoft Azure credentials with Microsoft Graph permissions (required for the Save to Microsoft Excel 365 node) Configure Google Cloud credentials with access to Google Drive, Google Sheets and Gmail APIs (the latter is required for the Send CSV via e-mail node).
by Dale Dunlop
WebSecScan: AI-Powered Website Security Auditor This n8n workflow provides comprehensive website security analysis by leveraging OpenAI's models to detect vulnerabilities, configuration issues, and security misconfigurations. The workflow generates a professional HTML security report delivered directly via Gmail. Key Features Dual-Layer Security Analysis:** Performs parallel security audits using specialized OpenAI agents: Header Configuration Audit: Analyzes HTTP headers, CORS policies, CSP implementation, and cookie security Vulnerability Assessment: Identifies XSS vectors, information disclosure, and client-side weaknesses Detailed Security Grading:** Automatically calculates a security grade (A+ to F) based on findings severity and quantity Professional Report Generation:** Creates a comprehensive HTML report with: Security grade visualization Color-coded vulnerability categories Detailed recommendations with example configuration fixes Header presence/absence indicators Implementation guidance for remediation Non-Invasive Testing:** Performs analysis without active scanning or exploitation attempts Technical Implementation Multi-Agent Architecture:** Utilizes two specialized OpenAI agents with custom prompts tailored for security analysis Advanced Header Analysis:** Detects presence and proper implementation of critical security headers: Content-Security-Policy Strict-Transport-Security X-Content-Type-Options X-Frame-Options Referrer-Policy Permissions-Policy Intelligent Issue Detection:** Uses JavaScript processing to analyze OpenAI outputs and count critical/warning issues Responsive HTML Report:** Dynamically generates a mobile-friendly report with detailed findings and recommendations Setup Requirements 1. OpenAI API Configuration Create an OpenAI API key at platform.openai.com In n8n, go to Settings → Credentials → New → OpenAI API Enter your API key and save 2. Gmail Integration Navigate to Settings → Credentials → New → Gmail OAuth2 API Complete the OAuth authentication flow Configure recipient email in the "Send Security Report" node 3. Workflow Customization (Optional) Modify the form title/description in the Landing Page node Upgrade from gpt-4o-mini to gpt-4o for more comprehensive analysis Add additional recipients to the email report Usage Instructions Activate the workflow and access the form via the generated URL Enter any website URL to analyze (including the http:// or https:// prefix) Receive a detailed security report via email within minutes Share findings with your development team to implement fixes This workflow represents a non-invasive security assessment tool. For production environments, complement with professional penetration testing services.
by Jimleuk
This n8n template monitors an Outlook mailbox for invoices, automatically parses/extracts data from them and then uploads the output to an Excel Workbook. One of my top workflow requests, this template can save many hours of manual labour for you or your finance/accounts team. How it works A scheduled trigger is set to fetch recent Outlook messages to the Accounts receivable mailbox. Each message is analysed to determine whether or not it from a supplier and is issuing/contains an invoice. For each valid message, the attachments are downloaded and non-invoice documents are filtered out via AI Vision classification. Invoices are then processed through a AI vision model again to extract the details. The extracted data can then be used for reconciliation or otherwise. For this demonstration, we'll just append the row to an Excel sheet for now. How to use Ensure your Microsoft365 credential points to the correct mailbox. If a shared folder is used, toggle "shared folder" option to "on" and for the principal ID, use the email address. If you receive lots of other types of messages such as replies and forwards, you may want to implement additional checks to prevent processing invoices twice. The "remove duplicates" node can help with this. Requirements Outlook for Mailbox Google Gemini for Document Understanding and Invoice Extraction Excel for Data Storage Customising this workflow Note the assumption for this template is that all invoices will come as a PDF attachment. In real life, this is rarely the case! Adding in document conversion to cover all invoice formats. Human feedback is also an important factor in AI workflows. Try tagging emails as a way to notify team members that the invoice was processed.
by Don Jayamaha Jr
Analyze exchange data, market indexes, and community sentiment from CoinMarketCap—powered by AI. This sub-agent provides access to exchange listings, token holdings, metadata, and high-level metrics like the CMC 100 Index and the Fear & Greed Index. It’s designed for use within your larger CoinMarketCap AI Analyst system or as a standalone workflow. This agent can be triggered by a supervisor or manually used with message and sessionId inputs. Supported Tools (5 Total) 🔍 Exchange Map Get CoinMarketCap IDs, names, and slugs for exchanges (used as lookup before deeper queries). 🧾 Exchange Info Metadata including launch date, social links, country, and operational status. 💰 Exchange Assets Token balances, wallet addresses, and total USD value held by a specific exchange. 📈 CoinMarketCap 100 Index Constituents and weights of the CMC 100 Index, updated live. 😱 Fear & Greed Index Market sentiment score updated daily, ranging from Extreme Fear to Extreme Greed. What You Can Do with This Agent 🔹 Map exchanges to retrieve their ID and slug 🔹 Analyze exchange holdings by token and blockchain 🔹 Pull metadata for major CEXs like Binance or Coinbase 🔹 Compare global sentiment using the Fear & Greed Index 🔹 Access index data to understand CMC’s top 100 crypto asset breakdown Example Queries You Can Use ✅ "What is the latest Fear and Greed Index reading?" ✅ "Get a list of all exchanges on CoinMarketCap." ✅ "What tokens are held by Binance?" ✅ "Retrieve metadata for Coinbase." ✅ "Show me the top assets in the CMC 100 Index." Agent Architecture AI Brain**: GPT-4o-mini Memory**: Window buffer memory using sessionId Tools**: 5 API-connected nodes Trigger**: External input via message and sessionId Setup Instructions Get a CoinMarketCap API Key Apply here: https://coinmarketcap.com/api/ Configure n8n Credentials Use HTTP Header Auth to store your CoinMarketCap API key. Optional: Trigger from a Supervisor Connect to a parent agent using Execute Workflow with message and sessionId inputs. Test Sample Prompts “Get all exchanges”, “Fetch CMC index”, “Show Binance token holdings” Sticky Notes Included Exchange & Community Guide – Explains agent purpose and component connections Usage & Examples – Walkthrough for sample use cases Error Handling & Licensing – Includes API error code reference and licensing details ✅ Final Notes This agent is part of a broader CoinMarketCap AI Analyst System. Visit my Creator profile to download all available sub-agents and supervisor flows. Understand exchange behavior and community sentiment—automated with AI and CoinMarketCap.
by Don Jayamaha Jr
Access real-time cryptocurrency prices, market rankings, metadata, and global stats—powered by GPT-4o and CoinMarketCap! This modular AI-powered agent is part of a broader CoinMarketCap multi-agent system designed for crypto analysts, traders, and developers. It uses the CoinMarketCap API and intelligently routes queries to the correct tool using AI. This agent can be used standalone or triggered by a supervisor AI agent for multi-agent orchestration. Supported API Tools (6 Total) This agent intelligently selects from the following tools to answer your crypto-related questions: 🔍 Tool Summary Crypto Map – Lookup CoinMarketCap IDs and active coins Crypto Info – Get metadata, whitepapers, and social links Crypto Listings – Ranked coins by market cap CoinMarketCap Price – Live prices, volume, and supply Global Metrics – Total market cap, BTC dominance Price Conversion – Convert between crypto and fiat What You Can Do with This Agent 🔹 Get live prices and volume for tokens (e.g., BTC, ETH, SOL) 🔹 Convert crypto → fiat or fiat → crypto instantly 🔹 Retrieve whitepapers, logos, and website links for any token 🔹 Analyze total market cap, BTC dominance, and circulating supply 🔹 Discover new tokens and track their CoinMarketCap IDs 🔹 View the top 100 coins ranked by market cap or volume Example Queries ✅ "What is the CoinMarketCap ID for PEPE?" ✅ "Show me the top 10 cryptocurrencies by market cap." ✅ "Convert 5 ETH to USD." ✅ "What’s the 24h volume for ADA?" ✅ "Get the global market cap and BTC dominance." AI Architecture AI Brain**: GPT-4o-mini Memory**: Session buffer with sessionId Agent Type**: Subworkflow AI tool Connected APIs**: 6 CoinMarketCap endpoints Trigger Mode**: Executes when called by a supervisor (via message and sessionId inputs) Setup Instructions Get a CoinMarketCap API Key Register here: https://coinmarketcap.com/api/ Configure Credentials in n8n Use HTTP Header Auth with your API key for each connected endpoint Connect This Agent to a Supervisor Workflow (Optional) Trigger this agent using Execute Workflow with inputs message and sessionId Test Prompts Try asking: “Convert 1000 DOGE to BTC” or “Top 5 coins in EUR” Included Sticky Notes Crypto Agent Guide – Agent overview, node map, and endpoint details Usage Instructions – Step-by-step usage and sample prompts Error Handling & Licensing – Troubleshooting and IP rights ✅ Final Notes This agent is part of the CoinMarketCap AI Analyst System, which includes multiple specialized agents for cryptocurrencies, exchanges, community data, and DEX insights. Visit my Creator profile to find the full suite of tools. Get smarter about crypto—analyze the market in real time with AI and CoinMarketCap.
by Don Jayamaha Jr
Gain full visibility into decentralized exchanges using CoinMarketCap’s DEXScan API—powered by AI. This workflow is part of the CoinMarketCap AI Analyst system and delivers real-time and historical insights on spot trading pairs, DEX liquidity, trading activity, and OHLCV data across chains like Ethereum, Polygon, Solana, and more. Use this workflow as a sub-agent triggered by a parent supervisor workflow, or run it manually with inputs sessionId and message. 🔧 Supported Tools (8 Total) DEX Metadata → Static info (name, launch date, logo, URLs) DEX Networks List → All supported DEX chains + network metadata DEX Listings Quotes → Ranked list of DEXs with live trading volume, market share DEX Pair Quotes (Latest) → Real-time liquidity, price, and buy/sell stats DEX OHLCV Historical → Time-series data (daily/hourly/1m) DEX OHLCV Latest → Today’s price, volume, open/close for pairs DEX Trades Latest → Up to 100 recent trades for any DEX pair DEX Spot Pairs Latest → Active token pairs across DEXs + filters (volume, liquidity, volatility) Agent Architecture AI Model**: gpt-4o-mini Context Memory**: Window buffer using sessionId Trigger Input**: message, sessionId Execution**: Via Execute Workflow or parent AI supervisor Design**: Tool-based LangChain agent with CMC DEXScan endpoints 💡 Use Cases 🔹 Find top DEXs by 24h volume 🔹 Get spot pairs with highest liquidity on a specific network 🔹 Track historical OHLCV for Uniswap pairs 🔹 View latest trades for SOL/USDC pool 🔹 Analyze tax, pooled % and holders for specific pairs 🔹 Filter pairs by 24h volume, percent change, liquidity, or number of transactions ✅ Example Queries ✅ "Top 5 DEXs by 24h volume on Ethereum" ✅ "Get historical OHLCV for SOL-USDC on Solana" ✅ "Latest trades for a PancakeSwap pair" ✅ "Show all spot pairs with over $500K in liquidity on Polygon" ✅ "Retrieve metadata for Uniswap and SushiSwap" 🛠️ Setup Instructions Get a CoinMarketCap API Key Sign up at: https://coinmarketcap.com/api/ Add API Key to Credentials in n8n Use HTTP Header Auth method Trigger from Parent Workflow (Optional) Use Execute Workflow and pass message and sessionId Test Prompt Ideas Try: "Compare liquidity of Uniswap and Curve pairs on Ethereum" Sticky Notes Included DEXScan Agent Guide – Workflow architecture + supported tools Usage & API Call Examples – Prompts, test inputs, setup flow Error Codes + Licensing – 400/401/429/500 troubleshooting, IP rights ✅ Final Notes This agent is part of the CoinMarketCap AI Analyst System, which includes multiple specialized agents for cryptocurrencies, exchanges, and community data. Visit my Creator profile to find the full suite of tools. Master DEX analytics with AI—get powerful liquidity, trading, and pair insights in seconds.
by irfan saeed
Auto-Generate YouTube Chapters with AI-Powered Transcript Analysis Overview This workflow uses YouTube Data API v3 and Google Gemini 1.5 Flash AI to automatically generate timestamped chapters for videos by analyzing SRT captions. It enhances viewer navigation, improves SEO , and saves creators time by automating manual tasks. Prerequisites YouTube API Setup Create a Google Cloud Project Go to the Google Cloud Console. Click Select a project > New Project and name it (e.g., "YouTube Chapters Automation") . Enable YouTube Data API v3 Navigate to APIs & Services > Library. Search for "YouTube Data API v3" and click Enable . Configure OAuth Consent Screen Go to APIs & Services > OAuth consent screen. Select External (public) or Internal (testing), then add required details (app name, support email) . Generate OAuth 2.0 Credentials Under Credentials, click Create Credentials > OAuth client ID. Choose Web app, then download the JSON key file . Add Credentials to n8n Other Requirements Google Gemini API**: Configure access for the gemini-1.5-flash-8b-exp-0924 model by getting the api key. Workflow Steps Set Video ID Input the target video ID (e.g., r1wqsrW2vmE) using the Set Video ID node. Fetch Video Metadata Use the YouTube API node to retrieve the video’s title, category, and existing description . Download SRT Captions Get Caption ID: Call https://www.googleapis.com/youtube/v3/captions to fetch the caption track ID . Download Transcript: Use the ID to retrieve SRT data via https://www.googleapis.com/youtube/v3/captions/{{ID}}?tfmt=srt . Analyze Transcript with Gemini AI Process the SRT file with Google Gemini AI to identify chapters using a prompt like: "Classify this transcript into timestamped chapters (e.g., 00:00 - Introduction)." Validate output with a structured parser (e.g., Structured Captions node) . Update Video Description Append chapters to the description using the YouTube API’s videos.update method . Value Proposition Viewer Experience**: Chapters improve navigation and reduce drop-off rates . SEO Benefits**: Structured descriptions enhance search visibility . Time Savings**: Eliminates manual chapter creation .
by Sagar
This template streamlines your Instagram content posting workflow by connecting Google Drive for image storage, using OpenAI for AI-generated captions, and leveraging Facebook Graph API for automated publishing. Pre-requisites Before setting up this workflow, ensure you have: A Google account with access to Google Drive An OpenAI API key for AI caption generation A Facebook Developer account with Instagram Graph API access An Instagram Business or Creator account connected to a Facebook Page n8n.io account with workflow access Setup Instructions Configure Data Source Create a Google Sheet with the following columns: Name: Filename of your image in Google Drive Caption: Optional custom caption (leave empty for AI-generated captions) URL: your Video Reel or Image in Google Drive Connect Google Drive Add your Google Drive credentials in the "Google Drive" node Specify the folder path where your Instagram image/Video are stored Configure the node to retrieve image files based on filenames from your Google Sheet Set Up OpenAI Integration Add your OpenAI API key to the credentials Configure the OpenAI node to generate engaging captions based on image content Adjust temperature and model parameters for desired creativity level Configure Facebook Graph API Connect your Facebook account with Instagram access Set up the Facebook Graph API node to post to your Instagram Business/Creator account Ensure proper image formatting (1:1, 4:5, or 16:9 aspect ratios supported by Instagram) Workflow Automation Setup Configure the scheduler node to run at your preferred frequency Set up error handling to notify you of any posting failures Add conditional nodes to use either custom or AI-generated captions Execution Instructions After completing all connections, test the workflow with a single image Monitor the execution in the n8n dashboard to ensure proper functioning View the "Executions" tab to track successful posts and troubleshoot any errors Adjust posting frequency and scheduling as needed This template saves hours of manual Instagram posting work while maintaining an authentic presence. Perfect for social media managers, content creators, and businesses looking to maintain consistent Instagram activity without the daily manual effort. The workflow handles image retrieval, caption generation or customization, proper Instagram API formatting, scheduled posting, and execution tracking - all in one automated solution.
by Łukasz
Who is it for? If you are getting a lot of emails into your Gmail inbox, then probably some of those can be solved easly by replying or by doing specific short tasks. But analyzing whole email thread content just to catch up with multiple threads can be very wasteful. So by using AI you can actually get simple propositions of what should be done before closing this specific email and actual proposed answer to that email. This is especially useful if you need to do some actions before replying to email. In that case you can simply assign task to specific person, await until it's done, copy-paste AI answer when it's done, and close. Another good use would be if on one inbox there are working multiple people. It can make the process much more streamlined. How It Works? Script runs on your selected trigger. If you are using section "Read and Star", then you may use "Email Trigger". Automation is looking for exiting open Todoist tasks, that have the same title as email If task does not exist, then we are asking AI to analyze thread and give output that is Todoist-API-ready: having summary of email content having proposed actions to be taken having proposed answer to this email If email was unstarred for some reason but task was not closed, then task is being closed automatically. Script FOR PURPOSE is not trying to unstar messagess which have closed tasks, because this could lead to some inconsistencies. How to set up? Select and setup your triggers, depending on your needs Setup connections using N8N instructions. You will need: Gmail Todoist AI (in this workflow OpenAI is used) (Optional) Remove "Read and Star" section if you don't want tasks automatically read and starred. (Optional) Adjust AI node - especially useful if you want to use different model or have response in different language NOTE Chat does not heave memory attached on purpose. The purpose is that it should analyze each inbox message separately, not in thread. When using memory, it can get lost easily. NOTE2 You might want to adjust limits on nodes "Get Unread From Inbox", "Get Starred From Inbox" and "Get Open Tasks", especially if having issues with model complying to output structure. And that's it. I hope that this automation will make your Gmail <-> Todoist process much more streamlined! What's More? There is actually more that you could do with this automation, but it really depends on your needs. For example, you could add Form trigger to handle incoming support requests. Another thing is that you could replace Todoist with Asana or any database (like NocoDB) if you are using it for your task management.
by Ranjan Dailata
Who this is for? This workflow is designed for professionals and teams who need real-time, structured insights from Google Search results without manual effort. What problem is this workflow solving? This n8n workflow solves the problem of automating Google Search result extraction, cleanup, summarization, and AI-enhanced formatting for downstream use like sending the results to a webhook or another system. What this workflow does Automates Google Search via Bright Data Uses Bright Data’s proxy-based SERP API to run a Google Search query programmatically. Makes the process repeatable and scriptable with different search terms and regions/zones. Cleans and Extracts Useful Content The Google Search Data Extractor uses LLM based cleaning to remove HTML/CSS/JS from the response and extract pure text data. Converts messy, unstructured web content into structured, machine-readable format. Summarizes Search Results Through the Gemini Flash + Summarization Chain, it generates a concise summary of the search results. Ideal for users who don’t have time to read full pages of search results. Formats Data Using AI Agent The AI Agent acts like a virtual assistant that: Understands search results Formats them in a readable, JSON-compatible form Prepares them for webhook delivery Delivers Results to Webhook Sends the final summary + structured search result to a webhook (could be your app, a Slack bot, Google Sheets, or CRM). Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. A Google Gemini API key (or access through Vertex AI or proxy). Update the Google Search query as you wish by navigating to the Set Google Search Query node. Update the Webhook HTTP Request node with the Webhook endpoint of your choice. How to customize This Workflow to your needs 1. Change the Search Input Default: It searches a fixed query or dataset. Customize: Accept input from a Google Sheet, Airtable, or a form. Auto-trigger searches based on keywords or schedules. 2. Customize Summarization Style (LLM Output) Default: General summary using Google Gemini or OpenAI. Customize: Add tone: formal, casual, technical, executive-summary, etc. Focus on specific sections: pricing, competitors, FAQs, etc. Translate the summaries into multiple languages. Add bullet points, pros/cons, or insight tags. 3.Choose Where the Results Go Options: Email, Slack, Notion, Airtable, Google Docs, or a dashboard. Auto-create content drafts for WordPress or newsletters. Feed into CRM notes or attach to Salesforce leads.