by Davide
This workflow allows you to send SMS messages globally using API without needing a physical phone number. 1. How It Works Consists of three main nodes: Manual Trigger**: The workflow starts when you click the "Test workflow" button in n8n. Set SMS Data**: This node defines the SMS message content and the recipient's phone number (including the international prefix). Send SMS**: This node sends the SMS using the ClickSend API. It uses HTTP Basic Authentication with your ClickSend credentials and sends a POST request to the ClickSend API endpoint with the message and recipient details. The workflow is simple and efficient, making it easy to automate SMS sending for various use cases, such as notifications, alerts, or marketing campaigns. 2. Set Up Steps To set up and use this workflow in n8n, follow these steps: Register on ClickSend: Go to ClickSend and create an account. Obtain your API Key and take advantage of the 2 € free credits provided. Set Up Basic Authentication in n8n: In the "Send SMS" node, configure the HTTP Basic Auth credentials: Username: Use the username you registered with on ClickSend. Password: Use the API Key provided by ClickSend. Configure the SMS Data: In the "Set SMS data" node, define: The message content (e.g., "Hi, this is my first message"). The recipient's phone number, including the international prefix (e.g., +39xxxxxxxx). Test the Workflow: Click the "Test workflow" button in n8n to trigger the workflow. The workflow will send the SMS using the ClickSend API, and you should receive the message on the specified phone number. Optional Customization: You can modify the workflow to dynamically set the message content or recipient phone number using data from other nodes or external sources. This workflow is a quick and efficient way to send SMS messages programmatically. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Mario
Template to get your public IP address and push it to Namecheaps Dynamic DNS per subdomain. Open "yourdomain.com" Insert your domain and your Namecheap DDNS password Open "subdomains" Replaces and insert your subdomains Execute Workflow Have fun!
by Agent Circle
This n8n template demonstrates walks you through a fully automated process to generate faceless videos - from script creation to final download - using AI-generated voice, images, and smart video editing. Use cases are many: This tool is perfect for YouTube and Shorts creators who want to publish daily content without showing their face, TikTok and Reels marketers automating voice-over-driven videos, and solopreneurs scaling up their content without hiring a team. It’s also ideal for agencies producing batches of faceless video ads, automation enthusiasts building smart media workflows in n8n, and anyone who’s rich in ideas but tired of spending hours editing. How It Works Phase 1: Provide Topic Input A short topic and idea should be entered into the Idea part in Node Fields - Set Idea inside the workflow in n8n. Trigger the process manually by clicking Test Workflow or Execute Workflow. Phase 2: Script Generation Your idea is passed to Google Gemini's chat model. The model returns a concise, 60-second faceless video script. The script is then reformatted into a structured layout optimized for voice generation and visual synchronization. Phase 3: Audio Generation The formatted script is passed to ElevenLabs, which turns the text into a high-quality voiceover audio. The generated audio is uploaded to Google Drive and made publicly accessible. At the same time, the audio is sent to OpenAI Whisper via a POST request to generate a transcription of the voiceover. Phase 4: Timestamps Generation The tool merges the original script and the OpenAI Whisper-generated transcription. The merged data is passed to Google Gemini's chat model to generate image prompts with precise timestamps. The output is parsed and cleaned using a structured parser to ensure it's in ready-to-use JSON format for image generation. Phase 5: Images Generation The full list of timestamped prompts is is split into individual entries. Each prompt is sent to Leonardo's API that turns text descriptions into visuals. A delay of 30 seconds is added to give the image generation engine enough time to complete rendering. Once completed, the workflow retrieves all final images for the next stage. Phase 6: Images To Video Conversion All generated images are sent to Leonardo's API, which stitches them together based on the structured prompts and timing. A 5-minute wait allows time for rendering. After the wait, the workflow retrieves the generated small videos and makes them downloadable. Then, the tool aggregates all downloaded videos into a single unified structure, preparing them for the final editing. Phase 7: Video Editing and Downloading The raw video, along with timestamps or subtitles, is sent to Shotstack, a video editing tool that supports advanced edits. A delay of 1 minute allows Shotstack to process the edit. Then, the tool checks whether the edited video is finished by Shotstack and ready to be downloaded. Once completed, you can download the final polished video to your local storage for later use. How To Use Download the workflow package. Import the package into your n8n interface. Set up necessary credentials for tools access and usability: For Google Gemini access, please connect to its API in the following nodes: Node Google Gemini Chat Model 1 Node Google Gemini Chat Model 2 For Google Drive access, please ensure connection in the following nodes: Node Upload Audio to Drive Node Make Audio File Public For ElevenLabs access, please connect to its API in the following node: Node Generate Voice For OpenAI Whisper access, please connect to its API in the following node: Node Transcribe Audio with OpenAI Whisper For Leonardo access, please allow connection to its API in the following nodes: Node Generate Images Node Generate Videos/Scenes For Shortstack access, please connect to its API in the following nodes: Node Edit with Shotstack Node Render Final Video with Shotstack Input your video idea or short description as a string in Node Fields - Set Idea in n8n. Run the workflow by clicking Execute Workflow or Test Workflow. Wait the process to run and finish. View the result in Node Download Final Video and download it in your local storage for later use. Requirements Basic setup in Google Cloud Console (OAuth or API Key method enabled) with enabled access to Google Drive. Google Gemini API** access with permission to use chat-based large language models. ElevenLabs API** access for generating high-quality voiceovers from scripts. OpenAI Whisper API** access to transcribe voiceovers into clean text. Leonardo API** access for both image and video generation tasks. Shotstack API** access for editing and rendering the final video with enhanced visuals and timing. How To Customize You can input your requested video topic or description directly in Node Fields – Set Idea. By default, the script length is set to around 60 seconds in Node 60 Second Script Writer. You can easily change this in the prompt to create shorter or longer videos based on your needs. While the default setup uses Google Gemini for script and prompt generation, you can replace it with OpenAI ChatGPT, Claude, or any other compatible chat-based model you prefer. The voiceover is currently created using ElevenLabs, but you’re free to substitute it with other text-to-speech engines like Google Cloud Text-to-Speech, HeyGen, etc. We're using OpenAI Whisper to transcribe the voiceover into text. You can switch to alternatives such as AssemblyAI, Deepgram, or other compatible providers depending on your preference. This workflow uses Leonardo for both image and video generation. You can swap it out for other compatible providers based on availability or style preference. Video editing is handled by Shotstack by default. You can plug in alternatives like Runway, FFmpeg, or other API-based editors depending on your editing needs or desired effects. If you’d like this workflow customized to fit your tools and platforms availability, or if you’re looking to build a tailored AI Agent for your own business - please feel free to reach out to Agent Circle. We’re always here to support and help you to bring automation ideas to life. Need Help? Join our community on different platforms for support, inspiration and tips from others. Website: https://www.agentcircle.ai/ Etsy: https://www.etsy.com/shop/AgentCircle Gumroad: http://agentcircle.gumroad.com/ Discord Global: https://discord.gg/d8SkCzKwnP FB Page Global: https://www.facebook.com/agentcircle/ FB Group Global: https://www.facebook.com/groups/aiagentcircle/ X: https://x.com/agent_circle YouTube: https://www.youtube.com/@agentcircle LinkedIn: https://www.linkedin.com/company/agentcircle
by Yaron Been
This workflow automatically monitors competitor pricing across multiple products and services to track market positioning and pricing strategies. It saves you time by eliminating the need to manually check competitor prices and provides real-time insights into pricing changes and market trends. Overview This workflow automatically scrapes competitor websites and pricing pages to extract current pricing information, product details, and promotional offers. It uses Bright Data to access pricing data without restrictions and AI to intelligently parse pricing information and detect changes over time. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping competitor pricing pages without being blocked OpenAI**: AI agent for intelligent pricing data extraction and analysis Google Sheets**: For storing pricing data and tracking changes over time How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your pricing tracking spreadsheet Customize: Define competitor URLs and pricing monitoring parameters Use Cases Pricing Strategy**: Stay competitive by monitoring market pricing trends Product Management**: Track competitor feature and pricing changes Sales Teams**: Provide up-to-date competitive pricing information Market Research**: Analyze pricing patterns and market positioning Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #pricemonitoring #competitorpricing #brightdata #webscraping #pricinganalysis #n8nworkflow #workflow #nocode #competitoranalysis #pricingdata #marketresearch #pricingtrends #competitiveintelligence #pricingtracking #marketanalysis #pricecomparison #competitormonitoring #businessintelligence #pricingstrategy #marketpositioning #pricinginsights #competitorresearch #pricingautomation #markettrends #pricealerts #dynamicpricing #pricingoptimization #competitivepricing
by David Harvey
🔮 Mystic Tarot Bot — AI-Powered iMessage Readings This magical n8n template turns your iMessage inbox into a soulful tarot reading experience powered by Blooio and AI. Users can send in questions or photos of their tarot spreads, and the bot replies like a mystical oracle — interpreting symbols, offering gentle insights, and guiding with poetic warmth. ✨ Ideal for solo reflection, spiritual creators, or client-based guidance services — no tech knowledge needed. 🌟 Use Cases Offer intuitive, emotionally resonant tarot readings via iMessage Support coaching, wellness, and metaphysical businesses with AI-enhanced readings Accept photos of real tarot card spreads or plain text questions Great for automating daily card pulls, client responses, or onboarding into spiritual flows 🧠 Good to Know Built using Blooio’s iMessage API — supports image attachments and conversational replies Includes visual recognition and symbolic interpretation of real tarot card photos Responses generated by OpenAI with a custom “Mystic Tarot Reader” persona Onboards users if they say “Hi” or request a virtual card draw Responds in poetic, spiritually attuned language — no markdown, no tech-speak ⚙️ How it Works Trigger: iMessage webhook via Blooio receives user message or image Check: Bot ignores self-sent messages to prevent loops Detect: If a photo is attached, it’s passed to AI for card recognition Interpret: The AI agent gives a heartfelt, symbolic interpretation Respond: A final, warm tarot reading is sent back through iMessage 📝 How to Use Set Up Blooio: Sign up at https://blooio.com Choose a Dedicated or Enterprise plan (image support required) Copy your API key from Settings → API Keys Paste it into the Send Message HTTP node as a Bearer token Customize the Experience: Adjust the prompt for a different tone or deck style Add journaling prompts, affirmations, or follow-ups Use other workflows to track users, create reading logs, or offer upsells Try It Out: Text your Blooio-connected number with: “Hi” → get onboarding “Draw a card for me” → get a virtual pull A tarot photo + question → get a full, soulful reading ✅ Requirements Blooio Account & API Token (Dedicated plan or higher for images) Optional: Tarot images, user questions, or both 🔧 Customizing This Workflow Add personalized spreads (e.g. past/present/future layouts) Send AI-generated visuals of the pulled cards Route readings into Notion, Airtable, or Google Sheets Expand to WhatsApp, web, or email with Blooio’s multichannel support 🃏 Let the cards speak. Let the messages flow.
by Itamar
🕵️ Company Research Agent (n8n + Explorium + LLM) This n8n workflow automates company research by combining Explorium’s MCP server, web scraping tools, and an AI agent. Results are written to a Google Sheet for easy use in GTM, product analysis, or competitive research. 🚀 What It Does Given a list of company domains or names, this workflow will: Look up company information using: 🧠 LLM Agent to guide the research 🔎 Explorium MCP Server for firmographic & tech signals 🌐 Website content and SerpAPI scraping (optional) Extract key commercial details (see below) Format the output in a consistent JSON structure Update a connected Google Sheet with the enriched results 🧩 Extracted Fields Each company is enriched with: domain linkedinUrl has_free_trial cheapest_plan has_enterprise_plan last_case_study_link market (e.g., B2B or B2C) integrations (e.g., Slack, Hubspot, MySQL) enrichment_status 📥 Input Sheet Format | input | |-------------| | Explorium | | n8n | | Apple | | ... | 📤 Output Sheet Format | domain | linkedinUrl | has_free_trial | cheapest_plan | has_enterprise_plan | last_case_study_link | market | integrations | enrichment_status | |--------------|----------------------------------|----------------|----------------|----------------------|-----------------------------|--------|---------------------------------------------------|-------------------| | Explorium.ai | https://linkedin.com/company/... | TRUE | 69 | TRUE | https://www.explorium.com | B2B | ["HubSpot", "Zapier", "Salesforce", ...] | done | | n8n.io | https://linkedin.com/company/... | TRUE | 20 | TRUE | https://n8n.io/case-studies | B2B | ["Slack", "Gmail", "MySQL", "Google Sheets", ...] | done | 🛠️ Tools Used n8n** (Automation platform) Explorium MCP Server** – rich company enrichment via API Anthropic Claude or OpenAI** – used by the AI researcher Google Sheets** – stores output data Structured Output Parser** – ensures clean, predictable JSON formatting 📦 How to Set It Up Add your company domains or names to the input sheet Configure your MCP and SerpAPI credentials in n8n Run the workflow using the Test Workflow trigger Watch the sheet populate with results You can adapt the system to output different formats or fields depending on your team's research goals. 📌 Use Cases Competitive landscape analysis Lead intelligence for outbound campaigns Feature benchmarking (e.g., who offers enterprise or free trial) VC/investment research 🧠 Notes This agent is easily customizable. Adjust the LLM prompt or Output Parser to extract different properties. Explorium MCP is leveraged as the core enrichment engine, ensuring signal accuracy and freshness.
by Weilun
🔄 n8n Workflow: Check and Update n8n Version This workflow automatically checks if the local n8n version is outdated and, if so, creates a file to signal an update is needed. 🖥️ Working Environment Operating System:** Ubuntu 24.04 n8n Installation:** Docker container 📁 Project Directory Structure n8n/ ├── check_update.txt ├── check-update.sh ├── compose.yml ├── update_n8n.cron 🧾 File Descriptions check_update.txt Contains a single word: true: Update is needed false: No update required check-update.sh #!/bin/bash PATH=/usr/local/sbin:/usr/local/bin:/usr/sbin:/usr/bin:/sbin:/bin if grep -q "true" /home/sysadmin/n8n/check_update.txt; then Place your update logic here echo "Update needed - please insert update logic" echo true > /home/sysadmin/n8n/check_update.txt fi Purpose: Checks the contents of check_update.txt If it contains true, executes update logic (currently a placeholder) Resets check_update.txt to true update_n8n.cron SHELL=/bin/sh 10 5 * * * /bin/sh /home/sysadmin/n8n/check-update.sh Purpose: Runs the check-update.sh script daily at 5:10 AM Uses /bin/sh as the shell environment 🧩 n8n Workflow Breakdown 1. Schedule Trigger 🕓 Purpose:** Triggers the workflow every day at 5:00 AM Node Type:** Schedule Trigger 2. Get the latest n8n version 🌐 Purpose:** Fetches the latest version of n8n from npm Endpoint:** https://registry.npmjs.org/n8n/latest Node Type:** HTTP Request 3. Get Local n8n version 🖥️ Purpose:** Retrieves the currently running n8n version Endpoint:** http://192.168.100.18:5678/rest/settings Node Type:** HTTP Request 4. If 🔍 Purpose:** Compares the local and latest versions Condition:** If not equal → update is needed 5. SSH1 🧾 Purpose:** Writes the result to a file on the host via SSH Logic:** echo "{{ $('If').params.conditions ? 'false' : 'true' }}" > check_update.txt Effect: Updates check_update.txt with "true" if an update is needed, "false" otherwise. 🛠️ Setting up Crontab on Ubuntu 1. Register the cron job with: crontab update_n8n.cron 2. Verify that your cron job is registered: crontab -l ✅ Result 5:00 AM** – n8n workflow checks versions and writes result to check_update.txt 5:10 AM** – Cron runs check-update.sh to respond to update flag
by Gulfiia
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Automated Data-Driven UX Persona Creation – Try It Out!* About You can create personas based on your website, region, and industry. Unlike traditional persona creation, this process uses reliable data sources and can estimate market size for each persona. UX personas have a wide range of applications: use them to better define your target users during product development, align your team around user goals during workshops, or inspire new features and ideas by deeply understanding user needs and behaviors. How It Works The flow is triggered via a web form Perplexity analyzes the market and creates a data foundation for the personas An AI agent transforms the data into detailed persona descriptions and publishes them in a Google Doc We use DALL·E 3 to generate an image for each persona, which is saved to your Google Drive How To Use Import the package into your N8N interface Set up the credentials in each node to access the necessary tools Wait for the process to run (it takes just a few seconds) Check the final output in Google Docs and your Google Drive Requirements Perplexity for research OpenAI for LLM and Image generation Google Doc Google Drive to upload images
by Max Tkacz
This n8n workflow template lets teams easily generate a custom AI chat assistant based on the schema of any Notion database. Simply provide the Notion database URL, and the workflow downloads the schema and creates a tailored AI assistant designed to interact with that specific database structure. Set Up Watch this quick set up video 👇 Key Features Instant Assistant Generation**: Enter a Notion database URL, and the workflow produces an AI assistant configured to the database schema. Advanced Querying**: The assistant performs flexible queries, filtering records by multiple fields (e.g., tags, names). It can also search inside Notion pages to pull relevant content from specific blocks. Schema Awareness**: Understands and interacts with various Notion column types like text, dates, and tags for accurate responses. Reference Links**: Each query returns direct links to the exact Notion pages that inform the assistant’s response, promoting transparency and easy access. Self-Validation**: The workflow has logic to check the generated assistant, and if any errors are detected, it reruns the agent to fix them. Ideal for Product Managers**: Easily access and query product data across Notion databases. Support Teams**: Quickly search through knowledge bases for precise information to enhance support accuracy. Operations Teams**: Streamline access to HR, finance, or logistics data for fast, efficient retrieval. Data Teams**: Automate large dataset queries across multiple properties and records. How It Works This AI assistant leverages two HTTP request tools—one for querying the Notion database and another for retrieving data within individual pages. It’s powered by the Anthropic LLM (or can be swapped for GPT-4) and always provides reference links for added transparency.
by n8n Team
This workflow integrates both web scraping and NLP functionalities. It uses HTML parsing to extract links, HTTP requests to fetch essay content, and AI-based summarization using GPT-4o. It's an excellent example of an end-to-end automated task that is not only efficient but also provides real value by summarizing valuable content. Note that to use this template, you need to be on n8n version 1.50.0 or later.
by Evoort Solutions
✨ AI Text Summarizer with Google Sheets Logging Summarize large blocks of text into concise outputs using the Text Summarizer AI API and automatically log results in Google Sheets. This automation is ideal for content creators, marketers, researchers, and teams who need efficient summarization and record-keeping—without writing a single line of code. 🚀 Features | Feature | Description | |--------------------------|-----------------------------------------------------------------------------| | Web Form Trigger | Collects title, content, mode (Paragraph/Bullet), and length preferences. | | API Integration | Sends content to Text Summarizer AI API via RapidAPI. | | Conditional Logic | Routes success and error cases appropriately. | | Google Sheets Logging| Stores summaries or error messages into Google Sheets. | | Error Handling | Captures failed summaries and ensures no data is lost. | 🧠 Problem Solved Manually summarizing long-form content is time-consuming and repetitive. Storing this output in structured logs (like Google Sheets) adds an extra layer of manual effort. This workflow solves that by: 🔁 Automating AI-driven summarization using Text Summarizer AI API. 📊 Storing results (or fallback errors) directly into a Google Sheet. ✅ Ensuring no request goes unlogged—even on API failure. 🎯 Use Cases 📝 Blog Writers: Quickly convert articles into summaries for social captions. 📚 Students: Break down textbook or lecture content. 🧠 Knowledge Management: Turn raw meeting notes into concise summaries. 📢 SEO Teams: Use bullet-point output for schema markup or meta descriptions. 🧩 Nodes in the Flow | Node Name | Purpose | |----------------------|-------------------------------------------------------------------------| | On form submission | Captures user input via form (title, content, mode, length). | | Mapping | Formats input to match Text Summarizer AI API specs. | | HTTP Request | Sends POST request to the summarization API on RapidAPI. | | If | Validates whether a summary was returned. | | Wait | Adds short delay before writing to spreadsheet (success). | | Google Sheets | Appends summary data to Google Sheet. | | Wait1 | Adds delay for error handling path. | | Google Sheets1 | Logs failure with an “Error occurred” message. | ✅ Benefits 🧠 AI-Powered: Uses Text Summarizer AI API for fast and contextual summaries. 📁 Organized Logs: Google Sheets integration ensures easy tracking and auditing. 🔐 Reliable: Captures all submissions, including failed ones. 🧩 Customizable: Easily adapt inputs or connect to other tools like Notion, Slack, or Airtable. 🛠️ Requirements ✅ n8n account (Cloud or Self-Hosted) ✅ Access to Text Summarizer AI API ✅ Google account (for Sheets integration) 🔑 API key from RapidAPI (used in the HTTP Request node) 🔄 Workflow Overview User submits text and preferences (mode, length) via a form. Workflow triggers and transforms the input. Formatted data is sent to the Text Summarizer AI API. If a valid summary is returned: Log it into Google Sheets. If the API fails: Log the error message instead. 📥 Import Instructions Open n8n and import the workflow JSON. Replace the x-rapidapi-key in the HTTP node with your personal RapidAPI key. Configure your Google Sheets credentials. Deploy and test. ✅ 🔗 Suggested Extensions 🟢 Notion database logging 🟣 Slack/Discord notification for each summary 🔵 CSV download or Airtable sync 🏷 Tags ai summarization text-processing rapidapi google-sheets automation markdown n8n Text Summarizer AI API Create your free n8n account and set up the workflow in just a few minutes using the link below: 👉 Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!
by scrapeless official
Brief Overview This workflow integrates Linear, Scrapeless, and Claude AI to create an AI research assistant that can respond to natural language commands and automatically perform market research, trend analysis, data extraction, and intelligent analysis. Simply enter commands such as /search, /trends, /crawl in the Linear task, and the system will automatically perform search, crawling, or trend analysis operations, and return Claude AI's analysis results to Linear in the form of comments. How It Works Trigger: A user creates or updates an issue in Linear and enters a specific command (e.g. /search competitor analysis). n8n Webhook: Listens to Linear events and triggers automated processes. Command identification: Determines the type of command entered by the user through the Switch node (search/trends/unlock/scrape/crawl). Data extraction: Calls the Scrapeless API to perform the corresponding data crawling task. Data cleaning and aggregation: Use Code Node to unify the structure of the data returned by Scrapeless. Claude AI analysis: Claude receives structured data and generates summaries, insights, and recommendations. Result writing: Writes the analysis results to the original issue as comments through the Linear API. Features Multiple commands supported /search: Google SERP data query /trends: Google Trends trend analysis /unlock: Unlock protected web content (JS rendering) /scrape: Single page crawling /crawl: Whole site multi-page crawling Claude AI intelligent analysis Automatically structure Scrapeless data Generate executable suggestions and trend insights Format optimization to adapt to Linear comment format Complete automation process Codeless process management based on n8n Multi-channel parallel logic distribution + data standardization processing Support custom API Key, regional language settings and other parameters Requirements Scrapeless API Key**: Scrapeless Service request credentials. Log in to the Scrapeless Dashboard Then click "Setting" on the left -> select "API Key Management" -> click "Create API Key". Finally, click the API Key you created to copy it. n8n Instance**: Self-hosted or n8n.cloud account. Claude AI**: Anthropic API Key (Claude Sonnet 3.7 model recommended) Installation Log in to Linear and get a Personal API Token Log in to n8n Cloud or a local instance Import the n8n workflow JSON file provided by Scrapeless Configure the following environment variables and credentials: Linear API Token Scrapeless API Token Claude API Key Configure the Webhook URL and bind to the Linear Webhook settings page Usage This automated job finder agent is ideal for: | Industry / Role | Use Case | | --------------------------------- | -------------------------------------------------------------------------------------------------- | | SaaS / B2B Software | | | Market Research Teams | Analyze competitor pricing pages using /unlock, and feature pages via /scrape. | | Content & SEO | Discover trending keywords and SERP data via /search and /trends to guide content topics. | | Product Managers | Use /crawl to explore product documentation across competitor sites for feature benchmarking. | | AI & Data-Driven Teams | | | AI Application Developers | Automate info extraction + LLM summarization for building intelligent research agents. | | Data Analysts | Aggregate structured insights at scale using /crawl + Claude summarization. | | Automation Engineers | Integrate command workflows (e.g., /scrape, /search) into tools like Linear to boost productivity. | | E-commerce / DTC Brands | | | Market & Competitive Analysts | Monitor competitor sites, pricing, and discounts with /unlock and /scrape. | | SEO & Content Teams | Track keyword trends and popular queries via /search and /trends. | | Investment / Consulting / VC | | | Investment Analysts | Crawl startup product docs, guides, and support pages via /crawl for due diligence. | | Consulting Teams | Combine SERP and trend data (/search, /trends) for fast market snapshots. | | Media / Intelligence Research | | | Journalists & Editors | Extract forum/news content from platforms like HN or Reddit using /scrape. | | Public Opinion Analysts | Monitor multi-source keyword trends and sentiment signals to support real-time insights. | Output