by gotoHuman
๐ผ Lead Outreach Agent This AI workflow helps you quickly react to new leads with an initial personalized outreach. A great start of your lead nurturing sequence to avoid loosing precious leads that could turn into paying customers. Most importantly it uses gotoHuman so you can review the AI-analysis and the AI-generated editable email draft before it is sent out in your name. How it works We receive a new form submission incl. the email address and company name of the prospect and extract the website URL from the address. We proceed only for company email addresses. We scrape the website using Firecrawl and summarize it with OpenAI Our AI agent runs an analysis based on the lead information and documents describing our own company and the defined Ideal Customer Profiles. It also fetches previously approved examples from gotoHuman so you're effectively creating a self-learning agent. It responds with the analysis and the drafted outreach email. Human Approval in gotoHuman. Allows editing the drafted email. We can now send our email including any edits made during the review and be sure that we are using high-quality content instead of AI slop. How to set up Most importantly, install the gotoHuman node before importing this template! (Just add the node to a blank canvas before importing) Set up your credentials for the different services In gotoHuman, select and create the pre-built review template "Lead Outreach Agent" or import the ID: T873fI1Xli5nt3eh33Rj Select this template in the gotoHuman node Requirements You need accounts for gotoHuman (Human Supervision) OpenAI (AI Agent) Typeform (Lead Form Submissions) Firecrawl (Website Scraping) Gmail Google Docs (Company Wiki) How to customize Replace the Typeform trigger with any other way you might receive or find new leads Provide the AI Sales Agent with more context to properly analyze the lead and create better personalized emails. Consider adding tools that allow the agent to fetch more infos about the prospect's company or personal profile, or to find out more about your specific product/service offerings and how your sales pitches look like.
by Khairul Muhtadin
Automatically extract job listings from any website URL, format them with AI, and publish directly to WordPress. Just send a URL via Telegram, and watch as the workflow scrapes the job details, enhances the content with GPT, and creates a polished post on your site. ๐ก Why Use Job Repost? โฐ Save countless hours Automatically extract, process, and publish job offers from any website, freeing your time from repetitive tasks. โ Eliminate human errors Say goodbye to typos and missed fields โ every job post is validated before going live. ๐ Boost engagement Fresh, well-structured job listings attract more candidates, improving your site's reach and authority. ๐ Stay ahead Leveraging AI with GPT means your content is not just automated but polished and SEO-friendly โ the digital assistant you never knew you needed. โก Perfect For Job board managers:** Want to aggregate listings from multiple sources with minimal effort Recruiters & HR teams:** Who need to streamline job posting workflows without technical hassles Content creators & marketers:** Looking to automate publishing while maintaining style and SEO standards ๐ง How It Works | Step | Process | Description | |------|---------|-------------| | ๐ฑ | Trigger | Send a job URL via Telegram bot to initiate the process | | ๐ฅ | Extract | Firecrawl API scrapes and extracts clean content from the provided URL | | ๐ | Process | Job data is extracted via AI, text split and cleaned, job categories and types mapped to your system | | ๐ค | Smart Logic | GPT crafts formatted job posts, intelligent validation ensures all key data is present, default values fill in the blanks if necessary | | ๐ | Output | Posts automatically published to WordPress with company logos uploaded, and success or error notifications sent via Telegram | | ๐ | Storage | Uses Supabase vector store for managing document embeddings, ensuring quick lookup and reference compliance | ๐ Quick Setup Import the provided JSON file into your n8n instances Add credentials: Firecrawl API key Google Drive OAuth2 (for RAG storage) OpenAI API WordPress API Telegram API Supabase Customize: Telegram bot token WordPress URLs Default images and category mappings if needed Update: URLs and API tokens where placeholders are used Test: Send a job URL to your Telegram bot to verify accurate extraction and posting ๐งฉ You'll Need โ Active n8n instances โ Firecrawl account with API access โ Google Drive account for RAG document storage โ OpenAI account with GPT API access โ WordPress site with autojob plugin and API enabled โ Telegram bot for URL submission and notifications โ Supabase account for vector store management ๐ ๏ธ Level Up Ideas ๐ Add multi-language support to expand global reach ๐ Support batch URL processing for multiple jobs at once ๐ฌ Integrate Slack or email notifications for wider team alerts ๐ฏ Use more AI nodes to summarize or rate job offers for quality control ๐ Schedule periodic cleanup of vector store for performance optimization ๐ Add analytics tracking for published jobs performance ๐ง Nodes Used Core Components: Firecrawl HTTP Request** (Web scraping and content extraction) Google Drive** (RAG document storage) Supabase Vector Store** OpenAI** (Embeddings, GPT Extraction) Code Nodes** for mapping categories Telegram Trigger & Message** HTTP Request** (for WordPress API and image uploads) Made by: Khaisa Studio Tags: automation recruitment job-posting wordpress AI web-scraping firecrawl Category: Human Resources, Recruitment, Wordpress, Scrapping Need a custom? contact me on LinkedIn or Web
by Trung Tran
Decodo Scraper API Workflow Template (n8n Automation Amazon Book Purchase Report) Watch the demo video below: > This workflow demos how to use Decodo Scraper API to crawl any public web page (headless JS, device emulation: mobile/desktop/tablet), extract structured product data from the returned HTML, generate a purchase-ready report, and automatically deliver it as a Google Doc + PDF to Slack/Drive. Whoโs it for Creators / Analysts** who need quick product lists (books, gadgets, etc.) with prices/ratings. Ops & Marketing teams** building weekly โtop picksโ reports. Engineers** validating the Decodo Scraper API + LLM extraction pattern before scaling. How it works / What it does Trigger โ Manually run the workflow. Edit Fields (manual) โ Provide inputs: targetUrl (e.g., an Amazon category/search/listing page) deviceType (desktop | mobile | tablet) Optional: maxItems, notes, reportTitle, reportOwner Scraper API Request (HTTP Request โ POST) Calls Decodo Scraper API with: URL to crawl, headless JS enabled Device emulation (UA + viewport) Optional waitFor / executeJS to ensure late-loading content is captured HTML Response Parser (Code/Function or HTML node) Pulls the HTML string from Decodo response and normalizes it (strip scripts/styles, collapse whitespace). Product Analyzer Agent (LLM + Structured Output Parser) Prompts an LLM to extract structured โbookโ objects from the HTML: The Structured Output Parser enforces a strict JSON schema and drops malformed items. Build ๐ Book Purchase Report (Code/LLM) Converts the JSON array into a Markdown (or HTML) report with: Executive summary (top picks, average price/rating) Table of items (rank, title, author, price, rating, link) โRecommended to buyโ shortlist (rules configurable) Notes / owner / timestamp Configure Google Drive Folder (manual) Choose/create a Drive folder for output artifacts. Create Document File (Google Docs API) Creates a Doc from the generated Markdown/HTML. Convert Document to PDF (Google Drive export) Exports the Doc to PDF. Upload report to Slack Sends the PDF (and/or Doc link) to a chosen Slack channel with a short summary. How to set up 1 Prerequisites n8n** (self-hosted or Cloud) Decodo Scraper API** key OpenAI (or compatible) API key** for the Analyzer Agent Google Drive/Docs** credentials (OAuth2) Slack** Bot/User token (files:write, chat:write) 2 Environment variables (recommended) DECODO_API_KEY OPENAI_API_KEY DRIVE_FOLDER_ID (optional default) SLACK_CHANNEL_ID 3 Nodes configuration (high level) Edit Fields (Set node) Scraper API Request (HTTP Request โ POST) HTML Response Parser (Code node) Product Analyzer Agent Build Book Purchase Report (Code/LLM) Create Document File Convert to PDF Upload to Slack Requirements Decodo**: Active API key and endpoint access. Be mindful of concurrency/rate limits. Model**: GPT-4o/4.1-mini or similar for reliable structured extraction. Google**: OAuth client (Docs/Drive scopes). Ensure n8n can write to the target folder. Slack**: Bot token with files:write + chat:write. How to customize the workflow Target site: Change targetUrl to any **public page (category, search, or listing). For other domains (not Amazon), tweak the LLM guidance (e.g., price/label patterns). Device emulation**: Switch deviceType to mobile to fetch mobile-optimized markup (often simpler DOMs). Late-loading pages**: Adjust waitFor.selector or use waitUntil: "networkidle" (if supported) to ensure full content loads. Client-side JS**: Extend executeJS if you need to interact (scroll, click โnextโ, expand sections). You can also loop over pagination by iterating URLs. Extraction schema**: Add fields (e.g., discount_percent, bestseller_badge, prime_eligible) and update the Structured Output schema accordingly. Filtering rules**: Modify recommendation logic (e.g., min ratings count, price bands, languages). Report branding**: Add logo, cover page, footer with company info; switch to HTML + inline CSS for richer Docs formatting. Destinations**: Besides Slack & Drive, add Email, Notion, Confluence, or a database sink. Scheduling: Add a **Cron trigger for weekly/monthly auto-reports.
by Sirhexalot
This n8n workflow allows you to update user roles in Zammad based on data from an Excel file. The workflow automates role assignments, ensuring efficient and consistent updates. Features Excel Integration**: Import user data from an Excel file containing emails and role assignments. Dynamic Updates**: Match Zammad users by email and update their roles. Error Handling**: Continue workflow execution even if some updates fail. Customizable Variables**: Configure Zammad API URL, API key, and Excel file URL. Usage Import the Workflow: Upload the provided .json file into your n8n instance. Set Variables: zammad_base_url: Your Zammad instance URL. excel_source_url: URL of the Excel file containing user data. Authentication for Zammad Create in the Node "Find Zammad User by email" and "Update User Roles" a Header Auth Authentication Name**: Authorization Value**: Bearer <put here your zammad api token> Run the Workflow: Execute the workflow to update user roles based on the Excel data. Issues and Suggestions For issues or suggestions, visit the GitHub Repository.
by Yaron Been
Ibm Granite Granite Speech 3.3 8b Text Generator Description Granite-speech-3.3-8b is a compact and efficient speech-language model, specifically designed for automatic speech recognition (ASR) and automatic speech translation (AST). Overview This n8n workflow integrates with the Replicate API to use the ibm-granite/granite-speech-3.3-8b model. This powerful AI model can generate high-quality text content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Optional Parameters seed** (integer, default: None): Random seed. Leave blank to randomize the seed. audio** (array, default: None): Audio inputs for the model. top_k** (integer, default: 50): The number of highest probability tokens to consider for generating the output. If > 0, only keep the top k tokens with highest probability (top-k filtering). top_p** (number, default: 0.9): A probability threshold for generating the output. If < 1.0, only keep the top tokens with cumulative probability >= top_p (nucleus filtering). Nucleus filtering is described in Holtzman et al. (http://arxiv.org/abs/1904.09751). prompt** (string, default: ): User prompt to send to the model. max_tokens** (integer, default: 512): The maximum number of tokens the model should generate as output. min_tokens** (integer, default: 0): The minimum number of tokens the model should generate as output. temperature** (number, default: 0.6): The value used to modulate the next token probabilities. chat_template** (string, default: None): A template to format the prompt with. If not provided, the default prompt template will be used. system_prompt** (string, default: None): System prompt to send to the model.The chat template provides a good default. How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate text content Access the generated output from the final node API Reference Model: ibm-granite/granite-speech-3.3-8b API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of text generation parameters
by Yaron Been
Google Veo 3 Video Generator Description Sound on: Googleโs flagship Veo 3 text to video model, with audio Overview This n8n workflow integrates with the Replicate API to use the google/veo-3 model. This powerful AI model can generate high-quality video content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Text prompt for video generation Optional Parameters seed** (integer, default: None): Random seed. Omit for random generations resolution** (string, default: 720p): Resolution of the generated video negative_prompt** (string, default: None): Description of what to discourage in the generated video How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate video content Access the generated output from the final node API Reference Model: google/veo-3 API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of video generation parameters
by Yaron Been
Prunaai Hidream E1.1 Image Generator Description Edit an image with a prompt. This is the hidream-e1.1 model accelerated with the pruna optimisation engine. Overview This n8n workflow integrates with the Replicate API to use the prunaai/hidream-e1.1 model. This powerful AI model can generate high-quality image content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Prompt Optional Parameters seed** (integer, default: -1): Random seed (-1 for random) image** (string, default: None): Input image to edit. speed_mode** (string, default: Juiced ๐ฅ (more speed)): Speed optimization level clip_cfg_norm** (boolean, default: True): Whether to use CLIP CFG normalization output_format** (string, default: webp): Output format guidance_scale** (number, default: 2.5): Guidance scale output_quality** (integer, default: 100): Output quality (for jpg and webp) refine_strength** (number, default: 0.3): Strength of refinement num_inference_steps** (integer, default: 28): Number of inference steps image_guidance_scale** (number, default: 1): Image guidance scale How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate image content Access the generated output from the final node API Reference Model: prunaai/hidream-e1.1 API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of image generation parameters
by Yaron Been
Bytedance Seedance 1 Lite Video Generator Description A video generation model that offers text-to-video and image-to-video support for 5s or 10s videos, at 480p and 720p resolution Overview This n8n workflow integrates with the Replicate API to use the bytedance/seedance-1-lite model. This powerful AI model can generate high-quality video content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Text prompt for video generation Optional Parameters fps** (string, default: 24): Frame rate (frames per second) seed** (integer, default: None): Random seed. Set for reproducible generation image** (string, default: None): Input image for image-to-video generation duration** (string, default: 5): Video duration in seconds resolution** (string, default: 720p): Video resolution aspect_ratio** (string, default: 16:9): Video aspect ratio. Ignored if an image is used. camera_fixed** (boolean, default: False): Whether to fix camera position last_frame_image** (string, default: None): Input image for last frame generation. This only works if an image start frame is given too. How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate video content Access the generated output from the final node API Reference Model: bytedance/seedance-1-lite API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of video generation parameters
by bangank36
This workflow automates the Mark as Fulfilled action in Shopify for each order, ensuring a seamless fulfillment process without manual intervention. How It Works This workflow retrieves all unfulfilled orders and processes their fulfillment automatically. Since Shopify requires a Fulfillment Order ID (not Order ID) to trigger fulfillment, the workflow follows these steps: 1๏ธโฃ Get all unfulfilled orders using the Shopify node. 2๏ธโฃ Retrieve the Fulfillment Order ID using the "List Fulfillment Orders" action. 3๏ธโฃ Create a fulfillment request using "Mark fulfillment order as fulfilled." 4๏ธโฃ Handle edge cases, such as partially fulfilled orders or API errors. This ensures that every valid order is marked as fulfilled efficiently. ๐ Ongoing discussions on this topic: Relevant Shopify API Discussion Step-by-step The workflow can be run as requested or on schedule You can adjust these parameters within the Shopify and filter nodes: Shopify Admin URL โ A Global node to customize the Shopify store URL. To find your Shopify store ID, login into your Shopify admin, then look at the URL in your browser's address bar, the subdomain portion (e.g., example_store_id.myshopify.com) is your store ID (in this case: example_store_id) Order Filtering โ Ensures only valid orders are fulfilled, particularly useful if: You sell digital downloads or gift cards exclusively. You use third-party fulfillment services for all products. Credentials To run this workflow, you'll need: Shopify API Key โ Required for authentication. Who Is This For? Shopify store owners looking to automate their fulfillment process. Merchants selling digital or personalized products who need instant fulfillment. Explore More Templates ๐ Check out my other n8n templates
by Yaron Been
Description This workflow automatically searches multiple flight booking websites to find the cheapest flights for your desired routes. It leverages web scraping to compare prices across platforms, helping you save money on air travel. Overview This workflow automatically searches multiple flight booking websites to find the cheapest flights for your desired routes. It uses Bright Data to scrape flight prices and can notify you when prices drop below your target threshold. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping flight prices from booking websites. Notification Services:** Email, SMS, or other messaging platforms. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Notifications: Configure your preferred notification method. Customize: Set your routes, date ranges, and price thresholds. Use Cases Frequent Travelers:** Find the best deals for your regular routes. Travel Agencies:** Monitor flight prices for client bookings. Budget Travelers:** Get notified when flights to your dream destination become affordable. Connect with Me YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #travel #flights #brightdata #dealalerts #webscraping #flightdeals #cheapflights #travelhacks #budgettravel #travelplanning #airfare #flightprices #travelautomation #n8nworkflow #workflow #nocode #traveltech #flightbooking #savemoney #traveltools #flightcomparison #bestflightdeals #travelsmarter #automatedtravel
by n8n Team
This workflow automatically imports data from a CSV file located at a specific URL and then updates the Google Sheets document with the provided data. Below is a step-by-step description of what this workflow does: The workflow is started manually using the "When you click 'Execute Workflow'" node. The CSV file is then uploaded from the specified URL "https://opendata.ecdc.europa.eu/covid19/testing/csv/data.csv" using the "Upload CSV" node. The "Import CSV" node accepts the uploaded CSV file and converts it into JSON formatted data. The "Add Unique Field" node generates a unique key by combining the 'country_code' and 'year_week' fields from the JSON data, which will be further used in the Google Sheets document. The 'Keep only DACH in 2023' node filters the data to keep only records where 'country_code' is either 'DE', 'AT', or 'CH' and 'year_week' starts with '2023'. Google's API has limitations on the speed of read and write operations, so only a subset of the data is taken. The filtered data is loaded into the specified Google Sheets document via the 'Load to Spreadsheet' node. The operation is set to 'appendOrUpdate' and the document ID and sheet name are specified. Also, the previously generated 'unique_key' key is set as the key to match the columns.
by Yaron Been
Aihilums Sehatsanjha AI Generator Description None Overview This n8n workflow integrates with the Replicate API to use the aihilums/sehatsanjha model. This powerful AI model can generate high-quality other content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Optional Parameters cookie** (string, default: None): Cookie to be returned unchanged user_id** (string, default: ): Unique session identifier audio_file** (string, default: None): Audio file user_state** (string, default: None): User state end_session** (boolean, default: False): End the current recording new_session** (boolean, default: False): Start a new recording How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate other content Access the generated output from the final node API Reference Model: aihilums/sehatsanjha API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of other generation parameters