by Ranjan Dailata
Who this is for? Indeed Data Scraper & Summarization with Airtable, Bright Data and Google Gemini is an automated workflow that extracts company profile information from Indeed using Bright Data Web Unlocker, transforms the data using Google Gemini's LLM, and forward the transformed response with the summary to a specified webhook for downstream use. This workflow is tailored for: Recruiters and HR teams who want quick summaries of companies listed on Indeed. Market researchers and analysts needing structured insights into businesses. Founders, investors, and consultants scouting potential competitors, partners, or clients. No-code enthusiasts looking to automate data extraction and enrichment pipelines without manual scraping or parsing. What problem is this workflow solving? Manually gathering structured information about companies on Indeed is time-consuming and inconsistent. Pages vary in structure, and extracting clean, digestible summaries can require technical scraping expertise. This workflow automates: Extracting company data from Indeed reliably using Bright Data Web Unlocker. Cleaning and summarizing the extracted content using Google Gemini LLM. Storing structured insights directly into Airtable for easy access and further workflows. Eliminates manual research, saves hours, and produces AI-enhanced, easily searchable records. What this workflow does Triggers on-demand. Pulls company page URLs from Airtable. Scrapes content from each Indeed company profile using Bright Data Web Unlocker. Sends the raw HTML to Google Gemini for extraction and summarization. Sends the summarized data to other platforms via a Webhook notification mechanism. Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials for Bright Data. The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. In n8n, configure the Google Gemini(PaLM) Api account with the Google Gemini API key (or access through Vertex AI or proxy). In n8n, configure the Airtable Personal Access Token account under Credentials. Update the Webhook Notifier with the Webhook endpoint of your choice. How to customize this workflow to your needs This workflow is built to be flexible - whether you're a company or a market researcher, entrepreneur, or data analyst. Here's how you can adapt it to fit your specific use case: Extend the scraper**: Modify Bright Data targets to pull job listings, salaries, or employee reviews via the Airtable data source. Customize the summary prompt**: Ask Gemini to extract different attributes hiring trends, practices etc. Routing the output to different destinations**: Send summaries or transformed response to Google Sheets, Airtable, or CRMs like HubSpot or Salesforce etc.
by Julian Kaiser
This automated workflow scrapes and processes the monthly "Who is Hiring" thread from Hacker News, transforming raw job listings into structured data for analysis or integration with other systems. Perfect for job seekers, recruiters, or anyone looking to monitor tech job market trends. How it works Automatically fetches the latest "Who is Hiring" thread from Hacker News Extracts and cleans relevant job posting data using the HN API Splits and processes individual job listings into structured format Parses key information like location, role, requirements, and company details Outputs clean, structured data ready for analysis or export Set up steps Configure API access to [Hacker News](https://github.com/HackerNews/API ) (no authentication required) Follow the steps to get your cURL command from https://hn.algolia.com/ Set up desired output format (JSON structured data or custom format) Optional: Configure additional parsing rules for specific job listing information Optional: Set up integration with preferred storage or analysis tools The workflow transforms unstructured job listings into clean, structured data following this pattern: Input: Raw HN thread comments Process: Extract, clean, and parse text Output: Structured job listing data This template saves hours of manual work collecting and organizing job listings, making it easier to track and analyze tech job opportunities from Hacker News's popular monthly hiring threads.
by Yaron Been
π Scrape Glassdoor with Bright Data Designed for sales teams, recruiters, and marketers aiming to automate job discovery and prospecting. This workflow scrapes Glassdoor job listings using Bright Data and automatically generates targeted pitches using AI, streamlining lead identification and outreach. π§© How It Works This automation leverages n8n, Bright Data, Google Sheets, and OpenAI: 1. Trigger Starts with a custom form input (Location, Keyword, Country). 2. Bright Data Job Scrape Triggers a Bright Data dataset snapshot via HTTP Request. Polls snapshot progress using a Wait node, ensuring data readiness. Retrieves full job listings dataset once ready. 3. Google Sheets Integration Writes detailed job data (company, role, location, overview, metrics) into a Google Sheet. Uses a pre-built template for organized data storage. 4. Automated Pitch Generation (AI) Splits listings into actionable parts: company name, title, and description. Sends data to OpenAI (via LangChain) to generate relevant pitches or icebreakers. Saves generated content back into the same sheet for easy access. β Requirements Ensure you have the following: Google Sheets Google account Template Sheet with columns for job details and AI-generated pitches Bright Data Active account with Dataset API access API key and dataset ID OpenAI Valid OpenAI API key for GPT models n8n Environment Nodes: HTTP Request, Wait, If, Google Sheets, Split Out, LangChain (OpenAI) Credentials: Google Sheets OAuth2 Bright Data API credentials OpenAI API key βοΈ Setup Instructions Step 1: Prepare Google Sheets Copy the provided Google Sheets template Do not change headers Step 2: Import & Configure Workflow in n8n Import the workflow JSON file Set Google Sheets node: Link to your copied sheet Confirm correct tab name Step 3: Configure Bright Data Replace <YOUR_BRIGHT_DATA_API_KEY> with your real key Set your dataset ID in all HTTP Request nodes Step 4: Configure OpenAI (LangChain) Connect OpenAI API key to the LangChain node Customize prompt to match tone and outreach style Step 5: Testing & Scheduling Test via manual form trigger Schedule runs or leave form enabled for on-demand use π§ Tips & Best Practices Use specific keywords and locations for better results Adjust polling intervals based on dataset size Refine AI prompts regularly to improve pitch quality Clean unused columns from your sheet to boost performance π¬ Support & Feedback For help or customization: π§ Email: Yaron@nofluff.online πΊ YouTube: @YaronBeen π LinkedIn: linkedin.com/in/yaronbeen π Bright Data Docs: docs.brightdata.com/introduction
by David Olusola
A complete, ready-to-deploy Telegram chatbot template for food delivery businesses. This intelligent assistant handles orders, payments, customer service, and order tracking with human-in-the-loop payment verification. β¨ Key Features π€ AI-Powered Conversations - Natural language order processing using Google Gemini π± Telegram Integration - Seamless customer interaction via Telegram π³ Payment Verification - Screenshot-based payment confirmation with admin approval π Order Tracking - Automatic Google Sheets logging of all orders π§ Memory Management - Contextual conversation memory for better customer experience π Multi-Currency Support - Easily customizable for any currency (USD, EUR, GBP, etc.) π Location Flexible - Adaptable to any city/country π Human Oversight - Manual payment approval workflow for security π οΈ What This Template Includes Core Workflow Customer Interaction - AI assistant takes orders via Telegram Order Confirmation - Summarizes order with total and payment details Information Collection - Gathers customer name, phone, and delivery address Payment Processing - Handles payment screenshots and verification Admin Approval - Human verification of payments before order confirmation Order Tracking - Automatic logging to Google Sheets with delivery estimates Technical Components AI Agent Node - Google Gemini-powered conversation handler Memory System - Maintains conversation context per customer Google Sheets Integration - Automatic order logging and tracking Telegram Nodes - Customer and admin communication Payment Verification - Screenshot detection and approval workflow Conditional Logic - Smart routing based on message types π Quick Setup Guide Prerequisites n8n instance (cloud or self-hosted) Telegram Bot Token Google Sheets API access Google Gemini API key Step 1: Replace Placeholders Search and replace the following placeholders throughout the template: Business Information [YOUR_BUSINESS_NAME] β Your restaurant/food business name [ASSISTANT_NAME] β Your bot's name (e.g., "Alex", "Bella", "Chef Bot") [YOUR_CITY] β Your city [YOUR_COUNTRY] β Your country [YOUR_ADDRESS] β Your business address [YOUR_PHONE] β Your business phone number [YOUR_EMAIL] β Your business email [YOUR_HOURS] β Your operating hours (e.g., "9AM - 11PM daily") Currency & Localization [YOUR_CURRENCY] β Your currency name (e.g., "USD", "EUR", "GBP") [CURRENCY_SYMBOL] β Your currency symbol (e.g., "$", "β¬", "Β£") [YOUR_TIMEZONE] β Your timezone (e.g., "EST", "PST", "GMT") [PREFIX] β Order ID prefix (e.g., "FB" for "Food Business") Menu Items (Customize Completely) [CATEGORY_1] β Food category (e.g., "Burgers", "Pizza", "Sandwiches") [ITEM_1] through [ITEM_8] β Your menu items [PRICE_1] through [DELIVERY_FEE] β Your prices Add or remove categories and items as needed Payment & Support [YOUR_PAYMENT_DETAILS] β Your payment information [YOUR_PAYMENT_PROVIDER] β Your payment method (e.g., "Venmo", "PayPal", "Bank Transfer") [YOUR_SUPPORT_HANDLE] β Your Telegram support username Step 2: Configure Credentials Telegram Bot - Add your bot token to Telegram credentials Google Sheets - Connect your Google account and create/select your orders spreadsheet Google Gemini - Add your Gemini API key Sheet ID - Replace [YOUR_GOOGLE_SHEET_ID] with your actual Google Sheet ID Step 3: Customize Menu Update the menu section in the AI Agent system message with your actual: Food categories Item names and prices Delivery fees Any special offerings or combos Step 4: Test & Deploy Import the template into your n8n instance Test the conversation flow with a test Telegram account Verify Google Sheets logging works correctly Test the payment approval workflow Activate the workflow π° Currency Examples USD Version π MENU & PRICES (USD) Burgers Classic Burger β $12.99 Cheese Burger β $14.99 Deluxe Burger β $18.99 Delivery Fee β $3.99 EUR Version π MENU & PRICES (EUR) Burgers Classic Burger β β¬11.50 Cheese Burger β β¬13.50 Deluxe Burger β β¬17.50 Delivery Fee β β¬3.50 π Google Sheets Structure The template automatically logs orders with these columns: Order ID Customer Name Chat ID Phone Number Delivery Address Order Info Total Price Payment Status Order Status Timestamp π§ Customization Options Easy Customizations Menu Items - Add/remove/modify any food items Pricing - Update to your local pricing structure Currency - Change to any currency worldwide Business Hours - Modify operating hours Delivery Areas - Add location restrictions Payment Methods - Update payment information# Header 1
by Joseph
Here is your refined template description with detailed step-by-step instructions, markdown formatting, and customization guidance. YouTube Transcript Extraction Workflow This n8n workflow extracts and processes transcripts from YouTube videos using the YouTube Transcript API on RapidAPI. It allows users to retrieve subtitles from YouTube videos, clean them up, and return structured transcript data for further processing. Table of Contents Problem Statement & Target Audience Pre-conditions & API Requirements Step-by-Step Workflow Explanation Customization Guide How to Set Up This Workflow Problem Statement & Target Audience Who is this for? This workflow is ideal for content creators, researchers, and developers who need to: Extract subtitles from YouTube videos automatically. Format and clean** transcript data for readability. Use transcripts for summarization, content repurposing, or language analysis. Pre-conditions & API Requirements API Required YouTube Transcript API** (RapidAPI) n8n Setup Prerequisites A running n8n instance (Installation Guide) A RapidAPI account to access the YouTube Transcript API An API key from RapidAPI to authenticate requests Step-by-Step Workflow Explanation 1. Input YouTube Video URL (Trigger) This step provides a simple input form where users enter a YouTube video URL. 2. HTTP Request Node (Retrieve Transcript Data) Makes a POST request to the YouTube Transcript API via RapidAPI. Passes the video URL received from the input form. Uses an environment variable to store the API key securely. 3. Function Node (Process Transcript) Receives* the API response containing the *raw transcript**. Processes and cleans** the transcript: Removes unwanted characters. Formats text for readability. Handles errors** when no transcript is available. Outputs* both the *raw and cleaned transcript** for further use. 4. Set Field Node (Response Formatting) Structures** the processed transcript data into a user-friendly format. Returns** the final transcript data to the client. Customization Guide 1. Modify Transcript Cleaning Rules Update the Function Node to apply custom text processing, such as: Removing timestamps. Changing the output format (e.g., JSON, plain text). 2. Store Transcripts in a Database Add a Database Node (e.g., MySQL, PostgreSQL, or Firebase) to save transcripts. 3. Generate Summaries from Transcripts Integrate AI services (e.g., OpenAI, Google Gemini) to summarize transcripts. 4. Convert Transcripts into Speech Use ElevenLabs API to generate an AI-powered voiceover from transcripts. How to Set Up This Workflow Step 1: Import the Workflow into n8n Download or copy the workflow JSON file. Import it into your n8n instance. Step 2: Set Up the API Key Sign up for the YouTube Transcript API. Subscribe to the api. Copy and paste your api key where the "your_api_key" is. Step 3: Activate the Workflow Start the workflow in n8n. Enter a YouTube video URL in the input form. The workflow will return a cleaned transcript. This workflow ensures seamless YouTube transcript extraction and processing with minimal manual effort. π
by Immanuel
Automated Research Report Generation with OpenAI, Wikipedia, Google Search, Gmail/Telegram and PDF Output Description What Problem Does This Solve? π οΈ This workflow automates the process of generating professional research reports for researchers, students, and professionals. It eliminates manual research and report formatting by aggregating data, generating content with AI, and delivering the report as a PDF via Gmail or Telegram. Target audience: Researchers, students, educators, and professionals needing quick, formatted research reports. What Does It Do? π Aggregates research data from Wikipedia, Google Search, and SerpApi. Refines user queries and generates structured content using OpenAI. Converts the content into a professional HTML report, then to PDF. Sends the PDF report via Gmail or Telegram. Key Features π Real-time data aggregation from multiple sources. AI-driven content generation with OpenAI. Automated HTML-to-PDF conversion for professional reports. Flexible delivery via Gmail or Telegram. Error handling for robust execution. Setup Instructions Prerequisites βοΈ n8n Instance**: Self-hosted or cloud n8n instance. API Credentials**: OpenAI API: API key with GPT model access, stored in n8n credentials. SerpApi (Google Search): API key from SerpApi, stored in n8n credentials (do not hardcode in nodes). Gmail API: Credentials from Google Cloud Console with Gmail scope. Telegram API: Bot token from BotFather on Telegram. Installation Steps π¦ Import the Workflow: Copy the workflow JSON from the "Template Code" section below. Import it into n8n via "Import from File" or "Import from URL". Configure Credentials: Add API credentials in n8nβs Credentials section for OpenAI, SerpApi, Gmail, and Telegram. Assign credentials to respective nodes. For example: In the SerpApi Google Search node, use n8n credentials for SerpApi: api_key={{ $credentials.SerpApiKey }}. In the Send Research PDF on Gmail node, use Gmail credentials. In the Send PDF to Telegram node, use Telegram bot credentials. Set Up Nodes: OpenAI Nodes (Research AI Agent, OpenAI Chat Model, OpenAI Chat Middle Memory): Update the model (e.g., gpt-4o) and prompt as needed. Input Validation (Input Validation node): Ensure your input query format matches the expected structure (e.g., topic: "AI ethics"). Delivery Options (Send Research PDF on Gmail, Send PDF to Telegram): Configure recipient email or Telegram chat ID. Test the Workflow: Run the workflow by clicking the "Test Workflow" node. Verify that the research report PDF is generated and sent via Gmail or Telegram. How It Works High-Level Steps π Query Refinement**: Refines the input query for better research. Aggregate Data**: Fetches data from Wikipedia, Google Search, and SerpApi. Generate Report**: Uses OpenAI to create a structured report. Convert to PDF**: Converts the report to HTML, then PDF. Deliver Report**: Sends the PDF via Gmail or Telegram. Detailed descriptions are available in the sticky notes within the workflow screenshot above. Node Names and Actions Research and Report Generation Test Workflow: Triggers the workflow for testing. Input Validation: Validates the input query. Query Refiner: Refines the query for better results. Research AI Agent: Coordinates research using OpenAI. OpenAI Chat Model: Generates content for the report. Structured Output Parser: Parses OpenAI output into structured data. OpenAI Chat Middle Memory: Retains context during research. Wikipedia Google Search: Fetches data from Wikipedia. SerpApi Google Search: Fetches data via SerpApi. Merge Split Items: Merges data from multiple sources. Aggregate: Aggregates all research data. Generate PDF HTML: Creates an HTML report. Convert HTML to PDF: Converts HTML to PDF. Download PDF: Downloads the PDF file. Send PDF to Telegram: Sends the PDF via Telegram. Send Research PDF on Gmail: Sends the PDF via Gmail. Customization Tips Expand Data Sources** π‘: Add more sources (e.g., academic databases) by adding nodes to Merge Split Items. Change Report Style** βοΈ: Update the Generate PDF HTML node to modify the HTML template (e.g., adjust styling or sections). Alternative Delivery** π§: Add nodes to send the PDF via other platforms (e.g., Slack). Adjust AI Model** π§ : Modify the OpenAI Chat Model node to use a different model (e.g., gpt-3.5-turbo).
by Sk developer
π₯ Bulk TikTok Video Download Without Watermark to Google Drive This workflow automates the process of downloading TikTok videos and uploading them to Google Drive. It reads TikTok URLs from a Google Sheet, downloads the video using the TikTok Video Downloader β a tool for downloading TikTok videos without watermark in HD quality β uploads it to Drive, makes it public, and updates the same sheet with the Drive link. π§ What It Does β Manually triggered when ready to run. π Reads TikTok URLs from a Google Sheet. π Loops through each URL one at a time. π Fetches video download links using the TikTok Video Downloader β a reliable TikTok video downloader without watermark. β¬οΈ Downloads each video in high-definition (HD) format using the direct media link. βοΈ Uploads the video to Google Drive. π Sets public sharing permission for the video. βοΈ Updates the original Google Sheet with the public Drive URL. π Google Sheet Example Make sure your sheet has at least these columns: | url | drive_link (to be auto-filled) | |-------------------------------------|--------------------------------| | https://www.tiktok.com/@user1... | (blank initially) | | https://www.tiktok.com/@user2... | (blank initially) | > The workflow reads from url and fills in drive_link after upload. π§© Nodes Used | Node Name | Type | Purpose | |------------------------------|-------------------|-------------------------------------------------------| | When clicking βExecuteβ | Manual Trigger | Starts the workflow manually | | Get Data From Google Sheets | Google Sheets | Fetches rows (TikTok URLs) | | Loop Over Items | Split In Batches | Iterates over each row | | Call TikTok Downloader | HTTP Request | Gets video download link from TikTok Video Downloader | | Wait | Wait | Optional delay to prevent overload | | Download File | HTTP Request | Downloads HD video using media link | | Upload File In Google Drive | Google Drive | Uploads the video to Google Drive | | Set Public Permission | Google Drive | Makes the uploaded file publicly accessible | | Update Row In Google Sheet | Google Sheets | Adds Drive link to the same row | | Sleep | Wait | Small delay between each iteration | π Requirements β Google API credentials (Service Account) with access to: Google Sheets Google Drive π RapidAPI Key for TikTok Video Downloader β a TikTok video downloader without watermark (HD supported) π A Google Sheet with a url column containing TikTok video URLs π§© Challenges Solved | β Challenge | β Solution | |-------------|-------------| | TikTok video URLs often have watermarks and low quality | Used TikTok Video Downloader API for HD + no watermark download links | | No easy way to bulk download and organize TikToks | Automated fetching, downloading, and uploading using n8n + Google Drive | | Manual video saving and re-uploading to Drive is time-consuming | Eliminated all manual steps with a fully automated workflow | | Tracking which videos are already processed | Automatically updates the Google Sheet row with the final Drive link | | Drive files are private by default | Automatically sets public sharing permission on uploaded videos | | Risk of API rate limits or throttling | Added Wait nodes and batch processing to avoid overload | π Benefits | π Benefit | π¬ Description | |------------|----------------| | π Saves Time | Fully automates a previously manual workflow | | π₯ High Quality Content | Videos downloaded are HD + watermark-free β ready for reuse or archives | | π Reusable Setup | Can process unlimited TikTok URLs via the Google Sheet | | π Organized Output | Keeps track of source URL and uploaded Drive link in a single sheet | | π Secure but Shareable | Drive links are auto-shared publicly while remaining under your control | | π Scalable | Can be run daily, weekly, or triggered by new rows β completely scalable | | πΈ Cost-Effective | No need for paid tools or manual freelancers β runs on n8n + free APIs | π‘ Use Cases Content curation from TikTok Archiving user-submitted TikToks Automating social-to-cloud workflows Bulk migration of video content Saving TikTok videos in HD without watermark for sharing or archiving π Tips Replace manual trigger with Cron for full automation. Use the TikTok Video Downloader responsibly β check API limits. Store metadata (e.g., uploader, hashtags) in additional Google Sheet columns. This tool helps ensure you're always downloading high-quality TikTok videos without watermark.
by Mario
Purpose This workflow snippet allows for advanced error catching during retry attempts. There are cases, where you want to check if an item exists first, so you can determine the following actions. Some APIβs do not support an endpoint (e.g. Todoist: completed tasks) to do so, which is why you would work with the error branch, only that this does not work well in combination with the retry functionality. How it works Instead of the builtin retry function of a Node a custom loop is used, to get more granular control in between the iterations If the main executed node fails, the error can be filtered for an expected error, which can trigger a separate action The retries only happen, if an unexpected error happened The workflow only stops, if the defined amount of retries exceeded Setup Copy the nodes into your existing workflow Replace the βReplace meβ placeholder with the Node you want to apply the retry logic on Follow the sticky notes for more instructions and optional settings
by Ventsislav Minev
UptimeRobot Alerts to Telegram with Visual Verification Automatically sends Telegram notifications with optional screenshots when monitors change status (β UP/π΄ DOWN/βΈοΈ PAUSED) Example Message in Telegram: Who Is This For? Teams or individuals needing to: Get alerts when websites/services go down Verify outages with visual screenshots Monitor infrastructure from Telegram What Does This Workflow Solve? π¨ Missed Alerts: Get immediate notifications in Telegram πΌοΈ Visual Verification: Optional screenshot confirmation of outages π Status Tracking: Clear records of when issues began/resolved π One-Click Access: Direct links to affected monitors β±οΈ Time Savings: No need to check dashboards manually Setup Guide 1. Pre-Requisites UptimeRobot Account**: With at least one monitor configured Gmail Account**: To receive alert notifications Telegram Account**: To receive alerts (mobile/desktop app recommended) (Optional) ScreenshotMachine free/paid account 2. Credentials Setup Make sure your n8n instance is connected with: Gmail Account** (via OAuth2) UptimeRobot API** (via API key) Telegram Bot** (via bot token) (Optional) ScreenshotMachine (via customer key) 3. Configure Your n8n Workflow Nodes 1. Alert Trigger Gmail Trigger**: Configure to watch for emails from alert@uptimerobot.com Set appropriate polling interval (e.g., every 5 minutes) 2. Monitor Configuration Conf Node**: Set your preferences: { "take_screenshot": true, "screenshotmachine_secret": "your-secret-here", "screenshotmachine_device": "desktop", "screenshotmachine_dimension": "1366xfull", "screenshotmachine_format": "png" } 3. Notification Settings Telegram Nodes**: Set your Chat ID (find with @getidsbot) Customize message formatting if needed 4. Service-Specific Setup UptimeRobot: Go to Dashboard β My Settings β API Settings Create API key with "Monitor Read" permissions Enable email alerts in monitor settings Telegram Bot: Message @BotFather to create new bot Get your Chat ID using @getidsbot Add bot token to n8n credentials ScreenshotMachine (Optional): Sign up at screenshotmachine.com Get Customer Key from account dashboard Set your secret phrase if using hash verification Final Steps Test your workflow by manually triggering a monitor status change Verify Telegram notifications arrive as expected Check screenshot quality if enabled Monitor for a few days to fine-tune alert preferences Happy Monitoring!
by Risper
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. How It Works This n8n workflow automatically discovers high-quality business leads from Reddit posts by analysing posts across targeted subreddits. Loads your business profile from a connected Google Sheet. Uses AI to identify relevant subreddits where your potential customers engage. Generates intent-based Reddit search queries based on your services, keywords, and client pain points. Searches Reddit in real time using the generated queries. Classifies posts based on whether they show lead potential. Analyses high-potential posts for service-fit, urgency, and estimated value. Filters and scores leads to prioritize high-conversion opportunities. Saves the most promising leads to a dedicated Google Sheet. Sends Slack alerts to notify your sales team for immediate follow-up. Requirements Before using this workflow, ensure the following services are connected and configured: Google Sheets (OAuth2): Reads your business profile and writes qualified leads Reddit (OAuth2) Perform Reddit post searches based on generated queries Google Gemini API Analyse posts, generate queries, and extract insights Slack API : Notify your team with qualified lead summaries Google Sheets Setup You will need two Google Sheets: Business Profile Sheet (Input) This sheet contains a single row describing your service business. The workflow reads this to generate relevant subreddit selections and search queries. Required Fields (as headers in row 1): profession industry primary_services service_keywords target_client_profile pain_points intent_signals urgency_indicators price_range Reddit Leads Sheet (Output) This sheet stores high-quality Reddit posts identified as potential leads. The workflow appends or updates rows based on post_id to avoid duplication. Expected Columns: post_id post_url post_title post_post post_subreddit post_date
by Joseph
(Image Generation β Hosting β Video Generation) This workflow is designed for creators, automation enthusiasts, and indie hackers who want to generate image-based videos automatically using AI tools β at a low cost. βοΈ Workflow Overview This automation performs the following steps: Trigger (Schedule or manual) Generate an image using Flux (choose between two APIs) Upload the image to Kraken.io to get a public URL Send the image to Runway ML (choose between two APIs) to generate a video Receive the video as a URL β ready for posting, download, or further automation π οΈ Step-by-Step Setup πΌοΈ Flux (Image Generation) You can use either of the following providers: Option 1: Flux by BlackForest Labs (Direct API) π Get your API key here: https://docs.bfl.ml/ Paste your API key in the HTTP Request node named Flux (Blackforest) You can customize prompts or styles inside the JSON body Option 2: Flux via RapidAPI π Subscribe and get your key here: https://rapidapi.com/poorav925/api/ai-text-to-image-generator-flux-free-api/playground/apiendpoint\_e38039ee-1912-4ef9-b4d4-270d72fca851 Enter your RapidAPI key in the X-RapidAPI-Key header Optional: tweak prompts, style, or resolution inside the JSON body π Kraken.io (Hosting the Image Publicly) Runway ML requires the image to be publicly accessible. We use Kraken.io to host the generated image and return a public URL. π Get your API credentials: https://kraken.io/account/api-credentials Setup: Copy your API Key and API Secret Open the Kraken Upload node in n8n Replace placeholders with your credentials The node uploads your image and gives back a public image URL for Runway to use π¬ RunwayML (Video Generation) You also have two options here: Option 1: Runway Official API π Get your credentials at: https://dev.runwayml.com/ Use the public image URL from Kraken in the JSON body Paste your Bearer token in the Authorization header Customize other settings like video length, style, FPS, etc. Option 2: Runway via RapidAPI π Subscribe and get your key here: https://rapidapi.com/fortunehoppers/api/runwayml/playground/apiendpoint\_93c8554d-8097-40cd-8252-3d4dec9c0e68 Paste your RapidAPI key in the request header Customize prompt and generation options in the body Use the Kraken-generated image URL as the input source π€ What to Do with the Video Once the video is generated, youβll get a direct video URL. You can: Save it to Google Sheets or Notion Send it via email Trigger a YouTube upload automation Or download manually for editing and reposting π‘ Optional Tips & Notes You can schedule this workflow to generate AI videos daily or weekly Combine it with a Google Sheet of prompts for bulk automation Try using a consistent visual style or theme for better branding This workflow is lightweight and affordable β perfect for indie projects or experimental content generation Great for shorts, quote visuals, music loops, AI art promos, etc. π Resources Flux (Blackforest) Docs Flux on RapidAPI RunwayML Official Docs Runway on RapidAPI Kraken.io API Dashboard π Need Help? Feel free to reach out: π¦ Twitter: @juppfy π§ Email: joseph@uppfy.com If youβd like to hire me for custom n8n workflows or product automations, donβt hesitate to get in touch.
by Ferenc Erb
Overview An automation workflow that creates a complete REST API for digitally signing PDF documents using n8n webhooks. This service demonstrates how to implement secure document signing functionality through standardized API endpoints with file upload and download capabilities. Use Case This workflow is designed for developers and automation specialists who need to implement digital document signing. It's particularly useful for: Integrating PDF signing capabilities into existing document workflows API-based automation of signature processes Creating proof-of-concept implementations for document verification systems Learning n8n's webhook capabilities and file handling techniques Testing PDF signing in development environments before production implementation What This Workflow Does API-Based Document Management Exposes RESTful webhook endpoints for all document operations Handles multipart/form-data uploads for PDF documents Processes JSON payloads for signing configuration Provides download functionality for completed documents Digital Certificate Handling Uploads existing PFX/PKCS#12 digital certificates Generates new certificates with customizable attributes Securely manages certificate storage and access Associates certificates with signing operations Cryptographic PDF Signing Applies digital signatures using industry-standard cryptographic methods Embeds signature information within PDF document structure Validates document integrity through cryptographic verification Preserves original document while adding signature elements Webhook Integration System Routes different API methods to appropriate handlers Validates request payloads and file content Manages authentication through webhook paths Returns structured responses for integration with other systems Technical Architecture Components API Gateway: n8n webhook nodes that receive external requests Request Router: Switch nodes that direct operations based on method parameters Document Processor: Function nodes for PDF manipulation and verification Certificate Manager: Specialized nodes for cryptographic key operations Storage Interface: File operation nodes for document persistence Response Formatter: Nodes that structure API responses Integration Flow Client Request β Webhook Endpoint β Method Router β Processing Engine β Digital Signing β Storage β Response Generation β Client Response Setup Instructions Prerequisites n8n installation (minimum version 0.214.0) Node.js 14 or higher Required environment variable: NODE_FUNCTION_ALLOW_EXTERNAL: "node-forge,@signpdf/signpdf,@signpdf/signer-p12,@signpdf/placeholder-plain" Configuration Steps Import Workflow Import the workflow JSON into your n8n instance Activate the workflow to enable the webhooks Configure Storage Set the storage path variables in the workflow Ensure proper permissions on the storage directories Test API Endpoints Use the included test scripts to verify functionality Test PDF upload, certificate generation, and signing Integration Document the webhook URLs for integration with other systems Configure error handling according to your requirements Testing Methods Test the workflow functionality using various HTTP requests and JSON data: Upload PDF documents to the document processing endpoint Upload or generate digital certificates Execute PDF signing operations Download signed documents from the download endpoint Webhook Endpoints The workflow exposes two primary webhook endpoints that form a complete API for PDF digital signing operations: 1. Document Processing Endpoint (/webhook/docu-digi-sign) This endpoint handles all document and certificate operations: Method: Upload PDF HTTP: POST Content-Type: multipart/form-data Parameters: method, uploadType, fileName, fileData Method: Upload Certificate HTTP: POST Content-Type: multipart/form-data Parameters: method, uploadType, fileName, fileData Method: Generate Certificate HTTP: POST Content-Type: application/json Parameters: method, subjectCN, issuerCN, serialNumber, validFrom, validTo, password Method: Sign PDF HTTP: POST Content-Type: application/json Parameters: method, inputPdf, pfxFile, pfxPassword 2. Document Download Endpoint (/webhook/docu-download) This endpoint handles the retrieval of processed documents: Method: Download Signed PDF HTTP: GET Content-Type: application/json Parameters: method, fileType, fileName Key Workflow Sections The workflow is organized into logical sections with clear responsibilities: Request Processing**: Parses incoming webhook data Method Routing**: Directs requests to appropriate handlers Document Management**: Handles file operations and storage Cryptographic Operations**: Manages signing and certificate functions Response Formatting**: Structures and returns results