by Yaron Been
LinkedIn AI Agent: Auto-Post Creator & Multi-Group Distributor Transform simple topic ideas into engaging LinkedIn posts and automatically distribute them across your profile and multiple LinkedIn groups. This powerful n8n workflow combines AI content generation with intelligent distribution, helping you maintain a consistent professional presence while maximizing your reach across relevant communities. 🚀 How It Works This sophisticated 6-step automation turns content ideas into LinkedIn success: Step 1: Smart Content Monitoring The workflow continuously monitors your Google Sheets for new post topics marked as "Pending", checking every minute for fresh content to process. Step 2: AI-Powered Content Generation GPT-4 transforms your basic topic into a professionally crafted LinkedIn post featuring: Compelling opening hooks that grab attention 3-4 informative paragraphs with valuable insights Strategic questions to encourage engagement 4-6 relevant hashtags for discoverability Professional emoji placement for visual appeal Optimized formatting for LinkedIn's platform Step 3: Professional Formatting The workflow ensures your content meets LinkedIn's technical requirements with proper JSON formatting, character limits, and special character handling. Step 4: Personal Profile Publishing Your generated post is automatically published to your personal LinkedIn profile, maintaining your professional brand presence. Step 5: Multi-Group Distribution The same content is intelligently distributed across all your specified LinkedIn groups, amplifying your reach to targeted professional communities. Step 6: Status Management The workflow automatically updates your Google Sheets to mark posts as "Posted", providing clear tracking of your content pipeline. ⚙️ Setup Steps Prerequisites Active LinkedIn account with API access Google Sheets access for content management OpenAI API key with GPT-4 access LinkedIn group memberships with posting permissions n8n instance (cloud or self-hosted) Required Google Sheets Structure Sheet 1 - Main Content: | ID | LinkedIn Post Title | Status | |----|-------------------|--------| | 1 | AI Trends in 2024 | Pending | | 2 | Remote Work Tips | Posted | Sheet 2 - Groups: | GroupIds | |-------------| | 123456789 | | 987654321 | | 456789123 | Note: Collect LinkedIn group IDs from groups where you have posting permissions. These can be found in the group URL or through LinkedIn's API. Configuration Steps Credential Setup Google Sheets OAuth2: Access your content spreadsheet OpenAI API Key: Required for AI content generation LinkedIn OAuth2: Enable profile and group posting HTTP Authentication: Configure LinkedIn API headers Google Sheets Preparation Create spreadsheet with the required two-sheet structure Populate group IDs from your joined LinkedIn groups Add initial post topics with "Pending" status Ensure proper column naming and data types LinkedIn Group Setup Join relevant professional LinkedIn groups Verify posting permissions in each group Collect group IDs using LinkedIn's interface or API Test posting permissions before full automation AI Content Customization The default prompt generates professional LinkedIn content, but can be customized for: Industry-specific terminology and trends Company voice and brand guidelines Target audience preferences Content style (educational, promotional, thought leadership) Workflow Activation Import the workflow JSON into your n8n instance Configure all credential connections Test with sample content before going live Activate the Google Sheets trigger 🎯 Use Cases Content Creators & Influencers Consistent Posting: Maintain regular LinkedIn presence without daily manual work Audience Growth: Reach multiple professional communities simultaneously Content Scaling: Transform brief ideas into full-length engaging posts Brand Building: Establish thought leadership across industry groups Marketing Teams Lead Generation: Share valuable content across targeted professional groups Brand Awareness: Increase visibility in relevant industry communities Thought Leadership: Position company experts as industry authorities Content Distribution: Maximize reach of marketing messages and insights Sales Professionals Pipeline Building: Share insights that attract potential clients Network Expansion: Engage with prospects across multiple professional groups Authority Building: Establish credibility through valuable content sharing Relationship Nurturing: Maintain visibility with existing connections Consultants & Freelancers Client Acquisition: Demonstrate expertise to potential clients Professional Branding: Build reputation across industry-specific groups Service Promotion: Share case studies and success stories broadly Network Building: Connect with peers and potential collaborators Business Leaders & Executives Industry Influence: Share strategic insights across professional networks Talent Attraction: Showcase company culture and opportunities Partnership Development: Connect with potential business partners Market Education: Share expertise to influence industry conversations 🔧 Advanced Customization Options Content Strategy Enhancement Multi-Tone Generation: Create different content styles for various audiences Industry Templates: Pre-built prompts for specific professional sectors Engagement Optimization: A/B testing different post formats and styles Content Calendar Integration: Schedule posts for optimal timing Distribution Intelligence Group Performance Tracking: Monitor engagement across different groups Selective Distribution: Choose specific groups based on content type Audience Segmentation: Tailor content for different professional communities Engagement Analysis: Track which groups provide best ROI Content Quality Control Approval Workflows: Add human review before automatic posting Content Scoring: Rate post quality before distribution Brand Compliance: Ensure posts meet company guidelines Duplicate Detection: Avoid posting similar content too frequently Extended Integration Options CRM Integration: Track leads generated from LinkedIn posts Analytics Platforms: Monitor engagement and conversion metrics Content Management: Connect to existing content planning tools Social Media Expansion: Extend to other professional platforms 📊 Content Generation Features AI Writing Capabilities The workflow generates LinkedIn posts that include: Professional Structure: Attention-grabbing opening statements Well-organized multi-paragraph content Clear value propositions and insights Strategic calls-to-action for engagement LinkedIn Optimization: Platform-specific formatting and spacing Proper hashtag research and placement Emoji usage that enhances readability Character count optimization for maximum impact Engagement Drivers: Thought-provoking questions to encourage comments Industry insights that spark discussions Personal anecdotes that build connections Actionable tips that provide immediate value Sample Output Input Topic: "Remote Work Productivity Tips" Generated Post: 🏠 Working from home has taught me that productivity isn't about being busy—it's about being intentional. After managing remote teams for 3 years, I've discovered that the most successful remote workers share three key habits that transform their daily performance. First, they create physical boundaries even in small spaces. A dedicated workspace signals to your brain that it's time to focus, even if it's just a corner of your kitchen table. Second, they batch similar tasks together. Instead of jumping between emails, calls, and deep work, they group activities to maintain mental flow and reduce cognitive switching costs. Third, they communicate proactively rather than reactively. They share progress updates before being asked and clarify expectations upfront to avoid confusion later. What's your most effective remote work strategy? I'd love to hear what's working for your team! 💪 #RemoteWork #Productivity #WorkFromHome #Leadership #TeamManagement #ProfessionalDevelopment 🛠️ Troubleshooting & Best Practices Common Issues & Solutions LinkedIn API Limitations Respect posting frequency limits to avoid account restrictions Monitor API usage and implement appropriate delays between posts Ensure compliance with LinkedIn's terms of service Maintain authentic engagement rather than purely automated interactions Group Posting Permissions Verify membership status and posting rights before adding group IDs Some groups require administrator approval for posts Monitor group rules and community guidelines Remove inactive or restricted groups from your list Content Quality Control Review AI-generated content periodically for brand consistency Adjust prompts based on engagement performance Maintain a balance between automation and personal touch Monitor comments and engage authentically with your audience Optimization Strategies Performance Enhancement Track engagement metrics across different groups A/B test posting times and content formats Refine prompts based on successful post patterns Gradually expand to new groups based on performance Content Strategy Develop content themes that resonate with your target audience Create series of related posts for deeper engagement Balance promotional content with value-driven insights Maintain consistency in voice and messaging Network Growth Engage with comments on your automated posts Connect with active commenters to expand your network Participate in group discussions beyond your own posts Build genuine relationships through authentic interactions 📈 Success Metrics Engagement Indicators Post Reach: Total views across profile and groups Interaction Rate: Comments, likes, and shares per post Network Growth: New connections from content engagement Group Performance: Which communities provide best engagement Business Impact Measurements Lead Generation: Connections and inquiries from LinkedIn posts Brand Awareness: Mentions and sharing of your content Thought Leadership: Recognition as industry expert Professional Opportunities: Speaking, collaboration, or job opportunities 📞 Questions & Support Need help setting up or optimizing your LinkedIn AI Agent workflow? 📧 Direct Technical Support Email: Yaron@nofluff.online Response Time: Within 24 hours on business days Expertise: LinkedIn API integration, AI prompt optimization, workflow scaling 🎥 Comprehensive Learning Resources YouTube Channel: https://www.youtube.com/@YaronBeen/videos Complete setup walkthrough and configuration Advanced customization techniques and strategies LinkedIn API best practices and limitations Content strategy optimization for maximum engagement Troubleshooting common integration issues 🤝 Professional Networking & Updates LinkedIn: https://www.linkedin.com/in/yaronbeen/ Connect for ongoing automation support and advice Share your LinkedIn growth success stories Get early access to new workflow templates and features Join discussions about LinkedIn marketing automation 💬 Support Request Guidelines Include in your support message: Your current LinkedIn strategy and goals Target audience and industry focus Specific LinkedIn groups you want to target Any technical errors or integration issues Current content creation process and pain points
by Batu Öztürk
Extract the main idea and key takeaways from YouTube videos and turn them into Airtable content ideas 📝 Description Automatically turn YouTube videos into clear, structured content ideas stored in Airtable. This workflow pulls new video links from Airtable, extracts transcripts using a RapidAPI service, summarizes them with your favourite LLM, and logs the main idea and key takeaways—keeping your content pipeline fresh with minimal effort. ⚙️ What It Does Scans Airtable for new YouTube video links every 5 minutes. Extracts the transcript of the video using a third-party API via RapidAPI. Summarizes the content to generate a main idea and takeaways. Updates the original Airtable entry with the insights and marks it as completed. 🛠 Prerequisites Before using this template, make sure you have: ✅ A RapidAPI account with access to the youtube-video-summarizer-gpt-ai API. ✅ A valid RapidAPI key. ✅ An OpenAI, Claude or Gemini account connected to n8n. ✅ An Airtable account with a base and table ready. 🧰 Setup Instructions Clone this template into your n8n workspace. Open the Get YouTube Sources node and configure your Airtable credentials. In the Get video transcript node: Enter your X-RapidAPI-Key under headers. The API endpoint is pre-configured. Connect your LLM credentials to the Extract detailed summary node. (Optional) Adjust the summarization prompt in the LangChain node to better suit your tone. Set your preferred schedule in the Trigger node. 📋 Airtable Setup Create a base (e.g., Content Hub) with a table named Ideas and the following columns: | Column Name | Type | Required | Notes | |-------------|------------|----------|----------------------------| | Type | Single select | ✅ | Must be set to Youtube Video | | Source | URL | ✅ | The YouTube video URL | | Status | Checkbox | ✅ | Leave empty initially; updated after processing | | MainIdea | Single line text | ✅ | Summary generated by OpenAI | | Key Takeaways | Long text | ✅ | List of takeaways extracted from the transcript Activate the workflow—and you're done!
by Angel Menendez
Who is this for? This workflow is for professionals and teams who want to automate LinkedIn message replies with intelligent, human-like responses — without losing control over tone or accuracy. Ideal for founders, sales teams, DevRel, or community managers handling high-volume inbound messages. What problem is this workflow solving? Responding to every LinkedIn message manually is slow and inconsistent. Basic AI bots generate replies without context or nuance. This subworkflow solves both problems by using structured message routing from Notion and profile insights from UniPile to craft smart, context-aware responses. What this workflow does This workflow takes the sender’s message and profile (from LinkedIn Auto Message Router with Request Detection) and references your centralized Notion database of message types. It uses that to either match the message to a known response or generate a new one using OpenAI's GPT model — all while following professional tone guidelines. This is the third workflow in a 3-part automation system: Receives data from LinkedIn Auto Message Router with Request Detection Uses UniPile LinkedIn Profile Lookup Subworkflow to enrich responses based on follower count or org data Example Use Case If a message comes from someone with low reach (e.g., under 1,000 followers), the AI politely deflects a meeting request. If an influencer reaches out, the AI immediately offers a booking link. Your team controls this logic by updating the Notion database — no edits to the workflow required. Setup Connect this workflow as a subworkflow in your router or Slack approval flow Store your Notion API key and database ID in n8n Provide the following parent inputs: message – The LinkedIn message text sender – Name of the sender chatid – Session ID (optional for memory) linkedinprofile – Enriched array with LinkedIn context (follower count, connection info, etc.) Add your preferred AI model credentials (supports OpenAI, Gemini, or Ollama) Optional: Customize system prompt to better match your brand voice How to customize this workflow to your needs Update the Notion schema to include industry-specific categories or actions Change the AI tone (e.g., humorous, more corporate, etc.) Add conditional logic for auto-sending messages without Slack approval Extend to support multiple platforms (e.g., email, X/Twitter, Instagram DMs)
by Zain Ali
🧠 Email real time RAG Assistant with Gmail, OpenAI & PGVector 📌 Who’s it for This workflow is ideal for: Professionals Project managers Sales and support teams Anyone managing high volumes of Gmail messages It enables fast and intelligent search through your email inbox using natural language queries. ⚙️ How it works / What it does Continuously monitors your Gmail inbox for new emails. Extracts email content and metadata (subject, body, sender, date). Converts email content into vector embeddings using OpenAI. Stores embeddings in a PostgreSQL database with PGVector. A conversational AI agent performs semantic search on your stored email history. Supports time-sensitive and context-aware responses via OpenAI Chat model. 🚀 How to set up Connect your Gmail account to the Gmail Trigger node (with API access enabled). Configure OpenAI credentials for the Embedding and Chat nodes. Set up a PostgreSQL database with the PGVector extension enabled. Import the workflow into your n8n instance (Cloud or Self-hosted). Customize parameters like polling frequency, embedding settings, or vector query depth. 📋 Requirements ✅ n8n instance (Self-hosted or Cloud) ✅ Gmail account with API access ✅ OpenAI API Key ✅ PostgreSQL database with PGVector extension installed 🛠️ How to customize the workflow Email Filtering**: Change filters in the Gmail Trigger to watch specific labels or senders. Text Splitting Granularity**: Adjust chunkSize and chunkOverlap in the text splitter node. Query Depth**: Modify topK in the vector search node to retrieve more or fewer similar results. Prompt Tuning**: Customize the system message or agent instructions in the RAG node. Workflow Extensions**: Add notifications, error logging, Slack/Telegram alerts, or data exports.
by Nurseflow
💼 LinkedIn Content Machine – AI-Powered Post Generator & Scheduler for X and LinkedIn How it works: This end-to-end workflow automates your personal or brand content strategy by: 🧠 Using Google Gemini or OpenAI to generate engaging LinkedIn/X content from a title or trending posts. 🗓️ Posting directly to LinkedIn and X (formerly Twitter). 📊 Pulling high-performing LinkedIn posts to inspire new ideas. ✍️ Saving AI-generated drafts to Google Sheets for review. 🔔 Notifying your team on Slack when drafts are ready. Steps to set up: Add your API keys for Google Gemini or OpenAI. Set up your LinkedIn, X (Twitter), Google Sheets, and Slack credentials. Customize prompt logic or post filters if needed. Schedule the idea generation module or trigger it manually. Start generating and posting consistent, high-quality content with zero manual effort!
by Louis Chan
How it works Transform medical documents into structured data using Google Gemini AI with enterprise-grade accuracy. Classifies document types (receipts, prescriptions, lab reports, clinical notes) Extracts text with 95%+ accuracy using advanced OCR Structures data according to medical taxonomy standards Supports multiple languages (English, Chinese, auto-detect) Tracks processing costs and quality metrics automatically Set up steps Prerequisites Google Gemini API key (get from Google AI Studio) Quick setup Import this workflow template Configure Google Gemini API credentials in n8n Test with a sample medical document URL Deploy your webhook endpoint Usage Send POST request to your webhook: { "image_url": "https://example.com/medical-receipt.jpg", "expected_type": "financial", "language_hint": "auto" } Get structured response: json{ "success": true, "result": { "documentType": "financial", "metadata": { "providerName": "Dr. Smith Clinic", "createdDate": "2025-01-06", "currency": "USD" }, "content": { "amount": 150.00, "services": [...] }, "quality_metrics": { "overall_confidence": 0.95 } } } Use cases Healthcare Organizations Medical billing automation - Process receipts and invoices automatically Insurance claim processing - Extract data from claim documents Clinical documentation - Digitize patient records and notes Data standardization - Consistent structured output format System Integrators EMR integration - Connect with existing healthcare systems Workflow automation - Reduce manual data entry by 90% Multi-language support - Handle international medical documents Quality assurance - Built-in confidence scoring and validation Supported Document Types Financial: Medical receipts, bills, insurance claims, invoices Clinical: Medical charts, progress notes, consultation reports Prescription: Prescriptions, medication lists, pharmacy records Administrative: Referrals, authorizations, patient registration Diagnostic: Lab reports, test results, screening reports Legal: Medical certificates, documentation forms
by simonscrapes
Use Case Transform and optimize images for web use: You need to host local images online You want to reduce image file sizes automatically You need image URLs for web content You want to generate and optimize AI-created images What this Workflow Does The workflow processes images through two services: Uploads images to ImgBB for hosting and URL generation (free but need API key) Optimizes images using ReSmush.it to reduce file size (free) Optional: Creates images using OpenAI's image generation Returns optimized image URLs ready for use Setup Create an ImgBB account and get your API key Add your ImgBB API key to the HTTP Request node (key parameter) Optional: Configure OpenAI credentials for image generation Connect your image input source How to Adjust it to Your Needs Skip OpenAI nodes if using your own image files Adjust image optimization parameters Customize image hosting settings Modify output format for your needs More templates and n8n workflows >>> @simonscrapes
by Nick Saraev
This workflow creates an end-to-end Instagram content pipeline that automatically discovers trending content from competitor channels, extracts valuable insights, and generates new high-quality scripts for your own content creation. The system helped scale an Instagram channel from 0 to 10,000 followers in just 15 days through intelligent content repurposing. Benefits: Complete Content Automation - Monitors competitor Instagram accounts, downloads new reels, and processes them without manual intervention AI-Powered Script Generation - Uses ChatGPT and Perplexity to analyze content, identify tools/technologies, and rewrite scripts with fresh angles Smart Duplicate Prevention - Automatically tracks processed content in a database to avoid redundant work Multi-Platform Intelligence - Combines Instagram scraping, AI transcription, web research, and content generation in one seamless flow Scalable Content Strategy - Process content from multiple niches and creators to fuel unlimited content ideas Revenue-Focused Approach - Specifically designed to identify monetizable tools and technologies for business-focused content How It Works: Instagram Content Discovery: Uses Apify's Instagram scraper to monitor specified creator accounts for new reels Automatically downloads video content and metadata from target accounts Filters content based on engagement metrics and relevance Intelligent Processing Pipeline: Transcribes video content using OpenAI Whisper for accurate text extraction Filters content using AI to identify tools, technologies, and automation opportunities Cross-references against existing database to prevent duplicate processing Enhanced Research & Analysis: Searches Perplexity AI for additional insights about discovered tools Generates step-by-step usage guides and implementation instructions Identifies unique angles and opportunities for content improvement Script Generation & Optimization: Creates new, original scripts optimized for your specific audience Maintains consistent brand voice while adding fresh perspectives Includes strategic call-to-action elements for audience engagement Required Google Sheets Database Setup: Before running this workflow, create a Google Sheets database with these exact column headers: Essential Columns: id - Unique Instagram post identifier (primary key for duplicate detection) timestamp - When the reel was posted caption - Original reel caption text hashtags - Hashtags used in the post videoUrl - Direct link to download the video file username - Account that posted the reel scrapedTranscript - Original transcript from video (added by workflow) newTranscript - AI-generated script for your content (added by workflow) Additional Tracking Columns: shortCode - Instagram's internal post code url - Public Instagram post URL commentsCount - Number of comments firstComment - Top comment on the post likesCount - Number of likes videoViewCount - View count metrics videoDuration - Length of video in seconds Setup Instructions: Create a new Google Sheet with these column headers in the first row Name the sheet "Reels" Connect your Google Sheets OAuth credentials in n8n Update the document ID in the workflow nodes The merge logic relies on the id column to prevent duplicate processing, so this structure is essential for the workflow to function correctly. Business Use Cases: Content Creators - Scale content production by 10x while maintaining quality and originality Marketing Agencies - Offer content research and ideation as a premium service Course Creators - Identify trending tools and technologies for educational content Revenue Potential: This exact system can be sold as a service for $3,000-$5,000 to growing channels or agencies. The automation saves 10+ hours weekly of manual research and content planning. Difficulty Level: Intermediate Estimated Build Time: 1-2 hours Monthly Operating Cost: ~$30 (API usage) Watch the Complete Build Process Want to see exactly how this system was built from scratch? Nick Saraev walks through the entire development process in this comprehensive tutorial, including all the debugging, dead ends, and problem-solving that goes into building real automation systems. 🎥 Watch: "The N8N Instagram Parasite System (10K Followers In 15 Days)" This 1.5-hour deep-dive shows the actual build process - not a polished demo, but real system development with all the thinking and iteration included. Set Up Steps: Configure Apify Integration: Sign up for Apify account and obtain API key Replace the bearer token in the "Run Actor Synchronously" node Customize the username array with your target Instagram accounts Set Up AI Services: Add OpenAI API credentials for transcription and content generation Configure Perplexity API for enhanced research capabilities Set up appropriate rate limiting for cost control Database Configuration: Create Google Sheets database with provided column structure Connect Google Sheets OAuth credentials Configure the merge logic for duplicate detection Content Filtering Setup: Customize the AI prompts for your specific niche and requirements Adjust the filtering criteria for tool/technology detection Set up the script generation template to match your brand voice Automation Schedule: Configure the schedule trigger for daily content monitoring Set optimal timing based on your content creation workflow Test the complete flow with a small number of accounts first Advanced Customization: Add additional content sources beyond Instagram Integrate with your existing content management systems Scale up monitoring to dozens of competitor accounts More AI Automation Systems:* For more advanced automation tutorials and business systems, check out My YouTube Channel where I share proven automation strategies that generate real revenue.
by Airtop
Automating Person Data Enrichment and CRM Update Use Case This automation enriches a person’s professional profile using their name and work email, scores them against an ICP (Ideal Customer Profile), and updates their record in HubSpot. It’s ideal for sales, marketing, and recruitment teams needing reliable contact insights. What This Automation Does This automation performs the following using the input parameters: Person name**: The full name of the individual. Work email**: The professional email address of the contact. Airtop Profile (connected to LinkedIn)**: An authenticated Airtop Profile used for LinkedIn-based enrichment. Hubspot object id**: The internal HubSpot ID for the contact to be updated. How It Works Initiates the workflow using a form or external trigger. Uses the name and email to extract and enrich the person’s data, including: LinkedIn profile and company page About section, job title, location ICP score, seniority level, AI interest, technical depth, connection and follower counts Formats and maps the enriched data. Pushes the updated data to HubSpot using the object ID. Setup Requirements Airtop API Key Airtop Profile logged in to LinkedIn. HubSpot access with object ID field for each contact to update. Next Steps Combine with Lead Generation**: Use as part of an end-to-end workflow that sources leads and enriches them in real time. Trigger from CRM**: Initiate this workflow when a new contact is added in HubSpot or another CRM. Customize Scoring Logic**: Tailor the ICP calculation to your team’s specific criteria. Read more about person data enrichment
by Ron
Objective In industry and production sometimes machine data is available in databases. That might be sensor data like temperature or pressure or just binary information. In this sample flow reads machine data and sends an alert to your SIGNL4 team when the machine is down. When the machine is up again the alert in SIGNL4 will get closed automatically. Setup We simulate the machine data using a Notion table. When we un-check the Up box we simulate a machine-down event. In certain intervals n8n checks the database for down items. If such an item has been found an alert is send using SIGNL4 and the item in Notion is updates (in order not to read it again). Status updates from SIGNL4 (acknowledgement, close, annotation, escalation, etc.) are received via webhook and we update the Notion item accordingly. This is how the alert looks like in the SIGNL4 app. The flow can be easily adapted to other database monitoring scenarios.
by Gain FLow AI
Overview This workflow automates the process of sending personalized cold email sequences to your prospects. It fetches un-emailed leads from your Google Sheet, validates their email addresses, and then dispatches tailored emails according to a predefined schedule. It updates your CRM (Google Sheet) with the status of each sent email, ensuring your outreach efforts are tracked and efficient. Use Case This workflow is perfect for: Sales Teams**: Automate the delivery of multi-stage cold email campaigns to a large volume of leads. Business Development**: Nurture prospects over time with a structured email sequence. Recruiters**: Send out introductory emails to potential candidates for open positions. Marketers**: Distribute personalized outreach for events, content, or product launches. Anyone doing cold outreach**: Ensure consistent follow-up and track email performance without manual effort. How It Works Scheduled Trigger: The workflow is set to run automatically at a defined interval (e.g., every 6 hours, as currently configured by the "Set Timer" node). This ensures regular outreach without manual intervention. Fetch Unsent Emails: The "Get Emails" node queries your Google Sheet to identify prospects who haven't yet received the current email in the sequence (i.e., "Email Sent " is "No"). Control Volume: A "Limit" node can be used to control the number of emails sent in each batch, preventing you from sending too many emails at once and potentially hitting sending limits. Loop Through Prospects: The "Loop Over Items" node processes each selected prospect individually. Email Validation (Conditional Send): An "If" node checks if the prospect's "Email Address" is valid and exists. This prevents sending emails to invalid addresses, improving deliverability. Send Email: "Send Email" Node: For valid email addresses, this node dispatches the personalized email to the prospect. It retrieves the recipient's email, subject, and body from your Google Sheet. "connect" Node: (Note: The provided JSON uses a generic emailSend node named "connect" that links to an SMTP credential. This represents the actual email sending mechanism, whether it's Gmail or a custom SMTP server.) Update CRM: After successfully sending an email, the "Update Records" node updates your Google Sheet. It marks the "Email Sent " column as "Yes" and records the "Sent on" timestamp and a "Message Id" for tracking. Delay Between Sends: A "Wait" node introduces a delay between sending emails to individual prospects. This helps mimic human sending behavior and can improve deliverability. How to Set It Up To set up your Automated Cold Email Sender, follow these steps: Google Sheet Setup: Duplicate the Provided Template: Make a copy of the Google Sheet Template (1TjXelyGPg5G8lbPDI9_XOReTzmU1o52z2R3v8dYaoQM) into your own Google Drive. This sheet should contain columns for "Name", "Email Address ", "Sender Email", "Email Subject", "Email Body", "Email Sent ", "Sent on", and "Message Id". Connect Google Sheets: Ensure your Google Sheets OAuth2 API credentials are set up in n8n and linked to the "Get Emails" and "Update Records" nodes. Update Sheet IDs: In both "Get Emails" and "Update Records" nodes, update the documentId with the ID of your copied template. Email Sending Service Credentials: Gmail: If using Gmail, ensure your Gmail OAuth2 credentials are configured and connected to the "Send Email" node (or the "connect" node, if that's your chosen sender). Other Email Services (SMTP): If you use a different email service, you'll need to set up an SMTP credential in n8n and connect it to the "connect" node. Refer to the "Sticky Note4" for guidance on non-Google email services. Configure Timer: In the "Set Timer" node, adjust the hoursInterval or other time settings to define how frequently you want the email sending process to run (e.g., every 6 hours, once a day, etc.). Control Volume (Optional): In the "Limit" node, you can set the maxItems to control how many emails are processed and sent in each batch. This is useful for managing email sending limits or gradual outreach. Import the Workflow: Import the provided workflow JSON into your n8n instance. Populate Your Sheet: Fill your copied Google Sheet with prospect data, including the email subject and body for each email you wish to send. Ensure the "Email Sent " column is initially "No". Activate and Monitor: Activate the workflow. It will begin fetching and sending emails based on your configured schedule. Monitor your Google Sheet to track the "Email Sent " status. This workflow provides a robust and automated solution for managing your cold email campaigns, saving you time and increasing your outreach efficiency.
by Airtop
Recursive Web Scraping Use Case Automating web scraping with recursive depth is ideal for collecting content across multiple linked pages—perfect for content aggregation, lead generation, or research projects. What This Automation Does This automation reads a list of URLs from a Google Sheet, scrapes each page, stores the content in a document, and adds newly discovered links back to the sheet. It continues this process for a specified number of iterations based on the defined scraping depth. Input Parameters: Seed URL: The starting URL to begin the scraping process. Example: https://example.com/ Links must contain: Restricts the links to those that contain this specified string. Example: https://example.com/ Depth: The number of iterations (layers of links) to scrape beyond the initial set. Example: 3 How It Works Starts by reading the Seed URL from the Google Sheet. Scrapes each page and saves its content to the specified document. Extracts new links from each page that match the Links must contain string, appends them to the Google Sheet. Repeats steps 2–3 for the number of times specified by Depth - 1. Setup Requirements Airtop API Key — free to generate. Credentials set up for Google Docs (requires creating a project on Google Console). Read how to. Credentials set up for Google Spreadsheet. Next Steps Add Filtering Rules**: Filter which links to follow based on domain, path, or content type. Combine with Scheduler**: Run this automation on a schedule to continuously explore newly discovered pages. Export Structured Data**: Extend the process to store extracted data in a CSV or database for analysis. Read more about website scraping for LLMS