by Basil Irfan
🚀 LinkedIn Lead-Gen Flywheel – Apify → GPT-4o → Google Sheets → Phantombuster What this workflow does Collect audience specs – simple web-form asks for your ideal company profile. Generate a laser-targeted Apollo search URL with GPT-4o (no manual filtering). Scrape the matching leads via an Apify actor (returns clean JSON). Craft hyper-personalized icebreakers for each lead using GPT-4o (ultra-short, human-sounding). Log everything to Google Sheets – name, LinkedIn URL, company site, summary, and the icebreaker. (Optional) Auto-launch Phantombuster to fire off those connection requests at scale. Why it matters Zero grunt work:** audience research, scraping, copy-writing, and outreach all run hands-free. Punchy personalization:** micro-icebreakers outperform canned intros, boosting accept rates. Scales with you:** flip a switch to go from 10 to 1 000+ connections/day. Node rundown | Step | Node | Key Inputs | Key Outputs | |------|------|-----------|-------------| | 1 | Form Trigger | Audience description | description_of_company | | 2 | OpenAI (GPT-4o) | Audience text | SearchUrl | | 3 | HTTP Request – Apify | SearchUrl, APIFY_TOKEN | Lead JSON | | 4 | OpenAI (GPT-4o) | Lead JSON | Icebreaker | | 5 | Google Sheets | Lead + Icebreaker | Row append/update | | 6 | Aggregate | Sheet rows | Batched output | | 7 | HTTP Request – Phantombuster | PHANTOM_KEY, AGENT_ID | Launch status | Prerequisites OpenAI API key** (GPT-4o access recommended) Apify API token** with access to actor id Google Service Account creds** shared with your target sheet Phantombuster API key** and Agent ID for your LinkedIn connector Active Apollo account to open the generated search URL (only required for debugging) Setup (5-minute sprint) Import the workflow into n8n. Add the required credentials in Credentials → OpenAI, Apify, Google Sheets, Phantombuster. Paste your Phantombuster Agent ID into the HTTP Request node URL. Publish the Form Trigger URL—this is where you (or your SDRs) describe the target audience. Hit Execute Workflow once to verify data flows end-to-end. Customization tips Titles & keywords:** tweak the prompt in the first GPT-4o node to lock in different roles or industries. Icebreaker style:** adjust the second GPT-4o prompt to match your brand voice. Data columns:** map extra fields from Apify into Google Sheets as needed. Skip outreach:** disable the Phantombuster node if you only want the leads + icebreakers.
by Agent Studio
Overview This workflow helps you compare Claude 3.5 Sonnet and Gemini 2.0 Flash when extracting data from a PDF This workflow extracts and processes the data within a PDF in one single step, instead of calling an OCR and then an LLM” How it works The initial 2 steps download the PDF and convert it to base64. This base64 string is then sent to both Claude 3.5 Sonnet and Gemini 2.0 Flash to extract information. This workflow is made to let you compare results, latency, and cost (in their dedicated dashboard). How to use it Set up your Google Drive if not already done Select a document on your Google Drive Modify the prompt in "Define Prompt" to extract the information you need and transform it as wanted. Get a Claude API key and/or Gemini API key Note that you can deactivate one of the 2 API calls if you don't want to try both Test the Workflow
by n8n Team
This workflow syncs your GitHub issues to your Notion database. Whenever a new issue is opened in your GitHub repository, it will be shown in your Notion database, syncing the status property (opened/edited/closed/deleted). In case there’s no Notion database existing yet, a new one will be created automatically. Prerequisites Notion account and Notion credentials GitHub account and GitHub credentials How it works Github trigger starts the workflow when a new issue is created in a GitHub repository. If node splits the workflow conditionally, showing whether the issue is new or an update of an existing issue. If data is new, the Notion node will create a new database page in Notion. If the data is not new, the Function node will create a Notion filter that will find its specific database page by issue ID. Switch node will then conditionally route the data into the appropriate Notion page, based on the update made upon it.
by Ibrahim
Overview This n8n workflow is designed to extract specific interests from messages in a Telegram chat and retrieve related information using the Facebook Graph API. It aims to provide a streamlined solution for parsing and analyzing user-provided interests within the Telegram platform. Features Interest Extraction:** Automatically identifies and extracts interests from messages that start with the hashtag "#interest". Data Retrieval:** Utilizes the Facebook Graph API to retrieve information related to the extracted interests. Structured Outputs:** Presents the retrieved data in an organized format for further analysis and review. Requirements Operational instance of n8n (self-hosted or cloud version). Basic understanding of n8n workflows and nodes. Setup and Configuration Import Workflow: Load the provided JSON workflow into your n8n instance. Configure Telegram Trigger Node: Ensure the Telegram trigger node is set up with the appropriate credentials and webhook ID. Configure and Test Nodes: Adjust node parameters as necessary and test the workflow to ensure proper functionality. How it Works Telegram Trigger: Listens for incoming messages in a specified Telegram chat. Check Message Contents: Verifies if the message begins with the specified hashtag and is from the designated chat ID. Extract Message: Extracts the content of the message for further processing. Split Message: Splits the extracted message to identify the interest and remaining content. Connect to Graph API: Utilizes the Facebook Graph API to search for information related to the extracted interest. Split Interests into a Table: Organizes the retrieved data into a structured table format. Get Variables: Maps the retrieved data to create a new JSON object containing specific fields related to the interest. Create a Spreadsheet: Generates a spreadsheet file in CSV format based on the retrieved and formatted data. Send the Spreadsheet File: Sends the generated spreadsheet file back to the original Telegram chat. Customization Modify the filtering conditions and fields to suit specific requirements. Adjust the frequency of the trigger node based on preference. Best Practices Regularly test the workflow to ensure consistent performance. Stay informed about any changes to external APIs that might affect the workflow's functionality. Contributing Your feedback and contributions are highly valued. Feel free to adapt, modify, and share enhancements with the n8n community.
by Joey D’Anna
This template is a set of building blocks to access Monday.com in ways not supported by the official Monday node. Prerequisites Monday account and Monday credentials. Included are setups to: Find a column value by the column's name (instead of a numerical index which can change when board structure is changed) Find a column value by the column's ID (again, instead of using a numerical index) Pull a board relation column, and get all the related pulses Pull an items subitems and split them out Upload a file to an item's files field Setup Create a Monday.com credential Update the nodes in the template to use your credential Copy/Paste the nodes you need from this template into any other workflow To retreive a column by name: Route a Monday.com node that gets an item to the COLUMN BY NAME node Edit the COLUMN BY NAME node, and enter the name in the first line of code. To retreive a column by its ID: Follow Monday.com's instructions to locate the column's ID Route a Monday.com node that gets an item to the COLUMN BY ID node -Edit the COLUMN BY ID node, and enter the ID in the first line of code. To retreive all linked pulses from a Board Relation column: Route a Monday.com node that gets an item to the GET BOARD RELATION node Edit the GET BOARD RELATION node to specify the column name. All linked pulses will be retrieved by the subsequent PULL LINKEDPULSE node To pull all subitems from an item: Route a Monday.com node that gets an item to the PULL SUBITEMS node All subitems will be retrieved by the subsequent GET EACH SUBITEM node To upload a File: Repalce the Convert to File node with whatever node you are using to output your binary file data Enable the MONDAY UPLOAD node If the destination column is named anything other then the default of "file" - edit the MONDAY UPLOAD node and change column_id:"file" in the first Value field to match the name of your file column
by Corentin Ribeyre
This template can be used to verify an email address with Icypeas. Be sure to have an active account to use this template. How it works This workflow can be divided into three steps : The workflow initiates with a manual trigger (On clicking 'execute'). It connects to your Icypeas account. It performs an HTTP request to verify an email address. Set up steps You will need a working icypeas account to run the workflow and get your API Key, API Secret and User ID. You will need an email address to perform the verification.
by Growth AI
Automated trend monitoring for content strategy Who's it for Content creators, marketers, and social media managers who want to stay ahead of emerging trends and generate relevant content ideas based on data-driven insights. What it does This workflow automatically identifies trending topics related to your industry, collects recent news articles about these trends, and generates content suggestions. It transforms raw trend data into actionable editorial opportunities by analyzing search volume growth and current news coverage. How it works The workflow follows a three-step automation process: Trend Analysis: Examines searches related to your topics and identifies those with the strongest recent growth Article Collection: Searches Google News for current articles about emerging trends and scrapes their full content Content Generation: Creates personalized content suggestions based on collected articles and trend data The system automatically excludes geo-localized searches to provide a global perspective on trends, though this can be customized. Requirements SerpAPI account (for trend and news data) Firecrawl API key (for scraping article content from Google News results) Google Sheets access AI model API key (for content analysis and recommendations - you can use any LLM provider you prefer) How to set up Step 1: Prepare your tracking sheet Duplicate this Google Sheets template Rename your copy and ensure it's accessible Step 2: Configure API credentials Before running the workflow, set up the following credentials in n8n: SerpAPI: For trend analysis and Google News search Firecrawl API: For scraping article content AI Model API: For content analysis and recommendations (Anthropic Claude, OpenAI GPT, or any other LLM provider) Google Sheets OAuth2: For accessing and updating your tracking spreadsheet Step 3: Configure your monitoring topics In your Google Sheet "Query" tab: Query column: Enter the main topics/keywords you want to monitor for trending queries (e.g., "digital marketing", "artificial intelligence", "sustainable fashion") Query to avoid column: Optionally add specific queries you want to exclude from trend analysis (e.g., brand names, irrelevant terms, or overly specific searches that don't match your content strategy) This step is crucial as these queries will be the foundation for discovering related trending topics. Step 4: Configure the workflow In the "Get Query" node, paste your duplicated Google Sheets URL in the "Document" field Ensure your Google Sheet contains your monitoring topics in the Query column Step 5: Customize language and location settings The workflow is currently configured for French content and France location. You can modify these settings in the SerpAPI nodes: Language (hl): Change from "fr" to your preferred language code Geographic location (geo/gl): Change from "FR" to your target country code Date range: Currently set to "today 1-m" (last month) but can be adjusted Step 6: Adjust filtering (optional) The "Sorting Queries" node excludes geo-localized queries by default. You can modify the AI agent's instructions to include location-specific queries or change filtering criteria based on your requirements. The system will also automatically exclude any queries you've listed in the "Query to avoid" column. Step 7: Configure scheduling (optional) The workflow includes an automated scheduler that runs monthly (1st day of each month at 8 AM). You can modify the cron expression 0 8 1 * * in the Schedule Trigger node to change: Frequency (daily, weekly, monthly) Time of execution Day of the month How to customize the workflow Change trend count: The workflow processes up to 10 related queries per topic but filters them through AI to select the most relevant non-geolocalized ones Adjust article collection: Currently collects exactly 3 news articles per query for analysis Content style: Customize the AI prompts in content generation nodes to match your brand voice Output format: Modify the Google Sheets structure to include additional data points AI model: Replace the Anthropic model with your preferred LLM provider Scraping options: Configure Firecrawl settings to extract specific content elements from articles Results interpretation For each monitored topic, the workflow generates a separate sheet named by month and topic (e.g., "January Digital Marketing") containing: Data structure (four columns): Query: The trending search term ranked by growth Évolution: Growth percentage over the last month News: Links to 3 relevant news articles Idée: AI-generated content suggestions based on comprehensive article analysis The workflow provides monthly retrospective analysis, helping you identify emerging topics before competitors and optimize your content calendar with high-potential subjects. Workflow limitations Processes up to 10 related queries per topic with AI filtering Collects exactly 3 news articles per query Results are automatically organized in monthly sheets Requires stable internet connection for API calls
by Evervise
🎯 AI Landing Page Analyzer & Optimizer Transform your landing page audits into a powerful lead generation machine with this professional n8n workflow powered by 4 specialized AI agents. What It Does This workflow analyzes any landing page and delivers a comprehensive, $2,000-value audit report in under 90 seconds. Perfect for agencies, consultants, and SaaS companies looking to generate high-quality leads while showcasing their expertise. 🤖 Four Specialized AI Agents Design Critic - Analyzes UX/UI, visual hierarchy, CTA placement, mobile responsiveness, and trust signals Copywriter - Reviews messaging, headlines, value propositions, and emotional triggers SEO Specialist - Audits technical SEO, meta tags, heading structure, and performance indicators Growth Strategist - Synthesizes all findings, assigns an A-F grade, and creates prioritized action plans ✨ Key Features Comprehensive Scoring**: 6-dimension scorecard with detailed justifications Actionable Insights**: Top 5 priorities ranked by impact vs. effort Conversion Lift Estimates**: Conservative, realistic, and optimistic projections Beautiful HTML Reports**: Professional email with all analyses, branded for your business Natural Upsell Path**: Built-in CTA for paid optimization services Fast Delivery**: Complete analysis in 60-90 seconds 📊 What Gets Analyzed Design & UX Visual hierarchy and layout structure CTA design, placement, and effectiveness Color scheme and whitespace usage Trust signals and social proof Mobile responsiveness indicators Copywriting Headline impact and hook effectiveness Value proposition clarity Benefits vs. features balance Emotional triggers and persuasion CTA copy strength and action orientation SEO & Technical Title tag and meta description optimization Heading structure (H1, H2 hierarchy) Image optimization and alt tags Mobile-friendliness Analytics tracking setup Page speed indicators Strategic Overview Overall grade (A+ to F) with detailed explanation Comprehensive scorecard across all dimensions Prioritized action items by ROI Estimated conversion lift potential Strategic recommendations and next steps 💼 Perfect For Digital Marketing Agencies**: Offer free audits to generate leads Freelance Consultants**: Showcase expertise and attract clients SaaS Companies**: Lead magnet for conversion optimization tools Web Development Agencies**: Pre-sales qualification and demonstrations CRO Specialists**: Automated initial assessments 📋 What You Need Required n8n instance (self-hosted or cloud) Anthropic API key (Claude Sonnet 4.5) Gmail account for sending reports (or any SMTP provider) Optional Enhancements Screenshot API (UrlBox, Puppeteer) for visual captures Google PageSpeed Insights API for performance testing CRM integration (HubSpot, Salesforce, Pipedrive) Slack notifications for lead alerts Calendar booking integration (Calendly) ⚙️ Technical Details AI Model**: Claude Sonnet 4.5 (4 separate agents) Average Runtime**: 60-90 seconds Cost Per Analysis**: ~$0.15-0.25 (API costs) Form Fields**: 6 (landing page URL, industry, goal, conversion rate, frustration, email) Output**: Beautiful HTML email report with comprehensive analysis 🎨 Customization Options The workflow includes detailed documentation and is fully customizable: Adjust agent prompts for your specific niche Modify scoring criteria and thresholds Customize email branding and design Add/remove form fields Integrate with your CRM Segment responses by score (different CTAs for different grades) 📈 Expected Results Lead Quality**: High (users actively seeking optimization) Conversion Rate**: 15-30% of audit recipients book calls Time Saved**: 2-3 hours of manual analysis per audit Perceived Value**: $2,000+ professional audit 🔧 Setup Difficulty Intermediate - Requires basic n8n knowledge and API key setup Setup Steps Import workflow to n8n Add Anthropic API credentials Configure Gmail/SMTP credentials Customize form and email branding Test with sample landing page Deploy form on your website 📚 Included Documentation Comprehensive sticky notes explaining each component Setup instructions and prerequisites Customization guide Monetization strategy breakdown Performance optimization tips Enhancement ideas for v2 🌟 Use Cases Lead Magnet: Embed form on your website to capture qualified leads Sales Tool: Use during discovery calls to demonstrate value Content Marketing: Offer audits in LinkedIn posts, email campaigns Partner Program: Provide white-labeled audits to partners Upsell Sequence: Follow up with paid optimization services ⚡ Why This Workflow? Unlike simple template-based audits, this workflow uses real AI intelligence to provide nuanced, contextual insights. Each analysis is unique and considers: Industry-specific best practices Stated business goals Current pain points Competitive landscape User's experience level The result? Reports that feel hand-crafted by experts, not generic checklists. 🤝 Support 📖 Website: https://evervise.ai/ ✨ Support: mark.marin@evervise.com N8N Link 📊 Version History v1.0** - Initial release with 4-agent analysis pipeline Coming soon: Screenshot capture, competitor comparison, visual mockups Tags lead-generation marketing-automation ai-agents conversion-optimization landing-pages anthropic claude audit seo copywriting design-analysis lead-magnet Ready to turn landing page audits into a lead generation machine? Import this workflow and start attracting high-quality clients today.
by Lucas Peyrin
How it works This workflow automates your initial hiring pipeline by creating an AI-powered CV scanner. It collects job applications through a web form, uses AI to analyze the candidate's CV against your job description, and neatly organizes the results in a Google Sheet. Here’s the step-by-step process: The Application Form:** A Form Trigger provides a public web form for candidates to submit their name, email, and CV (as a PDF). Initial Logging:** As soon as an application is submitted, the candidate's name and email are added to a Google Sheet. This ensures every applicant is logged, even if a later step fails. CV Text Extraction:* The workflow uses *Mistral's OCR** model to accurately extract all the text from the uploaded CV PDF. AI Analysis:* The extracted text is sent to *Google Gemini**. A detailed prompt instructs the AI to act as a hiring assistant, scoring the CV against the specific requirements of your job role and providing a detailed explanation for its score. Structured Output:** A JSON Output Parser ensures the AI's analysis is returned in a clean, structured format, making the data reliable. Final Record:** The AI-generated qualification score and explanation are added to the candidate's row in the Google Sheet, giving you a complete, analyzed list of applicants. Set up steps Setup time: ~15 minutes You'll need API keys for Mistral and Google AI, and to connect your Google account. Get Your Mistral API Key: Visit the Mistral Platform at console.mistral.ai/api-keys. Create and copy your API key. In the workflow, go to the Extract CV Text node, click the Credential dropdown, and select + Create New Credential. Paste your key into the API Key field and Save. Get Your Google AI API Key: Visit Google AI Studio at aistudio.google.com/app/apikey. Click "Create API key in new project" and copy the key. In the workflow, go to the Gemini 2.5 Flash Lite node, click the Credential dropdown, and select + Create New Credential. Paste your key into the API Key field and Save. Connect Your Google Account: Select the Create 'CVs' Spreadsheet node. Click the Credential dropdown and select + Create New Credential to connect your Google account. Repeat this for the Log Candidate Submission and Add CV Analysis nodes, selecting the credential you just created. Create Your Spreadsheet: Click the "play" icon on the Start Here node to run it. This will create a new Google Sheet in your Google Drive named "CVs" with the correct columns. Customize the Job Role: Go to the AI Qualification node. In the Text parameter, find the job_requirements section and replace the example job description with your own. Be as detailed as possible for the best results. Start Screening! Activate the workflow using the toggle at the top right. Go to the Application Form node and click the "Open Form URL" button. Fill out the form with a test application and upload a sample CV. Check your Google Sheet to see the AI's analysis appear within moments
by Derek Cheung
How it works: This project teaches you to create a personal AI assistant named Jackie that operates through Telegram. Jackie can summarize unread emails, check calendar events, manage Google Tasks, and handle both voice and text interactions. The assistant provides a comprehensive digital life management solution accessible via Telegram messaging. Key Features: Supports hands-free voice interaction Maintains conversation memory Integrates with major Google services Provides personalized assistance for email management, scheduling, and task organization Step-by-step: Telegram Trigger: The workflow starts with a Telegram trigger that listens for incoming message events. The system determines if the incoming message is voice or text input. Voice Processing: If a voice message is received, the workflow retrieves the voice file from Telegram and uses OpenAI's transcription API to convert speech to text. AI Assistant: The processed text (whether original text or transcribed voice) is passed to Jackie, the AI assistant powered by OpenRouter's language model. Tools Integration: Jackie is equipped with several productivity tools: Get Email: Uses Gmail API to fetch unread emails from the inbox with sender, date, subject, and summary information Google Calendar: Retrieves calendar events for specified dates, filtering out irrelevant future events Google Tasks: Both creates new tasks and retrieves existing tasks from Google Tasks lists API Keys Required: Telegram Bot API: Create a bot via @BotFather on Telegram to get your bot token OpenAI API: Required for voice-to-text transcription OpenRouter API: Powers the AI language model responses Google OAuth2: Needed for Gmail, Google Calendar, and Google Tasks integration Response Generation: The AI formulates intelligent responses based on the gathered information, current date context, and conversation history, then sends the response back to the user via Telegram in Markdown format.
by Khairul Muhtadin
Say goodbye to endless applications and hello to more time for perfecting your interview skills! The JOB Hunter Agent uses the power of Google Gemini and SerpAPI to find the perfect job match and generate a personalized cover letter. Result example: 💡 Why Use JOB Hunter Agent? Save Precious Time: Stop sifting through countless job boards; this agent does the heavy lifting, saving you hours every week. Land Your Dream Job Faster: Get laser-focused job matches and a custom-crafted cover letter that speaks directly to the hiring manager, increasing your chances of getting noticed. Never Miss an Opportunity: Your personal AI assistant works 24/7, ensuring you're always on top of the latest openings, even while you sleep! Stand Out from the Crowd: A perfectly tailored cover letter generated on the fly gives you an edge over generic applications, making you look like a superstar. ⚡ Perfect For Job Seekers: Anyone actively looking for a new role who wants to streamline their application process. Busy Professionals: Those with limited time who need an efficient way to find relevant opportunities. Career Changers: Individuals exploring new industries who need a helping hand in crafting compelling applications. 🔧 How It Works ⏱ Trigger: You submit your CV and job preferences through a simple n8n form. 📎 Process: Your CV is extracted from the PDF, and your preferences (location, job type, salary, email) are neatly organized. 🤖 Smart Logic: The "Job Hunter Agent" uses Google Gemini and SerpAPI to find the single best job match for you and then drafts a bespoke cover letter based on your CV and the job description. It's like magic, but with more code! 💌 Output: A beautifully formatted HTML email containing your profile summary, the best job match, your personalized cover letter, and handy application tips is sent straight to your inbox. 🗂 Storage: All the initial data from your form submission is processed and used to craft your perfect job application package. 🔐 Quick Setup Import JSON file to your n8n instances Add credentials: Google Gemini (Gemini 2.5 Pro model) and SerpAPI Customize: Adjust the system prompt in the "Job Hunter Agent" to fine-tune the cover letter tone or length, update the email footer, and expand job filters. Update: Ensure your Gmail OAuth2 credentials are valid for sending emails. Test: Run the workflow with your own CV and preferences to see the magic happen! 🧩 You'll Need Active n8n instances Google Gemini API key (for Gemini 2.5 Pro) SerpAPI account for Google Jobs search results A Gmail account for sending personalized job match emails 🛠️ Level Up Ideas Integrate with LinkedIn, Jobstreet, or Indeed APIs for a wider range of job sources. Allow the agent to find multiple job matches and present them as a curated list. Add an option to attach a parsed CV summary as a PDF to the email for quick reference. Made by: khaisa Studio Tags: AI, Gemini, Google Jobs, Job Search, Automation, Cover Letter Category: job hunter Need custom work? Contact me
by Sirhexalot
This n8n workflow allows you to update user roles in Zammad based on data from an Excel file. The workflow automates role assignments, ensuring efficient and consistent updates. Features Excel Integration**: Import user data from an Excel file containing emails and role assignments. Dynamic Updates**: Match Zammad users by email and update their roles. Error Handling**: Continue workflow execution even if some updates fail. Customizable Variables**: Configure Zammad API URL, API key, and Excel file URL. Usage Import the Workflow: Upload the provided .json file into your n8n instance. Set Variables: zammad_base_url: Your Zammad instance URL. excel_source_url: URL of the Excel file containing user data. Authentication for Zammad Create in the Node "Find Zammad User by email" and "Update User Roles" a Header Auth Authentication Name**: Authorization Value**: Bearer <put here your zammad api token> Run the Workflow: Execute the workflow to update user roles based on the Excel data. Issues and Suggestions For issues or suggestions, visit the GitHub Repository.