by Ranjan Dailata
1. Who this is for This workflow is specifically designed for Recruiters, HR analytics teams, and data-driven talent acquisition professionals seeking deeper insights from candidate resume. Valuable for HR tech developers, ATS/CRM engineers, and AI-driven recruitment platforms aiming to automate candidate research. Helps organizations build predictive hiring models and gain actionable talent intelligence. 2. What problem this workflow solves Recruiters often face information overload when analyzing candidate resume manually reviewing experiences, skills, and cultural fit is slow and inconsistent. Traditional scraping tools extract raw data but fail to produce actionable intelligence like career trajectory, skills alignment, and fit for a role. This workflow solves that by: Automating candidate resume data extraction through Decodo Structuring it into JSON Resume Schema Running deep AI-driven analytics using OpenAI GPT-4o-mini Delivering comprehensive candidate intelligence ready for ATS/CRM integration or HR dashboards 3. What this workflow does This n8n workflow combines Decodo’s web scraping with OpenAI GPT-4o-mini to produce advanced recruitment intelligence. Flow Breakdown: Manual Trigger – Start the workflow manually or schedule it in n8n. Set Input Fields – Define resume URL, location, and job description. Decodo Node – Scrapes the candidate’s profile (experience, skills, education, achievements, etc.). Structured Data Extractor (GPT-4o-mini) – Converts the scraped data into a structured JSON Resume Schema. Advanced Data Mining Engine (GPT-4o-mini) – Performs: Skills Analysis (strengths, gaps, transferable skills) Experience Intelligence (career trajectory, leadership, project complexity) Cultural Fit Insights (work style, communication style, agility indicators) Career Trajectory Forecasting (promotion trends, growth velocity) Competitive Advantage Analysis (market positioning, salary expectations) Summarizer Node – Produces an abstractive and comprehensive AI summary of the candidate profile. Google Sheets Node – Saves the structured insights automatically into your recruitment intelligence sheet. File Writer Node (Optional) – Writes the JSON report locally for offline storage or integration. The result is a data-enriched candidate intelligence report far beyond what traditional resume parsing provides. 4. Setup Prerequisites If you are new to Decode, please signup on this link visit.decodo.com n8n account with workflow editor access Decodo API credentials OpenAI API key Google Sheets account connected via OAuth2 Make sure to install the Decodo Community node. Setup Steps Import the workflow JSON into your n8n workspace. Set credentials for: Decodo Credentials account OpenAI API (GPT-4o-mini) Google Sheets OAuth2 In the “Set the Input Fields” node, update: url → Resume link geo → Candidate region or country jobDescription → Target job description for matching Ensure the Google Sheet ID and tab name are correct in the “Append or update row in sheet” node. Click Execute Workflow to start. 5. How to customize this workflow You can adapt this workflow for different recruitment or analytics scenarios: Add Sentiment Analysis Add another LLM node to perform sentiment analysis on candidate recommendations or feedback notes. Enrich with Job Board Data Use additional Decodo nodes or APIs (Indeed, Glassdoor, etc.) to compare candidate profiles to live job postings. Add Predictive Fit Scoring Insert a Function Node to compute a numerical "fit score" by comparing skill vectors and job requirements. Automate Candidate Reporting Connect to Gmail, Slack, or Notion nodes to automatically send summaries or reports to hiring managers. 6. Summary The Advanced Resume Intelligence & Data Mining via Decodo + OpenAI GPT-4o-mini workflow transforms traditional candidate sourcing into AI-driven intelligence gathering. It integrates: Decodo** → To perform webscraping of data GPT-4o-mini** → to interpret, analyze, and summarize with context Google Sheets** → to store structured results for real-time analysis With this system, recruiters and HR analysts can move from data collection to decision intelligence, unlocking faster and smarter talent insights.
by Joachim Hummel
This workflow automates the daily backup of all your n8n workflows to a designated folder in Nextcloud. It ensures that you always have the last 7 days of backups available while automatically deleting older ones to save space. 🔧 Features Scheduled Trigger: Runs automatically once per day (can be executed manually as well). Directory Management: Creates the /N8N-Backup directory in Nextcloud if it doesn't already exist. Backup Collection: Retrieves all workflows from the n8n instance. JSON Conversion: Converts each workflow into a JSON file. Upload to Nextcloud: Saves each backup file into the specified backup directory. Retention Control: Keeps only the latest 7 backups and deletes the rest from Nextcloud. 📌 Notes Make sure to manually create the /N8N-Backup directory in your Nextcloud account before using this flow. Update the Backup Path node if you wish to change the upload directory. Ideal for teams using n8n with self-hosted instances and requiring offsite backup via Nextcloud. 🔒 Requirements n8n instance with access to the Nextcloud node. Valid credentials for your Nextcloud account with API access. Update: 08/11/2025 “Backup Flows to Nextcloud” – Import format fixed Summary: The workflow now exports one clean JSON object per workflow (no arrays, no backup/meta fields), so files can be imported 1:1 via the n8n UI. What changed: Switched from “Convert to File” to a Set node that builds the JSON in binary data. Enabled filters.include = "all" on Get many workflows to include nodes, connections, settings, pinData, and tags. Sanitized filenames and removed IDs/metadata that can break UI imports. Fixed Nextcloud path and binary property mapping (data). Verification: Generated multiple backups and imported each via UI (“Import from file”) without errors. Each file begins with { (single object) and loads with full workflow structure. Notes: Keep “Binary Property” set to data in the Nextcloud node. Filenames are sanitized to avoid special-character issues.
by Oneclick AI Squad
Acts as a virtual receptionist for the restaurant, handling incoming calls via VAPI without human intervention. It collects user details (name, booking time, number of people) for table bookings, checks availability in a PostgreSQL database using n8n, books the table if available, and sends confirmation. It also provides restaurant details to users, mimicking a human receptionist. Key Insights VAPI must be configured to accurately capture user input for bookings and inquiries. PostgreSQL database requires a table to manage restaurant bookings and availability. Workflow Process Initiate the workflow with a VAPI call to collect user details (name, time, number of people). Use n8n to query the PostgreSQL database for table availability. If a table is available, book it using n8n and update the PostgreSQL database. Send a booking confirmation and hotel service details back to VAPI via n8n. Store and update restaurant table data in the PostgreSQL database using n8n. Usage Guide Import the workflow into n8n and configure VAPI and PostgreSQL credentials. Test with a sample VAPI call to ensure proper data collection and booking confirmation. Prerequisites VAPI API credentials for call handling PostgreSQL database with booking and availability tables Customization Options Modify the VAPI input fields to capture additional user details or adjust the PostgreSQL query for specific availability criteria.
by n8n Team
This workflow performs several data integration and synchronization tasks between Google Sheets and a MySQL database. Here is a step-by-step description of what this workflow does: Manual Trigger: The workflow starts when the user clicks "Execute Workflow." Schedule Trigger: This node schedules the workflow to run at specific intervals on weekdays (Monday to Friday) between 6 AM and 10 PM. It ensures regular data synchronization. Google Sheet Data: This node connects to a specific Google Sheets document and retrieves data from the "Form Responses 1" sheet, filtering by the "DB Status" column. SQL Get inquiries from Google: This node retrieves data from a MySQL database table named "ConcertInquiries" where the "source_name" is "GoogleForm." Rename GSheet variables: This node renames the columns retrieved from Google Sheets and transforms the data into a format suitable for MySQL, assigning a value for "source_name" as "GoogleForm." Compare Datasets: This node compares the data retrieved from Google Sheets and the MySQL database based on timestamp and source_name fields. It identifies changes and updates. No reply too long?: This node checks if there has been no reply within the last four hours, using the "timestamp" field from the Google Sheets data. DB Status assigned?: This node checks if the "DB Status" field is not empty in the compared dataset. Update GSheet status: If conditions are met in the previous nodes, this node updates the "DB Status" field in Google Sheets with the corresponding value from the MySQL dataset. DB Status in sync?: This node checks if the "source_name" field in Google Sheets is not empty. Sync MySQL data: If conditions are met in the previous nodes, this node updates the "source_name" field in the MySQL database to "GoogleFormSync." Send Notifications: If conditions are met in the "No reply too long?" node, this node sends notifications or performs actions as needed. Sticky Notes: These nodes provide additional information and documentation links for users.
by Yaron Been
This workflow provides automated access to the 0Xdino Cyberrealistic Pony V125 AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the 0Xdino Cyberrealistic Pony V125 model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the 0Xdino/cyberrealistic-pony-v125 AI model 0Xdino Cyberrealistic Pony V125**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Yaron Been
This workflow provides automated access to the Biggpt1 Olana Marz AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Biggpt1 Olana Marz model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Biggpt1/olana-marz AI model Biggpt1 Olana Marz**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically discovers and collects information about events and attendee data from event platforms. It saves you time by eliminating the need to manually browse through event listings and provides a centralized database of event information including categories, venues, and attendee feedback. Overview This workflow automatically scrapes event data from 10times.com and other event platforms to extract categories, featured events, attendee feedback, and venue information. It uses Bright Data to access event websites without being blocked and AI to intelligently parse event data into structured format for storage in Google Sheets. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping event websites without being blocked OpenAI**: AI agent for intelligent event data extraction and parsing Google Sheets**: For storing and organizing event information How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and specify the target spreadsheet Customize: Adjust the event platform URLs and event criteria you want to monitor Use Cases Event Planners**: Monitor competing events and industry trends Marketing Teams**: Identify events for sponsorship and networking opportunities Business Development**: Find relevant events for lead generation and partnerships Market Research**: Track event attendance patterns and industry insights Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #events #eventdiscovery #brightdata #webscraping #eventplanning #eventscraping #attendeedata #eventmarketing #n8nworkflow #workflow #nocode #eventautomation #eventmonitoring #eventresearch #10times #eventintelligence #venuedata #eventfeedback #eventtracking #eventcalendar #eventanalytics #businessevents #eventorganizer #eventtech #eventindustry #eventcollection #networkingevents #conferencedata
by Yaron Been
This workflow automatically analyzes website conversion funnels to identify optimization opportunities and track user journey performance. It saves you time by eliminating the need to manually analyze funnel metrics and provides detailed insights into conversion bottlenecks and improvement areas. Overview This workflow automatically scrapes website pages to analyze funnel elements including CTAs, tracking scripts, page structure, and conversion paths. It uses Bright Data to access websites without restrictions and AI to intelligently extract funnel data, identify conversion elements, and provide optimization recommendations. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping website pages without being blocked OpenAI**: AI agent for intelligent funnel analysis and optimization insights Google Sheets**: For storing funnel analysis data and recommendations How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your funnel analysis spreadsheet Customize: Define target website URLs and funnel analysis parameters Use Cases Conversion Optimization**: Identify and fix conversion funnel bottlenecks UX Analysis**: Analyze user experience and journey optimization opportunities Competitor Research**: Study competitor funnel strategies and implementations A/B Testing**: Monitor funnel performance changes over time Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #funnelanalysis #conversionoptimization #brightdata #webscraping #uxanalysis #n8nworkflow #workflow #nocode #websiteanalysis #funneloptimization #conversiontracking #userjourney #websiteoptimization #cro #digitalmarketing #funnelalyzer #websiteperformance #conversionanalytics #uxresearch #websitemetrics #funnelmonitoring #performanceanalysis #websiteinsights #conversionfunnel #userexperience #websiteaudit #funneltracking #optimizationanalysis
by Yaron Been
This workflow provides automated access to the Ligua033 Lorealcantara AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Ligua033 Lorealcantara model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Ligua033/lorealcantara AI model Ligua033 Lorealcantara**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Yaron Been
This workflow provides automated access to the Creativeathive Lemaar Door Urban AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Creativeathive Lemaar Door Urban model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Creativeathive/lemaar-door-urban AI model Creativeathive Lemaar Door Urban**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Yaron Been
This workflow provides automated access to the Alitas126 Alitas2 AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Alitas126 Alitas2 model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Alitas126/alitas2 AI model Alitas126 Alitas2**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Anthony
Disclaimer: This template only works on self-hosted for now, as it uses a community node. Use Case Web scrapers often break due to web page layout changes. This workflow attempts to mitigate this problem by auto-generating web scraping data extractor code via LLM. How It Works This workflow leverages ScrapeNinja n8n community node to: scrape webpage HTML, feed it into LLM (Google Gemini) and ask to write a JS extractor function code, then it executes the written JS extractor against scraped HTML to extract useful data from webpage (the code is safely executed in a sandbox) Installation To install ScrapeNinja n8n node, in your self-hosted instance, go to Settings -> Community nodes, enter "n8n-nodes-scrapeninja", and install. Make sure you are using at least v0.3.0. See this in action: https://www.linkedin.com/feed/update/urn:li:activity:7289659870935490560/