by Adam Janes
This workflow demonstrates a simple way to run evals on a set of test cases stored in a Google Sheet. The example we are using comes from an info extraction task dataset, where we tested 6 different LLMs on 18 different test cases. This workflow extends the functionality of my simple eval for benchmarking legal tasks here. Rather than running executions sequentially (waiting for each one to respond before making another request), we use parallel processing to fire 2 requests every second. You can see our sample data in this spreadsheet here to get started. Once you have this working for our dataset, you can plug in your own test cases matching different LLMs to see how it works with your own data. How it works Pull our test cases from Google Sheets. For each case, fire off an HTTP request to a webhook. That webhook grabs the relevant source file from Google Drive and converts it to text. The text gets sent to an LLM via Open Router (so we can easily swap out models). Results come back and are logged in Google Sheets. Set up steps: Add your credentials for Google Sheets, Google Drive, and OpenRouter. Make a copy of the original data spreadsheet so that you can edit it yourself. You will need to plug your version in the Update Results node to see the spreadsheet update on each run of the loop.
by n8n Team
This template quickly shows how to use RAG in n8n. Who is this for? This template is for everyone who wants to start giving knowledge to their Agents through RAG. Requirements Have a PDF with custom knowledge that you want to provide to your agent. Setup No setup required. Just hit Execute Workflow, upload your knowledge document and then start chatting. How to customize this to your needs Add custom instructions to your Agent by changing the prompts in it. Add a different way to load in knowledge to your vector store, e.g. by looking at some Google Drive files or loading knowledge from a table. Exchange the Simple Vector Store nodes with your own vector store tools ready for production. Add a more sophisticated way to rank files found in the vector store. For more information read our docs on RAG in n8n.
by Avkash Kakdiya
🔁 What This Workflow Does This automation fetches daily AI-related articles from trusted RSS feeds, summarizes them using OpenAI (GPT), and generates a ready-to-post LinkedIn update in your writing style. It then emails the post to you every morning for review and publishing. High-Level Steps: Triggers every morning via Cron. Fetches latest AI news from multiple RSS sources. Filters recent articles (last 24 hrs). Summarizes each article using OpenAI (ChatGPT). Generates a LinkedIn-style post using your tone. Sends the post to your Gmail for review. ⚙️ Setup Steps Estimated setup time: 15–30 minutes You’ll need: OpenAI API key Gmail account connected in n8n RSS feed URLs (defaults are provided) Add your email in the Gmail node to receive daily posts. Add your tone/style prompt in the ChatGPT nodes (instructions inside workflow).
by ikbendion
Reddit Poster to Discord This workflow checks Reddit every 15 minutes for new posts and sends selected posts to a Discord channel via webhook. Flow Overview: Schedule Trigger Runs every 15 minutes. Fetch Latest Posts Retrieves up to 3 new posts from any subreddit. Filter Posts Skips moderator or announcement posts based on author ID. Fetch Full Post Data Gets full details for the remaining post. Extract Image URL Parses the post to extract a direct image link. Send to Discord Sends the post title, image, and link to a Discord webhook. Setup Notes: Create a Reddit app and connect credentials in n8n. Add your subreddit name to both Reddit nodes. Connect a Discord webhook for posting.
by Matthieu
LinkedIn Profile Enrichment Automation Who is this for? This template is perfect for sales teams, recruiters, marketing professionals, and business development specialists who need to gather comprehensive LinkedIn profile data at scale. Ideal for lead generation teams building prospect databases, recruiters sourcing candidate information, sales professionals researching prospects, and marketing teams creating targeted outreach campaigns. What problem does this workflow solve? Manually collecting detailed information from LinkedIn profiles is incredibly time-consuming and prone to inconsistency. Visiting each profile individually to extract names, job titles, experience, education, skills, and contact details can take hours for even small prospect lists. This automation eliminates the tedious manual data entry while ensuring consistent, comprehensive profile enrichment. What this workflow does This workflow automatically enriches a list of LinkedIn profile URLs by extracting comprehensive professional data including: Personal details** (first name, last name, headline, location) Professional status** (hiring status, open to work indicators) Network metrics** (connections, followers count) Work experience** (up to 5 most recent positions with company details, dates, and roles) Education background** (up to 3 educational institutions with degrees and dates) Skills and languages** (complete skill sets and language proficiencies) Professional summary** (profile description and bio) The enriched data is automatically organized and updated in your Google Sheets database with structured formatting for easy analysis and outreach. Setup Create a Ghost Genius API account and obtain your API key for cookieless LinkedIn profile scraping Configure HTTP Request credentials with Header Auth using your Ghost Genius API key Set up your Google Sheets database using the provided template with columns: URL, Firstname, Lastname, Tagline, Location Connections, Followers, Hiring?, Open to work? Summary, Languages, Skills Experience 1-5, Education 1-3 Configure Google Sheets OAuth2 credentials following n8n's authentication setup guide Add LinkedIn profile URLs to the first column of your Google Sheet to begin enrichment Test the workflow with a small batch before processing larger lists How to customize this workflow Adjust batch processing settings** to handle larger volumes by modifying the batch size and interval timing Add data validation rules** to filter out incomplete or invalid profiles before processing Integrate with CRM systems** like HubSpot or Salesforce to automatically sync enriched data Set up automated scheduling** to regularly re-enrich profiles and capture profile updates Add email notifications** to alert when enrichment batches are completed or encounter errors Customize data mapping** to include additional profile fields or reorganize the output structure Add duplicate detection** to prevent re-processing the same profiles multiple times
by Stefan
Automate LinkedIn engagement without sounding like a bot. This workflow: 🌍 Detects language & tone (German / English) 👍 Chooses the right reaction (like / celebrate / support …) 🗣 Generates a personalised comment in your voice and mentions the author 📲 Optional Telegram review – approve ✅ or regenerate ❌ before posting 💸 Runs on cost-efficient GPT-4o mini or Claude 3.5 Haiku ☁️ Publishes comment + reaction via the Unipile API Setup (≈ 15-30 min) Unipile – connect LinkedIn → copy account_id, dsn, then create an Access-Token (X-API-KEY). Telegram (optional) – create a bot, add a credential named YOUR TELEGRAM ACCOUNT. OpenAI / Anthropic – add your API key and keep one LLM node (delete the other). Open the “Defining guardrails” node and replace the credential placeholders. (Optional) Tweak role, comment_length and openers_example_1-3 for your brand voice. Security: no live keys included – all secrets are placeholders. Best for: solopreneurs, marketing teams, personal-branding consultants.
by Alex
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. How It Works This template orchestrates a multi-step workflow that constructs a comprehensive four-zone automation matrix—Green, Yellow, Red, and White—grounded in the Human Agency Scale (HAS). When a user sends a job title via Telegram, the workflow routes both text and voice messages appropriately. Voice messages are transcribed via OpenAI's Whisper, while text inputs bypass transcription. Both streams merge into a single data flow. The AI Agent node, powered by GPT-4, analyzes the user's profession and core tasks. It also leverages live context by calling the Tavily search tool, ensuring the analysis incorporates up-to-date information. After the evaluation, the workflow formats and returns the completed matrix, with detailed task examples and rationales for each zone, back to the user via Telegram. Setup Instructions Create an OpenAI credential in n8n (model: GPT-4.1 mini). Add a Tavily credential with your API key (FREE plan available). Configure a Telegram Bot credential: API bot token. Import this JSON as a new workflow in n8n and map credentials in each node. Activate the workflow; test by sending sample job titles; adjust node timeouts and webhook settings as needed. Requirements n8n v1.0.0 or higher Active OpenAI API key (GPT-4.1 mini access) Tavily API key for web context search Telegram Bot token with correctly configured webhook Stable internet connectivity Audience & Problem This template is designed for consultants, HR professionals, and analysts who need a scalable, standardized approach to evaluate which routine tasks in a given profession can be automated, which require human oversight, and which should remain manual to preserve strategic judgment, creativity, and expertise.
by Emmanuel Bernard
This workflow illustrates how to use Perplexity AI in your n8n workflow. Perplexity is a free AI-powered answer engine that provides accurate, trusted, and real-time answers to any question. Credentials Setup 1/ Go to the perplexity dashboard, purchase some credits and create an API Key https://www.perplexity.ai/settings/api 2/ In the perplexity Request node, use Generic Credentials, Header Auth. For the name, use the value "Authorization" And for the value "Bearer pplx-e4...59ea" (Your Perplexity Api Key) AI Model Sonar Pro is the current top model used by perplexity. If you want to use a different one, check this page: https://docs.perplexity.ai/guides/model-cards
by Yaron Been
Automated monitoring system that tracks startup activities, funding events, and company updates in real-time, providing valuable market intelligence. 🚀 What It Does Real-time monitoring of startup activities Funding alerts and updates Competitor tracking Industry trend analysis Customizable watchlists 🎯 Perfect For Venture capitalists Startup founders Business development teams Market researchers Investment analysts ⚙️ Key Benefits ✅ Stay ahead of market movements ✅ Never miss important funding rounds ✅ Track competitor activities ✅ Identify emerging trends ✅ Save hours of manual research 🔧 What You Need Crunchbase API access n8n instance Notification preferences (email/Slack/Teams) 📊 Data Points Tracked New funding rounds Company updates Leadership changes Product launches Market expansions 🛠️ Setup & Support Quick Setup Deploy in 20 minutes with our step-by-step configuration guide 📺 Watch Tutorial 💼 Get Expert Support 📧 Direct Help Stay informed about the startup ecosystem with automated monitoring and alerts. Make data-driven decisions with timely, relevant information.
by PollupAI
Who is this for? This workflow is designed for professionals and teams who need to monitor multiple RSS feeds, filter the latest content, and distribute actionable updates as a Trello comment. Ideal for content managers, marketers, and team leads managing news or content pipelines. What problem is this workflow solving? Manually monitoring RSS feeds and keeping track of the latest content can be time-consuming. This workflow automates the aggregation, filtering, and distribution of news, ensuring that only relevant and timely updates are shared with your team or audience. What this workflow does: Aggregates RSS Feeds: Pulls data from up to three RSS feeds simultaneously. Filters Content: Filters articles based on their publication date (default: last 7 days). Organizes and Sorts: Sorts filtered articles by date for clarity. Formats Updates: Transforms news items into Markdown format for better readability. Publishes and Notifies: Posts comments to Trello cards and sends an email to a moderator to check the comment. Setup: Connect your RSS feeds by configuring the RSS Read nodes. Link your Trello and Gmail accounts for seamless integration. Adjust the schedule trigger to set how often the workflow should run (e.g., daily, weekly). Test the workflow to ensure all connections and configurations are correct. How to customize this workflow to your needs: Change the Number of RSS Feeds: Add or remove RSS Read nodes and update the merge configuration accordingly. Adjust the Date Filter: Modify the date logic in the “Filter by date” node to include more or fewer days. Limit the Number of Articles: Adjust the limit in the “Limit news to x” node. Custom Formatting: Update the Transform node to format the news items differently. Alternative Notifications: Replace Trello and Gmail with other integrations, such as Slack or Microsoft Teams. This workflow ensures your team stays informed with minimal effort and delivers content updates in an organized and professional manner.
by Yaron Been
This cutting-edge n8n automation is a sophisticated video intelligence tool designed to transform raw video content into actionable insights. By intelligently connecting Google Drive, AI analysis, and automated processing, this workflow: Discovers Video Content: Automatically retrieves videos from Google Drive Supports scheduled or on-demand analysis Eliminates manual content searching Advanced AI Analysis: Leverages Google Gemini AI Provides comprehensive video insights Extracts meaningful content summaries Intelligent Processing: Validates file status Prepares content for AI analysis Ensures high-quality insight generation Seamless Workflow Integration: Automated scheduling Cross-platform content processing Reduces manual intervention Key Benefits 🤖 Full Automation: Zero-touch video intelligence 💡 AI-Powered Insights: Advanced content analysis 📊 Comprehensive Processing: Detailed video understanding 🌐 Multi-Platform Synchronization: Seamless content flow Workflow Architecture 🔹 Stage 1: Content Discovery Scheduled Trigger**: Automated workflow initiation Google Drive Integration**: Video file retrieval Intelligent File Selection**: Identifies target videos Prepares for AI analysis 🔹 Stage 2: Content Preparation File Download** LLM Chain Processing** AI-Ready Content Formatting** 🔹 Stage 3: AI Analysis Gemini API Integration** Comprehensive Content Examination** Intelligent Insight Generation** 🔹 Stage 4: Result Structuring Analysis Result Formatting** Structured Insight Preparation** Ready-to-Use Intelligence** Potential Use Cases Content Creators**: Video content analysis Marketing Teams**: Content insight generation Educational Institutions**: Lecture and presentation review Research Organizations**: Automated video intelligence Media Companies**: Rapid content assessment Setup Requirements Google Drive Connected Google account Configured video folder Appropriate sharing settings Google Gemini API API credentials Configured analysis parameters Access to Gemini Pro model n8n Installation Cloud or self-hosted instance Workflow configuration API credential management Future Enhancement Suggestions 🤖 Multi-model AI analysis 📊 Detailed insight scoring 🔔 Automated reporting 🌐 Cross-platform insight sharing 🧠 Advanced content categorization Technical Considerations Implement robust error handling Use secure API authentication Maintain flexible content processing Ensure compliance with AI usage guidelines Ethical Guidelines Respect content privacy Maintain transparent analysis practices Ensure appropriate content usage Protect intellectual property rights Hashtag Performance Boost 🚀 #AIVideoAnalysis #ContentIntelligence #GeminiAI #VideoInsights #AutomatedLearning #AIWorkflow #MachineLearning #ContentAnalytics #TechInnovation #AIAutomation Workflow Visualization [Schedule Trigger] ⬇️ [Download from Drive] ⬇️ [LLM Chain Processing] ⬇️ [Check File Status] ⬇️ [Analyze Video] ⬇️ [Format Analysis Result] Connect With Me Ready to revolutionize your video intelligence? 📧 Email: Yaron@nofluff.online 🎥 YouTube: @YaronBeen 💼 LinkedIn: Yaron Been Transform your video content analysis with intelligent, automated solutions!
by bangank36
This workflow backup Squarespace website header and footer injections into Github How It Works The Squarespace injections are fetched when an URL is placed Setup Instructions First, edit HTTP Request's URL to put your Squarespace site URL there Next, to configure the Github, update the Globals node with the following values: repo.owner – Your GitHub username repo.name – The name of your GitHub repository storing the workflows repo.path – The folder path within the repository where workflows are stored For example, if your GitHub username is john-doe, your repository is named n8n-backups, and injections are stored in a squarespace-backup/ folder, you would set: repo.owner → john-doe repo.name → n8n-backups repo.path → squarespace-backup/ Each site's injections will be added into seperate folder Required Credentials GitHub API – Access to your repository Who Is This For? This template is made for Squarespace users who want to backup their header and footer injections at interval to or on demand Check out my other templates: 👉 My n8n Templates