by Geekaz / Kazen
Who is this for? This template is designed for social media managers, content creators, data analysts, and anyone who wants to automatically save and analyze their Meta Threads posts in Notion. It’s particularly useful for: Building a personal archive of your Threads content. Training AI models using your social media data. Tracking your online presence and engagement. What this workflow does This workflow uses the Meta Threads API to automatically retrieve your posts and import them into a Notion database. It retrieves the post content, date, and time, and stores them in designated properties within your Notion database. Setup Get Threads Access Token and ID: Obtain a long-lived access token and your Threads ID from the Meta Threads developer platform. This token auto-refreshes, ensuring uninterrupted workflow operation. Configure Credentials and Date Range: In the “Set Credentials” node (using edit fields), enter your token and ID. Set the since and until parameters in the “Set Date Range” node to specify the post import period. Connect to Notion and Create a Database: Connect to your Notion workspace and create a database with these properties (customize with the “Create Properties” node): a. Title: Threads post URL (Notion entry title). b. Threads ID: Unique post ID (prevents duplicate imports). c. Username: Post author (for future multi-account/source management). d. Post Date: Original post date. e. Source (Multi-Select): “Threads” tag (for future multi-platform import and filtering). f. Created: Import date and time. g. Import Check (Optional): For use with a separate post-categorization workflow.
by David Ashby
🛠️ Pushover Tool MCP Server Complete MCP server exposing all Pushover Tool operations to AI agents. Zero configuration needed - 1 operation pre-built. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works • MCP Trigger: Serves as your server endpoint for AI agent requests • Tool Nodes: Pre-configured for every Pushover Tool operation • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Uses official n8n Pushover Tool tool with full error handling 📋 Available Operations (1 total) Every possible Pushover Tool operation is included: 💬 Message (1 operations) • Push a message 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Resource IDs and identifiers • Search queries and filters • Content and data payloads • Configuration options Response Format: Native Pushover Tool API responses with full data structure Error Handling: Built-in n8n error management and retry logic 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • Other n8n Workflows: Call MCP tools from any workflow • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Complete Coverage: Every Pushover Tool operation available • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n error handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by Matthieu
LinkedIn Profile Tracker Automation Who is this for? This template is ideal for sales teams, recruiters, business development professionals, and relationship managers who need to monitor changes in their network's LinkedIn profiles. Perfect for agencies tracking client personnel changes, HR teams monitoring talent movements, sales professionals staying updated on prospect job changes, and content teams tracking influencer activity. What problem does this workflow solve? Manually checking LinkedIn profiles for updates like job changes, status modifications, profile edits, or latest posts is extremely time-consuming and easy to miss. This automation eliminates the need for constant manual monitoring while ensuring you never miss important changes that could signal new business opportunities, relationship updates, or content engagement opportunities. What this workflow does This workflow automatically monitors a list of LinkedIn profiles on a weekly schedule, detects any changes in: Personal information** (name, headline, summary) Job status** (hiring/open to work flags) Latest work experience** (new positions, company changes) Recent posts** (latest content activity) When changes are detected, it immediately sends Slack notifications with before/after comparisons and updates your tracking database to maintain historical records of all profile evolution. Setup Create a Ghost Genius API account and get your API key for LinkedIn profile scraping Configure HTTP Request nodes with Header Auth credentials using your Ghost Genius API key Set up your Google Sheets database with columns: Firstname, Lastname, LinkedIn URL, ID Tagline, Summary, Latest experience Open to work?, Hiring?, Latest post Configure Slack webhook integration for real-time notifications Set up credentials for Google Sheets and Slack following n8n documentation Add LinkedIn profile URLs to your Google Sheet to start monitoring How to customize this workflow Modify the schedule trigger** to check profiles daily, bi-weekly, or monthly based on your monitoring needs Customize Slack notification messages** to include additional context, mentions, or custom formatting Add email notifications** alongside Slack alerts for critical changes like job transitions Set up filtered notifications** to only alert on specific types of changes (e.g., job changes only, posts from key influencers) Add post content analysis** to detect mentions of your company or competitors Integrate with CRM systems** to automatically update lead records when profile changes occur
by Todsaporn Sangboon
📈 How it works This n8n workflow allows you to interact with Binance Spot Trading API directly to: Place Limit Buy and Limit Sell orders Place Market Buy and Market Sell orders Query account info* and *open orders** Cancel all open orders** for a specific symbol All requests are signed using Binance's HMAC SHA256 signature method for secure trading. ⚙️ Setup Steps Create Binance API Credentials in n8n: Go to Credentials > New Choose Binance API Add api_key and api_secret Save as Binance API Import this workflow into your n8n instance. Update default values: In Set Parameter nodes like LimitBuy Parameter, change: symbol (e.g. BTCUSDT) quantity, price as needed Run the workflow manually via the Execute workflow trigger. ✅ Notes Credential node is marked with instructions. HMAC signatures are automatically calculated before making each request. HTTP nodes are preconfigured for Binance API v3. 🔒 No API key or secret is included.
by n8n Team
This workflow let's a bot in Slack notify a specific channel when a new product in WooCommerce is published and live on the site. Prerequisites WooCommerce account Slack and a Slack bot How it works Listen for WooCommerce product creation If permalink starts with https://[your-url-here].com/product/ Slack bot notifies channel that a new product has been added. Please note, you must update the URL in the IF node to match your url. If your WooCommerce doesn't use the slug /product/, that will need to be updated too.
by Lucas
🎶 Add liked songs to a monthly playlist > This Workflow is a port of Add saved songs to a monthly playlist from IFTTT. When you like a song, the workflow will save this song in a monthly playlist. E.g.: It's June 2024, I liked a song. The workflow will save this song in a playlist called June '24. If this playlist does not exist, the workflow will create it for me. ⚙ How it works Each 5 minutes, the workflow will start automatically. He will do 3 things : Get the last 10 songs you saved in the "Liked song" playlist (by clicking on the heart in the app) and save them in a NocoDB table (of course, the workflow avoid to create duplicates). Check if the monthly playlist is already created. Otherwise, the playlist is created. The created playlist is also saved in NocoDB to avoid any problems. Check if the monthly playlist contains all the song liked this month by getting them from NocoDB. If they are not present, add them one by one in the playlist. You may have a question regarding the need of NocoDB. Over the last few weeks/months, I've had duplication problems in my playlists and some playlists have been created twice because Spotify wasn't returning all the information but only partial information. Having the database means I don't have to rely on Spotify's data but on my own, which is accurate and represents reality. 📝 Prerequisites You need to have : Spotify API keys, which you can obtain by creating a Spotify application here: https://developer.spotify.com/dashboard. Create a NocoDB API token 📚 Instructions Follow the instructions below Create your Spotify API credential Create your NocoDB credential Populate all Spotify nodes with your credentials Populate all Spotify nodes with your credentials Enjoy ! If you need help, feel free to ping me on the N8N Discord server or send me a DM at "LucasAlt" Show your support Share your workflow on X and mention @LucasCtrlAlt Consider buying me a coffee 😉
by Mark Shcherbakov
Video Guide I prepared a detailed guide that shows the whole process of building an AI tool to analyze Instagram Reels using n8n. Youtube Link Who is this for? This workflow is ideal for social media analysts, digital marketers, and content creators who want to leverage data-driven insights from their Instagram Reels. It's particularly useful for those looking to automate the analysis of video performance to inform strategy and content creation. What problem does this workflow solve? Analyzing video performance on Instagram can be tedious and time-consuming, requiring multiple steps and data extraction. This workflow automates the process of fetching, analyzing, and recording insights from Instagram Reels, making it simpler for users to track engagement metrics without manual intervention. What this workflow does This workflow integrates several services to analyze Instagram Reels, allowing users to: Automatically fetch recent Reels from specified creators. Analyze the most-watched videos for insights. Store and manage data in Airtable for easy access and reporting. Initial Trigger: The process begins with a manual trigger that can later be modified for scheduled automation. Data Retrieval: It connects to Airtable to fetch a list of creators and their respective Instagram Reels. Video Analysis: It handles the fetching, downloading, and uploading of videos for analysis using an external service, simplifying performance tracking through a structured query process. Record Management: It saves relevant metrics and insights into Airtable, ensuring that users can access and organize their video analytics effectively. Setup Create accounts: Set up Airtable, Edify, n8n, and Gemini accounts. Prepare triggers and modules: Replace credentials in each node accordingly. Configure data flow: Ensure modules are set to fetch and analyze the correct data fields as outlined in the guide. Test the workflow: Run the scenario manually to confirm that data is fetched and analyzed correctly.
by Lucas Peyrin
How it works This template is an interactive playground designed to help you master the most useful keyboard shortcuts in n8n and supercharge your building speed. Forget boring lists—this workflow gives you hands-on tasks to complete, turning learning into a practical exercise. The workflow is structured into four chapters, each focusing on a different aspect of workflow development: Node Basics: Learn the fundamentals of interacting with a single node, such as renaming, editing, duplicating, and deactivating. Canvas Navigation & Selection: Master the art of moving around the canvas and selecting multiple nodes efficiently. Advanced Actions: Discover powerful moves like tidying up messy connections and creating sub-workflows. Execution & Debugging: Uncover essential shortcuts for testing your workflows, like pinning data and navigating the executions panel. Each step provides a clear task in a sticky note, guiding you to perform the action yourself. Set up steps Setup time: 0 minutes! This workflow is a self-contained tutorial and requires no setup, credentials, or configuration. Open the workflow. Follow the instructions in the sticky notes, starting from the top. Perform the actions as described to build muscle memory for each shortcut. That's it! Get ready to become an n8n power user.
by Lucas Peyrin
How it works This template is an interactive, step-by-step tutorial designed to teach you the most important skill in n8n: using expressions to access and manipulate data. If you know what JSON is but aren't sure how to pull a specific piece of information from one node and use it in another, this workflow is for you. It starts with a single "Source Data" node that acts as our filing cabinet, and then walks you through a series of lessons, each demonstrating a new technique for retrieving and transforming that data. You will learn how to: Access a simple value from a previous node. Use n8n's built-in selectors like .last() and .first(). Get a specific item from a list (Array). Drill down into nested data (Objects). Combine these techniques to access data in an array of objects. Go beyond simple retrieval by using JavaScript functions to do math or change text. Inspect data with utility functions like Object.keys() and JSON.stringify(). Summarize data from multiple items using .all() and arrow functions. Set up steps Setup time: 0 minutes! This workflow is a self-contained tutorial and requires no setup or external credentials. Click "Execute Workflow" to run the entire tutorial. Follow the flow from the "Source Data" node to the "Final Exam" node. For each lesson, click on the node to see how its expressions are configured in the parameters panel. Read the detailed sticky note next to each lesson—it breaks down exactly how the expression works and why. By the end, you'll have the foundational knowledge to connect data and build powerful, dynamic workflows in n8n.
by David Olusola
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. MCP Gmail Workflow – AI-Powered Email Management ✨ What It Does A smart n8n workflow that connects Gmail with an AI agent (via MCP), letting you send, read, and organize emails using natural language. ⚙️ Key Features 🧠 AI Commands: “Send email to John about the budget” 📥 Inbox Control: Mark read/unread, apply/remove labels 🗂 Smart Organization: Auto-label based on content 🤖 MCP-Ready: Works with Claude, ChatGPT, etc. 🎯 Use Cases “📤 Send a follow-up to the client about yesterday’s meeting” “📬 Mark all newsletters as read and label ‘Newsletter’” “🧾 Summarize latest email from Sarah” “🗃 Label all Project X emails as ‘Project-X-2024’” “⭐ Find unread emails from my manager and mark as important” 🛠 Setup Guide 🔑 Prerequisites n8n (self-hosted or cloud) Gmail API credentials MCP-compatible AI (optional but powerful) 📥 1. Import Workflow Copy JSON → Open n8n → Import → Paste → Done ✅ 🔐 2. Gmail OAuth2 Setup Create Google project → Enable Gmail API Create OAuth2 creds → Add n8n redirect URI In n8n: Add Gmail OAuth2 → Paste Client ID/Secret → Connect 🧩 3. Update Credential References Find your credential ID in n8n Update each Gmail node with your ID 🧠 4. MCP Trigger (Optional) Use provided webhook URL in your AI system Send test prompts to verify connection 🧪 5. Test Key Actions ✅ “Send a test email” ✅ “Read latest email” ✅ “Label last email as ‘Test’” ✅ “Mark latest email as unread” ⚙️ 6. Advanced Tips Create custom labels in Gmail Use HTTPS + webhook auth Add retries and error handling in n8n 🧯 Troubleshooting ❗ Gmail Auth Error? → Re-auth and check redirect URI ❗ Webhook not firing? → Check endpoint + manual test ❗ Label errors? → Use correct label names or IDs ✅ Required Gmail Scopes: gmail.modify gmail.send 📈 Best Practices 🔁 Test regularly 🔒 Use minimal permissions 🏷 Consistent label naming 🔍 Monitor execution + webhook logs 🎉 You’re All Set! Control Gmail with your voice or text through AI. Make managing emails smarter, faster, and 100% automated 💌
by David Ashby
Complete MCP server exposing 4 AWS Cost and Usage Report Service API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add AWS Cost and Usage Report Service credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the AWS Cost and Usage Report Service API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to http://cur.{region}.amazonaws.com • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (4 total) 🔧 #X-Amz-Target=Awsorigamiservicegatewayservice.Deletereportdefinition (1 endpoints) • POST /#X-Amz-Target=AWSOrigamiServiceGatewayService.DeleteReportDefinition: Deletes the specified report. 🔧 #X-Amz-Target=Awsorigamiservicegatewayservice.Describereportdefinitions (1 endpoints) • POST /#X-Amz-Target=AWSOrigamiServiceGatewayService.DescribeReportDefinitions: Lists the AWS Cost and Usage reports available to this account. 🔧 #X-Amz-Target=Awsorigamiservicegatewayservice.Modifyreportdefinition (1 endpoints) • POST /#X-Amz-Target=AWSOrigamiServiceGatewayService.ModifyReportDefinition: Allows you to programatically update your report preferences. 🔧 #X-Amz-Target=Awsorigamiservicegatewayservice.Putreportdefinition (1 endpoints) • POST /#X-Amz-Target=AWSOrigamiServiceGatewayService.PutReportDefinition: Creates a new report using the description that you provide. 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native AWS Cost and Usage Report Service API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by David Ashby
Complete MCP server exposing 8 Bulk WHOIS API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add Bulk WHOIS API credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the Bulk WHOIS API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to http://localhost:5000 • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (8 total) 🔧 Batch (4 endpoints) • GET /batch: Get your batches • POST /batch: Create batch. Batch is then being processed until all provided items have been completed. At any time it can be get to provide current status with results optionally. • DELETE /batch/{id}: Delete batch • GET /batch/{id}: Get batch 🔧 Db (1 endpoints) • GET /db: Query domain database 🔧 Domains (3 endpoints) • GET /domains/{domain}/check: Check domain availability • GET /domains/{domain}/rank: Check domain rank (authority). • GET /domains/{domain}/whois: WHOIS query for a domain 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native Bulk WHOIS API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.