by Mark Shcherbakov
Video Guide I prepared a detailed guide that showed the whole process of integrating the Binance API and storing data in Airtable to manage funding statements associated with tokens in a wallet. Youtube Link Who is this for? This workflow is ideal for developers, financial analysts, and cryptocurrency enthusiasts who want to automate the process of managing funding statements and token prices. It’s particularly useful for those who need a systematic approach to track and report funding fees associated with tokens in their wallets. What problem does this workflow solve? Managing funding statements and token prices across multiple platforms can be cumbersome and error-prone. This workflow automates the process, allowing users to seamlessly fetch funding fees from Binance and record them alongside token prices in Airtable, minimizing manual data entry and potential discrepancies. What this workflow does This workflow integrates the Binance API with an Airtable database, facilitating the storage and management of funding statements linked to tokens in a wallet. The agent can: Fetch funding fees and current positions from Binance. Aggregate data to create structured funding statements. Insert records into Airtable, ensuring proper linkage between funding data and tokens. API Authentication: The workflow establishes authentication with the Binance API using a Crypto Node to handle API keys and signatures, ensuring secure and verified requests. Data Collection: It retrieves necessary data, including funding fees and current positions with properly formatted API requests to ensure seamless communication with Binance. Airtable Integration: The workflow inserts aggregated funding statements and token data into the corresponding Airtable records, managing token existence checks to avoid duplicate entries. Setup Set Up Airtable Database: Create an Airtable base with tables for Funding Statements and Tokens. Generate Binance API Key: Log in and create an API key with appropriate permissions. Set Up Authentication in N8N: Utilize a Crypto Node for Binance API authentication. Configure API Request to Binance: Set request method and headers for communication with the Binance API. Fetch Funding Fees and Current Positions: Retrieve funding data and current positions efficiently. Aggregate and Create Statements: Aggregate data to create detailed funding statements. Insert Data into Airtable: Input the structured data into Airtable and manage token records. Using Get Price Node: Implement a Get Price Node to maintain current token price tracking without additional setup.
by Zacharia Kimotho
Workflow documentation updated on 21 May 2025 This workflow keeps track of your brand mentions across different Facebook groups and provides an analysis of the posts as positive, negative or neutral and updates this to Googe sheets for further analysis This is useful and relevants for brands looking to keep track of what people are saying about their brands and guage the customer satisfaction or disatisfaction based on what they are talking about Who is this template for? This workflow is for you if You Need to keep track of your brand sentiments across different niche facebook groups Own a saas and want to monitor it across different local facebook Groups Are looking to do some competitor research to understand what others dont like about their products Are testing the market on different market offerings and products to get best results Are looking for sources other that review sites for product, software or service reviews Need to keep track of your brand sentiments across different niche facebook groups Are starting on market research and would like to get insights from differnt facebook groups on app usage, strngths weaknesses, features etc How it works You will set the desired schedule by which to monitor the groups This gets the brand names and facebook Groups to monitor. Setup Steps Before you begin You will need access to a Bright Data API to run this workflows Make a copy of the sheet below and add the urls for the facebook groups to scrap and the brand names you wish to monitor. Import the workflow json to your canvas Make a copy of this Google sheet to get started easily Set your APi key in the Map out the Google sheet to your tables You can use/update the current AI models to differnt models eg Gemini or anthropic Run the workflow Setup B Bright Data provides an option to receive the results on an external webhook via a POST call. This can be collected via the
by Sebastian/OptiLever
SWOT Analysis Generator That Produces PDF Reports In n8n Want to skip the manual work and instantly generate SWOT analyses for your business plans, investor decks, or strategy docs? 🚀 This workflow lets you automate the entire SWOT (Strengths, Weaknesses, Opportunities, Threats) process—using AI and no-code tools! Whether you're using this for yourself or creating documents for clients, this workflow will save you hours of time and help identify the key components of growing any business. Tools We’ll Be Using: OpenAI API**: Powers AI-driven extraction and generation of SWOT sections, introductions, conclusions, and more. n8n**: The no-code automation platform that orchestrates this workflow. Google Sheets**: Stores your company data and report outputs (download the provided "SWOT Analysis" template and fill it out). APITemplate.io**: Converts HTML reports into downloadable, multi-page PDFs. DeepSeek API* *(optional): An alternative to OpenAI for specific reasoning tasks. Gmail OAuth2**: Sends the final report via email. If you’re building a business, doing client work, or just want faster strategic planning—this will save you HOURS. How It Works This n8n workflow transforms structured company data from a Google Sheet into a fully formatted, investor-grade SWOT report in both HTML and PDF formats. Here’s the process in a few high-level steps: Load Data: Pulls company information from the "Company Info Input" tab in the "SWOT Analysis" Google Sheet. AI Analysis: Uses AI (via OpenAI or DeepSeek) to categorize data into Strengths, Weaknesses, Opportunities, and Threats, and generates detailed narrative sections for each. Create Report Sections: Produces a strategic introduction, conclusion, table of contents, and title page using AI. Format and Combine: Converts all sections into styled HTML and merges them into a single document. Generate PDF: Converts the HTML into a polished, multi-page PDF using APITemplate.io. Save and Share: Saves the report back to Google Sheets and emails the PDF to a specified recipient. The result is a professional, investor-ready SWOT report with minimal manual effort, ready for sharing or review. Setup Steps Setting up this workflow typically takes 15-30 minutes, depending on your familiarity with the tools. Here’s what you’ll need to do: Configure Credentials**: Set up API keys and OAuth2 credentials for Google Sheets, OpenAI, APITemplate.io, Gmail, and optionally DeepSeek. Prepare Your Google Sheet**: Download the provided "SWOT Analysis" template, fill out the "Company Info Input" tab with your company data, and ensure it’s accessible. Import the Workflow**: Copy the workflow JSON into your n8n instance and connect your credentials. Detailed instructions for each step, including credential setup and customization options, are available in the sticky notes within the workflow. Refer to these for step-by-step guidance. Benefits Efficiency**: Automates the entire SWOT report generation process, saving hours of manual work. Professional Output**: Delivers fully styled HTML and PDF reports with headings, paragraphs, and custom layouts. AI-Driven Insights**: Ensures clarity and relevance in every section through targeted AI analysis. Seamless Sharing**: Automatically saves reports to Google Sheets and emails them to stakeholders. This workflow is perfect for entrepreneurs, consultants, or strategists looking to streamline their SWOT analysis process and deliver high-quality, professional reports effortlessly.
by David Ashby
Complete MCP server exposing 1 IP2WHOIS Domain Lookup API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add IP2WHOIS Domain Lookup credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the IP2WHOIS Domain Lookup API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to https://api.ip2whois.com/v2 • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (1 total) 🔧 General (1 endpoints) • GET /: Lookup WHOIS Data 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native IP2WHOIS Domain Lookup API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by Simeon
🔄 Reddit Content Operations via MCP Server 🧑💼 Who is this for? This workflow is built for content creators, marketers, Reddit automation enthusiasts, and AI agent developers who want structured, programmable access to Reddit content. If you're researching niche communities, tracking trends, or automating Reddit engagement — this is for you. 💡 What problem is this workflow solving? Reddit has valuable content scattered across subreddits, but manual analysis or engagement is inefficient. This workflow acts as a centralized API interface to: Query and manage Reddit posts Create, fetch, delete, and reply to comments Analyze subreddit metadata and behavior Enable AI agents to autonomously operate on Reddit data It does this using an MCP (Model Context Protocol) Server over Server-Sent Events (SSE). ⚙️ What this workflow does This template sets up a custom MCP Server that listens for JSON-based operation commands sent via SSE. Based on the operation, it routes the request to one of the following branches: 🟥 Post CRUD Create a new Reddit post Search posts across subreddits Fetch posts by ID Delete existing posts 🟩 Comment CRUD Create or reply to comments Fetch multiple comments from posts Delete specific comments 🟦 Subreddit Read Operations Get information about subreddits List subreddit posts Retrieve subreddit rules 🛠 Setup Import this workflow into your self-hosted n8n instance. Configure Reddit credentials (OAuth2). Connect your input system to the MCP Server Trigger node via SSE. Send operation payloads to the server like this: { "operation": "post_search", "params": { "query": "AI agents", "subreddit": "machinelearning" } } The workflow will route to the appropriate node based on operation type. 🧩 Supported Operations post_create post_get_many post_search post_delete post_get_by_id comment_create comment_reply comment_get_many comment_delete subreddit_get_about subreddit_get_many subreddit_get_rules 🧠 How to customize this workflow to your needs Add new operations to the operation_switch node for additional API functionality. Chain results into Notion, Slack, Airtable, or external APIs. Integrate with OpenAI/GPT to summarize posts or filter content. Add logic to score and sort content by engagement, sentiment, or keywords. 🟨 Sticky Notes Each operation group is color-coded (Posts, Comments, Subreddits). Sticky Notes explain the purpose and dependencies of each section. Easy to maintain and extend with clear logical separation. ⚠️ This template uses a custom MCP Server node and only works in self-hosted n8n. 🖼 Workflow Preview
by David Ashby
Complete MCP server exposing 1 Recommendation API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add Recommendation API credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the Recommendation API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to https://api.ebay.com{basePath} • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (1 total) 🔧 Find (1 endpoints) • POST /find: Get Promoted Listings Recommendations 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native Recommendation API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by David Ashby
Complete MCP server exposing 1 Buy Marketing API operations to AI agents. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Credentials Add Buy Marketing API credentials Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works This workflow converts the Buy Marketing API into an MCP-compatible interface for AI agents. • MCP Trigger: Serves as your server endpoint for AI agent requests • HTTP Request Nodes: Handle API calls to https://api.ebay.com/buy/marketing/v1_beta • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Returns responses directly to the AI agent 📋 Available Operations (1 total) 🔧 Merchandised_Product (1 endpoints) • GET /merchandised_product: Fetch Merchandised Products 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Path parameters and identifiers • Query parameters and filters • Request body data • Headers and authentication Response Format: Native Buy Marketing API responses with full data structure Error Handling: Built-in n8n HTTP request error management 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Cursor: Add MCP server SSE URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n HTTP request handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by Alfonso Corretti
Who is this for? 🧑🏻🫱🏻🫲🏻🤖 Humans and Robots alike. This workflow can be used as a Chat Trigger, as well as a Workflow Trigger. It will take a natural language request, and then generate a SQL query. The resulting query parameter will contain the query, and a sqloutput parameter will contain the results of executing such query. What's the use case? This template is most useful paired with other workflows that extract e-mail information and store it in a structured Postgres table, and use LLMs to understand inquiries about information contained in an e-mail inbox and formulate questions that needs answering. Plus, the prompt can be easily adapted to formulate SQL queries over any kind of structured database. Privacy and Economics As LLM provider I'm using Ollama locally, as I consider my e-mail extremely sensitive information. As model, phi4-mini does an excellent job balancing quality and efficiency. Setup Upon running for the first time, this workflow will automatically trigger a sub-section to read all tables and extract their schema into a local file. Then, either by chatting with the workflow in n8n's interface or by using it as a sub-workflow, you will get a query and a sqloutput response. Customizations If you want to work with just one particular table yet keep edits at bay, append a condition to the List all tables in a database step, like so: WHERE table_schema='public' AND table_name='my_emails_table_name' To repurpose this workflow to work with any other data corpus in a structured database, inspect the AI Agent user and system prompts and edit them accordingly.
by Huzaifa Tahir
🎬 What it does This workflow creates an engaging YouTube Short with a single click — from script to voiceover, to visuals and background music. It combines several AI tools to automate content creation and final video assembly. ⚙️ How it works Accepts an input prompt or topic Generates script using GPT Converts script to voiceover using ElevenLabs Generates b-roll style images via Leonardo.Ai Matches background music Assembles a vertical 1080×1920 MP4 video using JSON render config Optionally uploads to YouTube or saves to Cloudinary 🧰 Setup steps Add your credentials: Leonardo API (image generation) ElevenLabs (voiceover) Cloudinary (upload destination) Any GPT-based text generator Drop your audio/music file in the right node Replace API expressions with your own credentials > 🟨 Full step-by-step instructions are in sticky notes inside the workflow.
by Alex Gurinovich
AI powered Automated Crypto Insights with Chart-img and BrowserAI Tired of paying for costly crypto updates? Or reading long analyses? This n8n workflow automates the delivery of personalized crypto insights, using Chart-img for capturing coin graphs of BTC, ETH, SOL, and XRP as base64 images, and BrowserAI for web scraping and information gathering of news and articles. This setup ensures thorough market coverage and timely updates, without breaking the bank. Overview Designed for crypto enthusiasts, traders, and analysts, this workflow automates the process of collecting and distributing valuable crypto information. It’s perfect for anyone wanting consistent and accurate updates conveniently. Setup Instructions Pre-conditions Chart-img Account: Register for a Chart-img account and obtain an API key here. BrowserAI Account: Sign up for BrowserAI and get your API key from your BrowserAI dashboard. Step-by-Step Setup 🗓️ Schedule and Date Calculation Triggers twice daily at 8AM and 8PM to ensure up-to-date insights, and can be changed to your like. Calculates yesterday’s date dynamically for accurate data retrieval. 📊 Coin Graph Capture with Chart-img Uses Chart-img API to capture 24-hour graphs for BTC, ETH, SOL, and XRP. Converts images to base64 strings for easy integration into analysis. 🌐 Web Scraping with BrowserAI Creates tasks in BrowserAI to gather the latest crypto news and insights. Automates data extraction for comprehensive market analysis. ⌛ Monitor and Complete Tasks Incorporates status checks to ensure BrowserAI tasks complete successfully before proceeding. ✏️ Analyze and Synthesize Information Combines graph data with web-scraped insights for an enriched summary. Uses AI to generate simple, informative descriptions under 60 words to not overload you. 📩 Deliver Insights Efficiently Sends the compiled analysis to your Telegram, with easy options to switch to WhatsApp, email, or any other communication channel. Customization Guidance Content Personalization:** Customize the datasets and keywords for tailored updates. Modify Schedule:** Adjust triggering times according to your needs using n8n’s scheduling options. This workflow delivers a seamless and cost-effective approach to staying informed about crypto market trends, combining the latest technology for superior insights. ++WARNING:++ This template is intended for personal use only and does not constitute financial advice. Any actions taken using this tool are solely the user's responsibility.
by Jan Oberhauser
With this workflow you get a fully automated AI powered Support-Agent for your WooCommerce webshop. It allows customers to request information about things like: the status of their order the ordered products shipping and billing address current DHL shipping status How it works The workflow receives chat messages from an in a website integrated chat. For security and data-privacy reasons, does the website transmit the email address of the user encrypted with the requests. That ensures that user can just request the information about their own orders. An AI agent with a custom tool supplies the needed information. The tool calls a sub-workflow (in this case, in the same workflow for convenience) to retrieve the required information. This includes the full information of past orders plus the shipping information from DHL. If otherr shipping providers are used it should be simple to adjust the workflow to query information from other APIs like UPS, Fedex or others.
by Tiberiu S - Makeitfuture.com
This workflow will allow you to use OpenAI Assistant API together with a chatting platform. This version is configured to work with Hubspot, however, the Hubspot modules can be replaced by other platform and it will work similarly. Prerequisites: Create a Hubspot Chat (Live chat available on free plan) or Chatflow (paid hubspot only) and configure it to send all replies toward an n8n webhook (you need to create a custom app for that. I will create a separate article on how to do it, meanwhile, feel free to message me if you need support. Setup: Create a OpenAI Assistant, define its functionality and functions Update the Hubspot modules with the Hubspot API Key Update the OpenAI modules with OpenAI API Key Create an Airtable or any other database where you keep a reference between the thread id in Hubspot and Assistant API If you need help deploying this solution don't hesitate to email me or schedule a call here.