by Richard Uren
This template extracts all customers from shopify using GraphQL and the shopify admin API and sync them into a Baserow table. Setup Notes Update the Endpoint in GraphQL node to reflect your Shopify store. In Baserow create a shopify database with a customer table in Baserow. Create columns in the Baserow customer table for first_name, last_name, and email. It takes about 1 second per row to insert.
by Stefan
Automate LinkedIn engagement without sounding like a bot. This workflow: ๐ Detects language & tone (German / English) ๐ Chooses the right reaction (like / celebrate / support โฆ) ๐ฃ Generates a personalised comment in your voice and mentions the author ๐ฒ Optional Telegram review โ approve โ or regenerate โ before posting ๐ธ Runs on cost-efficient GPT-4o mini or Claude 3.5 Haiku โ๏ธ Publishes comment + reaction via the Unipile API Setup (โ 15-30 min) Unipile โ connect LinkedIn โ copy account_id, dsn, then create an Access-Token (X-API-KEY). Telegram (optional) โ create a bot, add a credential named YOUR TELEGRAM ACCOUNT. OpenAI / Anthropic โ add your API key and keep one LLM node (delete the other). Open the โDefining guardrailsโ node and replace the credential placeholders. (Optional) Tweak role, comment_length and openers_example_1-3 for your brand voice. Security: no live keys included โ all secrets are placeholders. Best for: solopreneurs, marketing teams, personal-branding consultants.
by David Roberts
AI evaluation in n8n This is a template for n8n's evaluation feature. Evaluation is a technique for getting confidence that your AI workflow performs reliably, by running a test dataset containing different inputs through the workflow. By calculating a metric (score) for each input, you can see where the workflow is performing well and where it isn't. How it works This template shows how to calculate a workflow evaluation metric: whether a specific tool was called by an agent. We use an evaluation trigger to read in our dataset It is wired up in parallel with the regular trigger so that the workflow can be started from either one. More info We make sure that the agent outputs the list of tools that it used We then check whether the expected tool (from the dataset) is in that list Finally we pass this information back to n8n as a metric
by Zakaria Ben
This workflow template is designed for dental assistants and anyone looking to automate appointment scheduling. It integrates Google Calendar for booking appointments and Google Sheets as a database to store patient information. How It Works The user interacts with the chatbot to schedule an appointment. The chatbot collects necessary details and checks availability via Google Calendar. If the requested time is available, the AI books the appointment. If unavailable, the AI suggests alternative time slots. Once booked, the AI logs the appointment details into Google Sheets for record-keeping. Setup Instructions ๐ Watch this ๐ฅ Setup Video for detailed instructions on running and customizing this workflow. Step 1: Set Up Credentials OpenAI API Key (for chatbot functionality). Google Account (for Google Sheets & Google Calendar integration). Step 2: Choose the Right Tools Select the correct Google Calendar in the Google Calendar tool. Choose the appropriate Google Sheets file in the Google Sheets tool. Step 3: Test Run a test to ensure everything works correctly. Once tested. Example Templates Below are sample Google Sheets template to help you get started.
by VKAPS IT
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. ๐ฏ How it works This workflow captures new lead information from a web form, enriches it with Apollo.io data, qualifies the lead using AI, andโif the lead is strongโautomatically sends a personalized outreach email via Gmail and logs the result in Google Sheets. ๐ ๏ธ Key Features ๐ฉ Lead form capture with validation ๐ Enrichment via Apollo API ๐ค Lead scoring using AI (LangChain + Groq) ๐ง Dynamic email generation & sending via Gmail ๐ Logging leads with job title & org into Google Sheets โ Conditional email sending (score โฅ 6 only) ๐งช Set up steps Estimated time: 15โ20 minutes Add your Apollo API Key to the HTTP Header credential (never hardcode!) Connect your Gmail account for sending emails Connect your Google Sheets account and set up the correct spreadsheet & sheet name Enable LangChain/Groq credentials for lead scoring and AI-generated emails Update the form endpoint to your live webhook if needed ๐ Sticky Notes Add the following mandatory sticky notes inside your workflow: FormTrigger Node: "Collects lead info via form. Ensure your form is connected to this endpoint." HTTP Request Node: "Enrich lead using Apollo.io API. Add your API key via header-based authentication." AI Agent (Lead Score): "Scores lead from 1-10 based on job title and industry match. Only leads with score โฅ 6 proceed." AI Agent (Email Composer): "Generates a concise, polite email using leadโs job title & company. Modify tone if needed." Google Sheets Append: "Logs enriched lead with job title, org, and LinkedIn URL. Customize sheet structure if needed." Gmail Node: "Sends personalized outreach email if lead passes score threshold. Uses AI-generated content." ๐ธ Free or Paid? Free โ No paid API services are required (Apollo has a free tier).
by Jimleuk
This template is for Self-Hosted N8N Instances only. This n8n demonstrates how to build a simple SQLite MCP server to perform local database operations as well as use it for Business Intelligence. This MCP example is based off an official MCP reference implementation which can be found here -https://github.com/modelcontextprotocol/servers/tree/main/src/sqlite How it works A MCP server trigger is used and connected to 5 tools: 2 Code Node and 3 Custom Workflow. The 2 Code Node tools use the SQLLite3 library and are simple read-only queries and as such, the Code Node tool can be simply used. The 3 custom workflow tools are used for select, insert and update queries as these are operations which require a bit more discretion. Whilst it may be easier to allow the agent to use raw SQL queries, we may find it a little safer to just allow for the parameters instead. The custom workflow tool allows us to define this restricted schema for tool input which we'll use to construct the SQL statement ourselves. All 3 custom workflow tools trigger the same "Execute workflow" trigger in this very template which has a switch to route the operation to the correct handler. Finally, we use our Code nodes to handle select, insert and update operations. The responses are then sent back to the the MCP client. How to use This SQLite MCP server allows any compatible MCP client to manage a SQLite database by supporting select, create and update operations. You will need to have a SQLite database available before you can use this server. Connect your MCP client by following the n8n guidelines here - https://docs.n8n.io/integrations/builtin/core-nodes/n8n-nodes-langchain.mcptrigger/#integrating-with-claude-desktop Try the following queries in your MCP client: "Please create a table to store business insights and add the following..." "what business insights do we have on current retail trends?" "Who has contributed the most business insights in the past week?" Requirements SQLite for database. MCP Client or Agent for usage such as Claude Desktop - https://claude.ai/download Customising this workflow If the scope of schemas or tables is too open, try restrict it so the MCP serves a specific purpose for business operations. eg. Confine the querying and editing to HR only tables before providing access to people in that department. Remember to set the MCP server to require credentials before going to production and sharing this MCP server with others!
by Jimleuk
This n8n demonstrates how to build your own Github MCP server to personalise it to your organisation's repositories, issues and pull requests. This n8n implementation, though not as fully featured as the official MCP server offered by Github, allows you to control precisely what access and/or functionality is granted to users which can make MCP use simpler and in some cases, more secure. The use-case in this template is to simply view and comment on issues within a specific repository but can be extended to meet the needs of your team. This MCP example is based off an official MCP reference implementation which can be found here https://github.com/modelcontextprotocol/servers/tree/main/src/github How it works A MCP server trigger is used and connected to 3 custom workflow tools. We're using custom workflow tools as there is quite a few nodes required for each task. Behind these tools are regular Github nodes although preconfigured with credentials and targeted repository. The "Get Issue Comments" and "Create Issue Comment" tools depend on obtaining an Issue Number first. The agent should call the "Get Latest Issues" tool for this. How to use This Github MCP server allows any compatible MCP client to view and comment on Github Issues. You will need to have a Github account and repository access available before you can use this server. Connect your MCP client by following the n8n guidelines here - https://docs.n8n.io/integrations/builtin/core-nodes/n8n-nodes-langchain.mcptrigger/#integrating-with-claude-desktop Try the following queries in your MCP client: "Can you get me the latest issues about MCP?" "What is the current progress on Issue 12345?" "Please can you add a comment to Issue 12345 that they should try installing the latest version and see if that works?" Requirements Github for account and repository access. The repository need not be your own but you'll still need to ensure you have the correct permissions. MCP Client or Agent for usage such as Claude Desktop - https://claude.ai/download Customising this workflow Extend this template to interactive with pull requests or workflows within your own company's Github repositories. Alternatively, pull in metrics and generate reports for programme managers. Remember to set the MCP server to require credentials before going to production and sharing this MCP server with others!
by Josh Universe
How the sequence works: A "Schedule Trigger" node activates the automation at a defined schedule. An "Airtable" node will search for previously posted questions in your question database. Airtable Base Template: here An "Aggregate" node will take all the questions from Airtable and compress them to a single output. ChatGPT, or a model of your choice, will generate a discussion question based on the options in the system prompt. The discussion question will be posted to the subreddit of your choice by the "Reddit" node. You can choose between a text, image, or link post! The recently-posted discussion question will then be uploaded to your Airtable base using the "Airtable" node. This will be used to prevent ChatGPT from creating duplicate questions. Setup Steps The setup process will take about 5 minutes. An Airtable base template is also pre-made for you here: https://airtable.com/app6wzQqegKIJOiOg/shrzy7L9yv8BFRQdY Set the recurrence in the "Schedule" node Create an Airtable account, you can use the link here. Get an Airtable personal access token here. Configure the "Get Previous Discussions" Airtable node. Configure the options in brackets in the "Generate New Discussion" node. Set the desired subreddit to post to and the post type(text, image, or link) in the "Post Discussion" node. Configure the "Create Archived Discussion" node to map to the Airtable base(required) and specific subreddit(optional).
by HoangSP
Medical Q&A Chatbot for Urology using RAG with Pinecone and GPT-4o This template provides an AI-powered Q&A assistant for the Urology domain using Retrieval-Augmented Generation (RAG). It uses Pinecone for vector search and GPT-4o for conversational responses. ๐ง Use Case This chatbot is designed for clinics or medical pages that want to automate question answering for Urology-related conditions. It uses a vector store of domain knowledge to return verified responses. ๐ง Requirements โ OpenAI API key (GPT-4o or GPT-4o-mini) โ Pinecone account with an active index โ Verified Urology documents embedded into Pinecone โ๏ธ Setup Instructions Create a Pinecone vector index and connect it using the Pinecone credentials node. Upload Urology-related documents to embed using the Create Embeddings for Urology Docs node. Customize the chatbot system message to reflect your medical specialty. Deploy this chatbot on your website or link it with Telegram via the chat trigger node. ๐ ๏ธ Components chatTrigger: Listens for user messages and starts the workflow. Medical AI Agent: GPT-based agent guided by domain-specific instructions. RAG Tool Vector Store: Fetches relevant documents from Pinecone using vector search. Memory Buffer: Maintains conversation context. Create Embeddings for Urology Docs: Encodes documents into vector format. ๐ Customization You can replace the knowledge base with any other medical domain by: Updating the documents stored in Pinecone. Modifying the system prompt in the AI Agent node. ๐ฃ CTA This chatbot is ideal for clinics, medical consultants, or educational websites wanting a reliable AI assistant in Urology.
by Jimleuk
This n8n workflow demonstrates how to create a really simple yet effective customer support channel and pipeline by combining Slack, Linear and AI tools. Built on n8n's ability to integrate anything, this workflow is intended for small support teams who want to maximise re-use of the tools they already have with an interface which is doesn't require any onboarding. Read the blog post here: https://blog.n8n.io/automated-customer-support-tickets-with-n8n-slack-linear-and-ai/ How it works The workflow is connected to a slack channel setup with the customer to capture support issues. Only messages which are tagged with a "โ " reaction are captured by the workflow. Messages are tagged by the support team in the channel. Each captured support issue is sent to the AI model to classify, prioritise and rewrite into a support ticket. The generated support ticket is uploaded to Linear for the support team to investigate and track. Support team is able to report back to the user via the channel when issue is fixed. Requirements Slack channel to be monitored Linear account and project Customising this workflow Don't have Linear? This workflow can work just as well with traditional ticketing systems like JIRA.
by Guillaume Duvernay
This template provides a fully automated system for monitoring news on any topic you choose. It leverages Linkup's AI-powered web search to find recent, relevant articles, extracts key information like the title, date, and summary, and then neatly organizes everything in an Airtable base. Stop manually searching for updates and let this workflow deliver a curated news digest directly to your own database, complete with a Slack notification to let you know when it's done. This is the perfect solution for staying informed without the repetitive work. Who is this for? Marketing & PR professionals:** Keep a close eye on industry trends, competitor mentions, and brand sentiment. Analysts & researchers:** Effortlessly gather source material and data points on specific research topics. Business owners & entrepreneurs:** Stay updated on market shifts, new technologies, and potential opportunities without dedicating hours to reading. Anyone with a passion project:** Easily follow developments in your favorite hobby, field of study, or area of interest. What problem does this solve? Eliminates manual searching:** Frees you from the daily or weekly grind of searching multiple news sites for relevant articles. Centralizes information:** Consolidates all relevant news into a single, organized, and easily accessible Airtable database. Provides structured data:** Instead of just a list of links, it extracts and formats key information (title, summary, URL, date) for each article, ready for review or analysis. Keeps you proactively informed:** The automated Slack notification ensures you know exactly when new information is ready, closing the loop on your monitoring process. How it works Schedule: The workflow runs automatically based on a schedule you set (the default is weekly). Define topics: In the Set news parameters node, you specify the topics you want to monitor and the time frame (e.g., news from the last 7 days). AI web search: The Query Linkup for news node sends your topics to Linkup's API. Linkup's AI searches the web for relevant news articles and returns a structured list containing each article's title, URL, summary, and publication date. Store in Airtable: The workflow loops through each article found and creates a new record for it in your Airtable base. Notify on Slack: Once all the news has been stored, a final notification is sent to a Slack channel of your choice, letting you know the process is complete and how many articles were found. Setup Configure the trigger: Adjust the Schedule Trigger node to set the frequency and time you want the workflow to run. Set your topics: In the Set news parameters node, replace the example topics with your own keywords and define the news freshness that you'd like to set. Connect your accounts: Linkup: Add your Linkup API key in the Query Linkup for news node. Linkup's free plan includes โฌ5 of credits monthly, enough for about 1,000 runs of this workflow. Airtable: In the Store one news node, select your Airtable account, then choose the Base and Table where you want to save the news. Slack: In the Notify in Slack node, select your Slack account and the channel where you want to receive notifications. Activate the workflow: Toggle the workflow to "Active", and your automated news monitoring system is live! Taking it further Change your database:* Don't use Airtable? Easily swap the *Airtable* node for a *Notion, **Google Sheets, or any other database node to store your news. Customize notifications:* Replace the *Slack* node with a *Discord, **Telegram, or Email node to get alerts on your preferred platform. Add AI analysis:** Insert an AI node after the Linkup search to perform sentiment analysis on the news summaries, categorize articles, or generate a high-level overview before saving them.
by Yang
Who is this for? This workflow is for content creators, social media managers, marketing teams, and virtual assistants who want to automatically repurpose YouTube videos into ready-to-post social media content. If you need to quickly turn long-form videos into short posts for platforms like Instagram, Facebook, or LinkedIn, this workflow saves you hours of manual work. What problem is this workflow solving? Manually extracting ideas from YouTube videos, writing captions, creating images, and preparing social media posts takes a lot of time and effort. This workflow automates the entire process: it reads the video, generates posts with captions and AI images, and organizes everything into Airtable. It lets you focus more on growing your audience instead of spending hours repurposing content. What this workflow does Watches a YouTube channel RSS feed for new videos. Extracts the video transcript automatically using Dumpling AI. Summarizes and transforms the transcript into 3 social media captions (Instagram, Facebook, LinkedIn) using OpenAI. Generates 3 unique AI image prompts. Sends the prompts to Dumpling AI to create realistic social media images. Saves the captions and attaches the AI images into Airtable, ready for posting. Setup RSS Feed Setup Get your YouTube channelโs RSS feed URL. Insert the URL into the RSS Trigger node. This will monitor for new YouTube uploads automatically. Dumpling AI Setup for Transcript Extraction Sign up at Dumpling AI. Get your Dumpling AI API Key. In the first HTTP Request node after the RSS trigger, insert your API Key (use HTTP Header Authentication). This sends the YouTube URL to Dumpling AIโs /extract-transcript endpoint. OpenAI Setup for Caption and Prompt Generation Get your OpenAI API Key. In the OpenAI node, connect your account. The AI will: Generate 3 platform-specific captions. Generate 3 creative prompts to design images related to the video. Edit Fields Node This node organizes the generated captions and prompts into separate fields for easy Airtable mapping. Captions are split for Instagram, Facebook, and LinkedIn. Dumpling AI Setup for AI Image Generation After the Edit Fields node, the second HTTP Request node sends the image prompt to Dumpling AIโs /generate-image endpoint. This returns a realistic AI-generated image. Airtable Setup for Saving Posts (Without Image First) Create a new base in Airtable with the following fields: Platform (Single select: Instagram, Facebook, LinkedIn) Content (Long text field) Image (Attachment field) Connect your Airtable Personal Access Token to the Airtable node. The Airtable node saves the generated captions into separate records, initially without images. Upload Generated Images Back to Airtable The third HTTP Request node PATCHES the Airtable record. It updates the Image field with the generated AI image from Dumpling AI. Credentials Required Dumpling AI API Key (for transcript extraction and AI image generation) OpenAI API Key (for caption and prompt creation) Airtable Personal Access Token (for inserting and updating records) How to customize this workflow to your needs Change the OpenAI prompt to generate captions in your brand tone (e.g., friendly, professional, witty). Modify the image prompts to match your design style better. Adjust the Airtable base fields if you want to add more platforms or content formats. Add scheduling tools like Buffer or Metricool to automatically post from Airtable. โก Quick Tips Make sure Dumpling AI credits are active to allow transcript and image generation. Set Airtable permissions properly so PATCH requests can update attachments.