by Nijan
This workflow turns Slack into your content control hub and automates the full blog creation pipeline — from sourcing trending headlines, validating topics, drafting posts, and preparing content for your CMS. With one command in Slack, you can source news from RSS feeds, refine them with Gemini AI, generate high-quality blog posts, and get publish-ready output — all inside a single n8n workflow. ⸻ ⚙️ How It Works 1.Trigger in Slack Type start in a Slack channel to fetch trending headlines. Headlines are pulled from your configured RSS feeds. 2.Topic Generation (Gemini AI) Gemini rewrites RSS headlines into unique, non-duplicate topics. Slack displays these topics in a numbered list (e.g., reply with 2 to pick topic 2). 3.Content Validation When you reply with a number, Gemini validates and slightly rewrites the topic to ensure originality. Slack confirms the selected topic back to you. 4.Content Creation Gemini generates a LinkedIn/blog-style draft: Strong hook introduction 3–5 bullet insights A closing takeaway and CTA Optionally suggests asset ideas (e.g., image, infographic). 5.CMS-Ready Output Final draft is structured for publishing (markdown or plain text). You can expand this workflow to automatically send the output to your CMS (WordPress, Ghost, Notion, etc.). ⸻ 🛠 Setup Instructions Connect your Slack Bot to n8n. Configure your RSS Read nodes with feeds relevant to your niche. Add your Gemini API credentials in the AI node. Run the workflow: Type start in Slack → see trending topics. Reply with a number (e.g., gen 3) → get a generated blog draft in the same Slack thread. ⸻ 🎛 Customization Options • Change RSS sources to match your industry. • Adjust Gemini prompts for tone (educational, casual, professional). • Add moderation filters (skip sensitive or irrelevant topics). • Connect the final output step to your CMS, Notion, or Google Docs for publishing. ⸻ ✅ Why Use This Workflow? • One-stop flow: Sourcing → Validation → Writing → Publishing. • Hands-free control: Everything happens from Slack. • Flexible: Easily switch feeds, tone, or target CMS. • Scalable: Extend to newsletters, social posts, or knowledge bases.
by Summer
LinkedIn Job Search Automation Creator: Summer Chang Setup Instructions This n8n workflow automatically searches for senior designer jobs on LinkedIn every day at 5am and saves them to a Notion database. Prerequisites n8n instance (cloud or self-hosted) Notion account with API access A Notion database set up to receive job listings Setup Steps ✅ 1. Create Your Notion Database Or duplicate my template ✅ 2. Connect Notion to n8n In the "Save to Notion" node, click on the Notion credentials Follow the authentication flow to connect your Notion account Select your job search database from the dropdown ✅ 3. Customize Your Search Criteria In the "Set Search Criteria" node, modify these parameters to match your job preferences: search_keywords: Job titles to search for (comma-separated) Default: senior product designer, product design lead, senior UX designer, AI designer excluded_keywords: Terms to filter out (comma-separated) Default: contract, freelance location: Where you want to work (comma-separated) Default: remote, san francisco f_TPR: Time filter for job postings r86400 = Last 24 hours r604800 = Last week r2592000 = Last month sortBy: How to sort results DD = Most recent first R = Most relevant first ✅ 4. Adjust the Schedule In the "Everyday @5am" node: Click on the node Modify the schedule to your preferred time You can set it to run daily, weekly, or at custom intervals ✅ 5. Set Result Limits In the "Limit1" node: Default: Processes 10 jobs per run Adjust the maxItems value to get more or fewer results ✅ 6. Configure Wait Time (Optional) The "Wait2" node adds a 10-second delay between requests to avoid rate limiting: Default: 10 seconds Increase if you're getting blocked by LinkedIn Decrease for faster processing (not recommended) How It Works Trigger: Runs automatically every day at 5am Search: Queries LinkedIn with your specified criteria Parse: Extracts job title, company, location, and URL from search results Filter: Removes any jobs with missing critical information Wait: Delays between requests to avoid rate limiting Fetch Details: Retrieves full job descriptions and poster information Save: Adds each job to your Notion database
by Guillaume Duvernay
Move beyond generic AI-generated content and create articles that are high-quality, factually reliable, and aligned with your unique expertise. This template orchestrates a sophisticated "research-first" content creation process. Instead of simply asking an AI to write an article from scratch, it first uses an AI planner to break your topic down into logical sub-questions. It then queries a Super assistant—which you've connected to your own trusted knowledge sources like Notion, Google Drive, or PDFs—to build a comprehensive research brief. Only then is this fact-checked brief handed to a powerful AI writer to compose the final article, complete with source links. This is the ultimate workflow for scaling expert-level content creation. Who is this for? Content marketers & SEO specialists:** Scale the creation of authoritative, expert-level blog posts that are grounded in factual, source-based information. Technical writers & subject matter experts:** Transform your complex internal documentation into accessible public-facing articles, tutorials, and guides. Marketing agencies:** Quickly generate high-quality, well-researched drafts for clients by connecting the workflow to their provided brand and product materials. What problem does this solve? Reduces AI "hallucinations":** By grounding the entire writing process in your own trusted knowledge base, the AI generates content based on facts you provide, not on potentially incorrect information from its general training data. Ensures comprehensive topic coverage:** The initial AI-powered "topic breakdown" step acts like an expert outliner, ensuring the final article is well-structured and covers all key sub-topics. Automates source citation:** The workflow is designed to preserve and integrate source URLs from your knowledge base directly into the final article as hyperlinks, boosting credibility and saving you manual effort. Scales expert content creation:** It effectively mimics the workflow of a human expert (outline, research, consolidate, write) but in an automated, scalable, and incredibly fast way. How it works This workflow follows a sophisticated, multi-step process to ensure the highest quality output: Decomposition: You provide an article title and guidelines via the built-in form. An initial AI call then acts as a "planner," breaking down the main topic into an array of 5-8 logical sub-questions. Fact-based research (RAG): The workflow loops through each of these sub-questions and queries your Super assistant. This assistant, which you have pre-configured and connected to your own knowledge sources (Notion pages, Google Drive folders, PDFs, etc.), finds the relevant information and source links for each point. Consolidation: All the retrieved question-and-answer pairs are compiled into a single, comprehensive research brief. Final article generation: This complete, fact-checked brief is handed to a final, powerful AI writer (e.g., GPT-5). Its instructions are clear: write a high-quality article using only the provided information and integrate the source links as hyperlinks where appropriate. Implementing the template Set up your Super assistant (Prerequisite): First, go to Super, create an assistant, connect it to your knowledge sources (Notion, Drive, etc.), and copy its Assistant ID and your API Token. Configure the workflow: Connect your AI provider (e.g., OpenAI) credentials to the two Language Model nodes (GPT 5 mini and GPT 5 chat). In the Query Super Assistant (HTTP Request) node, paste your Assistant ID in the body and add your Super API Token for authentication (we recommend using a Bearer Token credential). Activate the workflow: Toggle the workflow to "Active" and use the built-in form to generate your first fact-checked article! Taking it further Automate publishing:* Connect the final *Article result* node to a *Webflow* or *WordPress** node to automatically create a draft post in your CMS. Generate content in bulk:* Replace the *Form Trigger* with an *Airtable* or *Google Sheet** trigger to automatically generate a whole batch of articles from your content calendar. Customize the writing style:* Tweak the system prompt in the final *New content - Generate the AI output** node to match your brand's specific tone of voice, add SEO keywords, or include specific calls-to-action.
by Dr. Christoph Schorsch
Rename Workflow Nodes with AI for Clarity This workflow automates the tedious process of renaming nodes in your n8n workflows. Instead of manually editing each node, it uses an AI language model to analyze its function and assign a concise, descriptive new name. This ensures your workflows are clean, readable, and easy to maintain. Who's it for? This template is perfect for n8n developers and power users who build complex workflows. If you often find yourself struggling to understand the purpose of different nodes at a glance or spend too much time manually renaming them for documentation, this tool will save you significant time and effort. How it works / What it does The workflow operates in a simple, automated sequence: Configure Suffix: A "Set" node at the beginning allows you to easily define the suffix that will be appended to the new workflow's name (e.g., "- new node names"). Fetch Workflow: It then fetches the JSON data of a specified n8n workflow using its ID. AI-Powered Renaming: The workflow's JSON is sent to an AI model (like Google Gemini or Anthropic Claude), which has been prompted to act as an n8n expert. The AI analyzes the type and parameters of each node to understand its function. Generate New Names: Based on this analysis, the AI proposes new, meaningful names and returns them in a structured JSON format. Update and Recreate: A Code Node processes these suggestions, updates all node names, and correctly rebuilds the connections and expressions. Create & Activate New Workflow: Finally, it creates a new workflow with the updated name, deactivates the original to avoid confusion, and activates the new version.
by Budi SJ
Automated Brand DNA Generator Using JotForm, Google Search, AI Extraction & Notion The Brand DNA Generator workflow automatically scans and analyzes online content to build a company’s Brand DNA profile. It starts with input from a form, then crawls the company’s website and Google search results to gather relevant information. Using AI-powered extraction, the system identifies insights such as value propositions, ideal customer profiles (ICP), pain points, proof points, brand tone, and more. All results are neatly formatted and automatically saved to a Notion database as a structured Brand DNA report, eliminating the need for manual research. 🛠️ Key Features Automated data capture, collects company data directly from form submissions and Google search results. Uses AI-powered insight extraction with LLMs to extract and summarize brand-related information from website content. Fetches clean text from multiple web pages using HTTP requests and a content extractor. Merges extracted data from multiple sources into a single Brand DNA JSON structure. Automatically creates a new page in Notion with formatted sections (headings, paragraphs, and bullet points). Handles parsing failures and processes multiple pages efficiently in batches. 🔧 Requirements JotForm API Key, to capture company data from form submissions. SerpAPI Key, to perform automated Google searches. OpenRouter / LLM API, for AI-based language understanding and information extraction. Notion Integration Token & Database ID, to save the final Brand DNA report to Notion. 🧩 Setup Instructions Connect your JotForm account and select the form containing the fields Company Name and Company Website. Add your SerpAPI Key. Configure the AI model using OpenRouter or LLM. Enter your Notion credentials and specify the databaseId in the Create a Database Page node. Customize the prompt in the Information Extractor node to modify the tone or structure of AI analysis (Optional). Activate the workflow, then submit data through the JotForm to test automatic generation and Notion integration. 💡 Final Output A complete Brand DNA Report containing: Company Description Ideal Customer Profile Pain Points Value Proposition Proof Points Brand Tone Suggested Keywords All generated automatically from the company’s online presence and stored in Notion with no manual input required.
by Stephan Koning
Recruiter Mirror is a proof‑of‑concept ATS analysis tool for SDRs/BDRs. Compare your LinkedIn or CV to job descriptions and get recruiter‑ready insights. By comparing candidate profiles against job descriptions, it highlights strengths, flags missing keywords, and generates actionable optimization tips. Designed as a practical proof of concept for breaking into tech sales, it shows how automation and AI prompts can turn LinkedIn into a recruiter‑ready magnet. Got it ✅ — based on your workflow (Webhook → LinkedIn CV/JD fetch → GhostGenius API → n8n parsing/transform → Groq LLM → Output to Webhook), here’s a clear list of tools & APIs required to set up your Recruiter Mirror (Proof of Concept) project: 🔧 Tools & APIs Required 1. n8n (Automation Platform) Either n8n Cloud or self‑hosted n8n instance. Used to orchestrate the workflow, manage nodes, and handle credentials securely. 2. Webhook Node (Form Intake) Captures LinkedIn profile (LinkedIn_CV) and job posting (LinkedIn_JD) links submitted by the user. Acts as the starting point for the workflow. 3. GhostGenius API Endpoints Used: /v2/profile → Scrapes and returns structured CV/LinkedIn data. /v2/job → Scrapes and returns structured job description data. Auth**: Requires valid credentials (e.g., API key / header auth). 4. Groq LLM API (via n8n node) Model Used: moonshotai/kimi-k2-instruct (via Groq Chat Model node). Purpose: Runs the ATS Recruiter Check, comparing CV JSON vs JD JSON, then outputs a structured JSON per the ATS schema. Auth**: Groq account + saved API credentials in n8n. 5. Code Node (JavaScript Transformation) Parses Groq’s JSON output safely (JSON.parse). Generates clean, recruiter‑ready HTML summaries with structured sections: Status Reasoning Recommendation Matched keywords / Missing keywords Optimization tips 6. n8n Native Nodes Set & Aggregate Nodes** → Rebuild structured CV & JD objects. Merge Node** → Combine CV data with job description for comparison. If Node** → Validates LinkedIn URL before processing (fallback to error messaging). Respond to Webhook Node** → Sends back the final recruiter‑ready insights in JSON (or HTML). ⚠️ Important Notes Credentials**: Store API keys & auth headers securely inside n8n Credentials Manager (never hardcode inside nodes). Proof of Concept: This workflow demonstrates feasibility but is **not production‑ready (scraping stability, LinkedIn terms of use, and API limits should be considered before real deployments).
by takuma
Who's it for This template is for home cooks, small restaurant owners, or anyone who wants to streamline their meal planning, ingredient cost tracking, leftover management, nutritional analysis, and social media promotion. It's ideal for those looking to optimize their kitchen operations, reduce food waste, maintain a healthy diet, and efficiently share their culinary creations. How it works / What it does This advanced workflow acts as a comprehensive culinary assistant. Triggered by a new menu item, it performs several key functions: Cost and Ingredient Tracking:** A "Menu Agent" uses AI to analyze your input (e.g., a recipe or dish) and extract a detailed list of ingredients, their associated costs, unit prices, and total cost, then logs this into a Google Sheet as a "Recipe List." Leftover Management:** A "Leftovers Agent" identifies any unused ingredients from your planned dish and suggests three new recipes to utilize them, helping to minimize food waste. This information is also recorded in a Google Sheet. Nutritional Diary:** A "Nutritionist Agent" generates a diary-style entry with dietary advice based on the meal, highlighting key nutrients and offering personalized suggestions. This entry is appended to a "Diary" Google Sheet. Social Media Promotion:** A "Post Agent" takes the nutritional diary entry and transforms it into an engaging social media post (specifically for X/Twitter in this template), which is then sent as a direct message, ready for you to share with your followers. How to set up Webhook Trigger: The workflow starts with a Webhook. Copy the webhook URL from the "Webhook" node. You will send your menu item input to this URL. Google Sheets Integration: You need to set up a Google Sheets credential for your n8n instance. Create a Google Sheet document (e.g., "Recipe List"). Within this document, create three sheets: "Recipe: This sheet will store your menu items, ingredients, costs, etc. Ensure it has columns for Date, Item, Ingredients, Ingredient Cost, Unit Price, Quantity, Total, Cost, and Leftover Ingredients. "leftovers" (Leftovers): This sheet will store suggested recipes for leftover ingredients. Ensure it has columns for Date and Ingredients. "diary" (Diary): This sheet will store your nutritional diary entries. Ensure it has a column for Diary. In the "Append row in sheet", "Append row in sheet1", and "Append row in sheet2" nodes, replace the Document ID with the ID of your Google Sheet. For "Sheet Name," ensure you select the correct sheet (e.g., "レシピ", "diary", "leftovers") from the dropdown. OpenRouter Chat Model: Set up your OpenRouter credentials in the "OpenRouter Chat Model" nodes. You will need your OpenRouter API key. Twitter Integration: Set up your Twitter credentials for the "Create Direct Message" node. In the "Create Direct Message" node, specify the User (username) to whom the direct message should be sent. This is typically your own Twitter handle or a test account. Requirements An n8n instance. A Google account with Google Sheets enabled. An OpenRouter API key. A Twitter (X) account with developer access to send Direct Messages. How to customize the workflow Input Data:** The initial input to the "Webhook" node is expected to be the name of a dish or recipe. You can modify the "Menu Agent" to accept more detailed inputs if needed. Google Sheets Structure:** Adjust the column mappings in the Google Sheets nodes if your spreadsheet column headers differ. AI Agent Prompts:** Customize the System Message in each AI Agent node (Menu Agent, Leftovers Agent, Nutritionist Agent, Post Agent) to refine their behavior and the kind of output they generate. For example, you could ask the Nutritionist Agent to focus on specific dietary needs. Social Media Platform:** The "Create Direct Message" node is configured for Twitter. You can swap this with another social media node (e.g., Mastodon, Discord) if you prefer to post elsewhere, remembering to adjust the "Post Agent" system message accordingly. Output Parser:** The "Structured Output Parser" is configured for a specific JSON structure. If you change the "Menu Agent" to output a different structure, you'll need to update this parser.
by Dahiana
Description Who's it for: Content creators, marketers, and businesses who publish on both YouTube and blog platforms. What it does: Monitors your YouTube channel for new videos and automatically creates SEO-optimized blog posts using AI, then publishes to WordPress or Webflow. How it works: RSS Feed Trigger polls YouTube videos (every X amount of time) Extracts video metadata (title, description, thumbnail) YouTube node extracts full description for extra context Uses OpenAI (you can choose any model) to generate 600-800 word blog post Publishes to WordPress AND/OR Webflow with error handling Sends notifications to Telegram if publishing fails Requirements: YouTube channel ID (avoid tutorial channels for better results) OpenAI API key (or similar) WordPress OR Webflow credentials Telegram bot (optional, for error notifications) Setup steps: Replace YOUR_CHANNEL_ID in RSS Feed Trigger Add OpenAI credentials in AI generation node Configure WordPress and/or Webflow credentials Add Telegram bot for error notifications (optional). If you choose to set up Telegram, you need to input your channel ID. Test with manual execution first Customization: Modify AI prompt for different content styles Adjust polling frequency (30-60 minutes recommended) Add more CMS platforms Add content verification (is content larger than 600 characters? if not, improve)
by noda
Price Anomaly Detection & News Alert (Marketstack + HN + DeepL + Slack) Overview This workflow monitors a stock’s closing price via Marketstack. It computes a 20-day moving average and standard deviation (±2σ). If the latest close is outside ±2σ, it flags an anomaly, fetches related headlines from Hacker News, translates them to Japanese with DeepL, and posts both original and translated text to Slack. When no anomaly is detected, it sends a concise “normal” report. How it works 1) Daily trigger at 09:00 JST 2) Marketstack: fetch EOD data 3) Code: compute mean/σ and classify (normal/high/low) 4) IF: anomaly? → yes = news path / no = normal report 5) Hacker News: search related items 6) DeepL: translate EN → JA 7) Slack: send bilingual notification Requirements Marketstack API key DeepL API key Slack OAuth2 (bot token / channel permission) Notes Edit the ticker in Get Stock Data. Adjust N (days) and k (sigma multiplier) in Calculate Deviation. Keep credentials out of HTTP nodes (use n8n Credentials).
by Rahul Joshi
Description Automatically detect customer churn risks from Zendesk tickets, log them into Google Sheets for tracking, and send instant Slack alerts to your customer success team. This workflow helps you spot unhappy customers early and take proactive action to reduce churn. 🚨📊💬 What This Template Does Fetches Zendesk tickets daily on schedule (8:00 PM). ⏰ Processes and formats ticket data into clean JSON (priority, age, urgency). 🧠 Identifies churn risks based on negative satisfaction ratings. ⚠️ Logs churn risk tickets into Google Sheets for analysis and reporting. 📈 Sends formatted Slack alerts with ticket details to the CS team channel. 📢 Key Benefits Detects unhappy customers before they churn. 🚨 Centralized churn tracking for reporting and team reviews. 🧾 Proactive alerts to reduce response delays. ⏱️ Clean, structured ticket data for analytics and filtering. 🔄 Strengthens customer success strategy with real-time visibility. 🌐 Features Schedule Trigger – Runs every weekday at 8:00 PM. 🗓️ Zendesk Integration – Fetches all tickets automatically. 🎫 Smart Data Processing – Adds ticket age, urgency, and priority mapping. 🧮 Churn Risk Filter – Flags tickets with negative satisfaction scores. 🚩 Google Sheets Logging – Saves churn risk details with metadata. 📊 Slack Alerts – Sends formatted messages with ID, subject, rating, and action steps. 💬 Requirements n8n instance (cloud or self-hosted). Zendesk API credentials with ticket read access. Google Sheets OAuth2 credentials with write permissions. Slack Bot API credentials with channel posting permissions. Pre-configured Google Sheet for churn risk logging. Target Audience Customer Success teams monitoring churn risk. 👩💻 SaaS companies tracking customer health. 🚀 Support managers who want proactive churn alerts. 🛠️ SMBs improving retention through automation. 🏢 Remote CS teams needing instant notifications. 🌐 Step-by-Step Setup Instructions Connect your Zendesk, Google Sheets, and Slack credentials in n8n. 🔑 Update the Schedule Trigger (default: daily at 8:00 PM) if needed. ⏰ Replace the Google Sheet ID with your churn risk tracking sheet. 📊 Confirm the Slack channel ID for alerts (default: zendesk-churn-alerts). 💬 Adjust churn filter logic (default: satisfaction_score = "bad"). 🎯 Run a test to fetch Zendesk tickets and validate Sheets + Slack outputs. ✅
by Sk developer
🚀 All-In-One Video Downloader to Google Drive (via RapidAPI best All-In-One Video Downloader) Description: This n8n workflow automates the process of downloading videos from any supported platform (like LinkedIn, Facebook, or Instagram) using the RapidAPI best All-In-One Video Downloader. It then uploads the video to your Google Drive and shares it publicly, while logging any failures in Google Sheets for tracking. 📦 Node-by-Node Breakdown | 🧩 Node Name | 📝 One‑Line Explanation | |-------------------------------|-------------------------------------------------------------------------------| | On form submission | Triggers the workflow when a user submits a video URL through a web form. | | All in one video downloader | Sends a POST request to RapidAPI best All-In-One Video Downloader to fetch downloadable video links. | | If | Checks whether the API response includes an error and routes accordingly. | | Download mp4 | Downloads the video using the direct media URL received from the API. | | Upload To Google Drive | Uploads the MP4 file to a designated folder in your Google Drive. | | Google Drive Set Permission | Makes the uploaded file publicly shareable with a viewable link. | | Wait | Adds a short delay before logging errors to prevent duplicate entries. | | Google Sheets Append Row | Logs failed download attempts with the original URL and status as N/A. | ✅ Benefits of This Flow 🔁 End-to-End Automation: From user input to shareable video link—no manual steps required. 🌐 Supports Multiple Platforms: The RapidAPI best All-In-One Video Downloader supports sites like Instagram, Facebook, Twitter, LinkedIn, and more. ⚠️ Smart Error Handling: Automatically logs failed download attempts into Google Sheets for retry or audit. ☁️ Cloud Ready: Videos are stored in Google Drive with instant public access. 📊 Trackability: Logs failures, timestamps, and source URLs for easy debugging or analytics. 🧩 Modular Setup: Easily expand this in n8n to include Slack notifications, email alerts, or tagging. 🔁 Use Cases 🎬 Social Media Video Archiving: Download and store content (Reels, posts, stories) into Drive for future use. 🧑🏫 Educational Sharing: Teachers can collect useful videos and share links with students. 📚 Content Curation: Bloggers or content managers can create a media archive from multiple platforms. 🤝 Team Automation: Teams submit links, and the workflow handles download + Drive share link generation. 📉 Error Tracking for Ops: Failed URLs are tracked in Google Sheets for retry, monitoring, or debugging. 🧠 Final Thoughts This workflow leverages the power of n8n and RapidAPI best All-In-One Video Downloader to create a fully automated pipeline for capturing video content from across the web. It’s ideal for educators, marketers, content curators, or developers who want to streamline video storage and access using Google Drive. 🔑 How to Get API Key from RapidAPI Best All-In-One Video Downloader Follow these steps to get your API key and start using it in your workflow: Visit the API Page 👉 Click here to open Best All-In-One Video Downloader on RapidAPI Log in or Sign Up Use your Google, GitHub, or email account to sign in. If you're new, complete a quick sign-up. Subscribe to a Pricing Plan Go to the Pricing tab on the API page. Select a plan (free or paid, depending on your needs). Click Subscribe. Access Your API Key Navigate to the Endpoints tab. Look for the X-RapidAPI-Key under Request Headers. Copy the value shown — this is your API key. Use the Key in Your Workflow In your n8n workflow (HTTP Request node), replace: "x-rapidapi-key": "your key" with: "x-rapidapi-key": "YOUR_ACTUAL_API_KEY" ✅ You’re now ready to use the Best All-In-One Video Downloader with your automated workflows!
by Sabrina Ramonov 🍄
Description This AI Agent Carousel Maker uses ChatGPT and Blotato to write, generate, and auto-post social media carousels to 5 social platforms: Instagram, Tiktok, Facebook, Twitter, and Pinterest. Simply chat with the AI agent, confirm which prebuilt viral carousel template you want to use, then the AI Agent populates the template with your personalized information and quotes, and posts to social media on autopilot. Who Is This For? This is perfect for entrepreneurs, small businesses, content creators, digital marketing agencies, social media marketing agencies, and influencers. How It Works 1. Chat: AI Agent Carousel Maker Chat with AI agent about your desired carousel Confirm quotes and carousel template to use 2. Carousel Generation AI agent calls corresponding Blotato tool to generate carousel Wait and fetch completed carousel 3. Publish to Social Media via Blotato Choose your social accounts Either post immediately or schedule for later Setup Sign up for OpenAPI API access and create credential Sign up for Blotato.com Generate Blotato API Key by going to Settings > API > Generate API Key (paid feature only) Create Blotato credential If you're using n8n, ensure you have ""Verified Community Nodes"" enabled in your n8n Admin Panel. Then, install ""Blotato"" verified community node. Click ""Open chat"" to test workflow Complete SETUP sticky notes in BROWN in this template Tips & Tricks AFTER your first successful run, open each carousel template tool call (i.e. pink nodes attached to AI Agent Carousel Maker) and tweak the parameters, but DO NOT change the ""quotes"" parameter unless you're an n8n expert. DO NOT edit the 'quotes' parameter unless you're an n8n expert. When adding a new template, DO NOT duplicate an existing node. Instead, click '+ Tool' > Blotato Tool > Video > Create > select new template. This ensures template parameters are correctly loaded. While testing: enable only 1 social platform, and deactivate the rest. Add optional parameter 'scheduledTime' so that you don't accidentally post to social media. Check your content calendar here: https://my.blotato.com/queue/schedules 📄 Documentation Full Tutorial Troubleshooting Check Blotato API Dashboard and logs to review requests, responses, and errors. Verify template parameters and n8n node configuration if runs fail. You can also: View all video/carousel templates available Check how your carousels look Need Help? In the Blotato web app, click the orange button on the bottom right corner. This opens the Support messenger where I help answer technical questions. Connect with me: Linkedin | Youtube