by Alfred Nutile
How it works This workflow provides a streamlined process for uploading files to Digital Ocean Spaces, making them publicly accessible. The process happens in three main steps: User submits the form with file, in this case I needed it to upload images I use in my seo tags. File is automatically uploaded to Digital Ocean Spaces using S3-compatible storage Form completion confirmation is provided Setup steps Initial setup typically takes 5-10 minutes Configure your Digital Ocean Spaces credentials and bucket settings Test the upload functionality with a small sample file Verify public access permissions are working as expected Important notes Credentials are tricky check the screenshot above for how I set the url, bucket etc. I am just using the S3 Node Set the ACL as seen below Troubleshooting Bucket name might be incorrect Region Wrong Check Space permissions if uploads fail Verify API credentials are correctly configured You can see a video here. (live in 24 hours) https://youtu.be/pYOpy3Ntt1o
by Ludwig
How it works: This workflow automates tagging for WordPress posts using AI: Fetch blog post content and metadata. Generate contextually relevant tags using AI. Verify existing tags in WordPress and create new ones if necessary. Automatically update posts with accurate and optimized tags. Set up steps: Estimated time: ~15 minutes. Configure the workflow with your WordPress API credentials. Connect your content source (e.g., RSS feed or manual input). Adjust tag formatting preferences in the workflow settings. Run the workflow to ensure proper tag creation and assignment. This workflow is perfect for marketers and content managers looking to streamline their content categorization and improve SEO efficiency.
by Madame AI
AI Image Remix & Design Bot for Telegram with BrowserAct & Gemini This workflow transforms your Telegram bot into an intelligent creative assistant. It can chat conversationally, fetch trending image prompts from PromptHero for inspiration, or perform a deep "remix" of any photo you upload by analyzing its composition and regenerating it with high-fidelity prompt engineering. Target Audience Digital artists, designers, content creators, and hobbyists looking for AI-assisted inspiration and image generation. How it works Traffic Control: The workflow starts with a Telegram Trigger and immediately splits traffic: new messages go one way, while interactive button clicks (like "Regenerate") go another. Intent Classification: An AI Agent analyzes text inputs to decide if the user wants to "Chat" (small talk) or "Start" a creative session (fetch inspiration). Inspiration Mode: If "Start" is detected, BrowserAct scrapes trending prompts from PromptHero and saves them to a Google Sheet. Visual Forensics: If the user uploads an image, an AI Vision Agent (using OpenRouter/Gemini) analyzes it in extreme detail (lighting, composition, subjects) and saves the description. Master Prompt Engineering: Specialized AI Agents expand these inputs (either scraped prompts or image descriptions) into massive, detailed prompts using the "Rule of Multiplication." Production: Google Gemini generates the new image, which is sent back to Telegram with interactive buttons to "Regenerate" or move to the "Next" idea. ⚠️ Complex Workflow This workflow is complex. Please proceed using the tutorial video. How to set up Configure Credentials: Connect your Telegram, Google Sheets, BrowserAct, Google Gemini, and OpenRouter accounts in n8n. Prepare BrowserAct: Ensure the Image Remix & Design Bot template is saved in your BrowserAct account. Setup Google Sheet: Create a Google Sheet with four tabs: PromptHero, Current State, UserImage, and Current Image. Connect Sheet: Open all Google Sheets nodes in the workflow and paste your spreadsheet ID. Configure Telegram: Ensure your bot is created via BotFather and the API token is added to the Telegram credentials. Activate: Turn on the workflow. Requirements BrowserAct* account with the *Image Remix & Design Bot** template. Telegram** account (Bot Token). Google Sheets** account. Google Gemini** account. OpenRouter** account (or compatible LLM credentials). How to customize the workflow Change Art Style: Modify the system prompt in the Generate Image agents to enforce a specific style (e.g., "Cyberpunk," "Watercolor," or "Photorealistic"). Add More Sources: Update the BrowserAct template to scrape prompts from other sites like Civitai or Midjourney feed. Switch Image Model: Replace the Gemini image generation node with Stable Diffusion or DALL-E 3 if you prefer different aesthetics. Need Help? How to Find Your BrowserAct API Key & Workflow ID How to Connect n8n to BrowserAct How to Use & Customize BrowserAct Templates Workflow Guidance and Showcase Video How To create stateful n8n Workflows | AI Image Remix Bot with n8n & BrowserAct & Telegram 🎨
by Gilbert Onyebuchi
Automate video creation: AI generates ideas, Vertex AI renders videos, and auto-uploads to Google Drive with complete tracking. What You Get Gemini AI for creative prompts Vertex AI video generation Auto-upload to Google Drive Complete Google Sheets logging Smart retry logic Base64 to MP4 conversion Setup Enable Vertex AI in Google Cloud Get Gemini API key Run gcloud auth print-access-token for ACCESS TOKEN Import workflow & configure credentials Add prompts & test Flow Schedule → Gemini AI → Vertex AI → Wait → Convert → Upload → Log Resources Google Sheets Template ⚠️ Note: ACCESS TOKEN expires hourly - refresh using gcloud auth print-access-token 📧 LinkedIn: linkedin.com/in/yourprofile 🔗 More n8n Products: Click here
by Madame AI
Generate visual resumes from Telegram inputs using Google Gemini This workflow transforms text-based resume data into visually stunning images by leveraging Google Gemini's reasoning and vision capabilities. It autonomously analyzes the candidate's profile, selects an appropriate design template based on their industry, and renders a high-quality resume image directly in Telegram. Target Audience Job seekers, career coaches, resume writers, and recruitment agencies looking to automate design generation. How it works Classify Input: The workflow starts with a Telegram trigger. A Google Gemini agent analyzes the incoming message to determine if it is a casual chat or a resume generation request. Fetch Context: If it is a resume request, a BrowserAct node triggers a workflow (using the "AI Resume Replicant" template) to fetch necessary external context or data. Ingest Designs (Optional): If a reference image is provided, CloudConvert standardizes the file, and Google Gemini Vision reverse-engineers the layout and style, saving the "Visual DNA" to Google Sheets. Draft Blueprint: The "Resume Writer" AI agent selects a stored design template that matches the candidate's industry (e.g., "Corporate" for Finance, "Creative" for Design) and maps the text content to the layout. Generate Prompt: A "Visualizer" AI agent converts the structured blueprint into a highly detailed natural language prompt for image generation. Render & Deliver: Google Gemini generates the final resume image, which is then sent back to the user via Telegram. How to set up Configure Credentials: Connect your Telegram, Google Gemini, Google Sheets, CloudConvert, and BrowserAct accounts in n8n. Prepare BrowserAct: Ensure the AI Resume Replicant template is saved in your BrowserAct account. Setup Google Sheet: Create a new Google Sheet with the required header (listed below). Connect Sheet: Open the Google Sheets nodes (Clear, Get, Append) and select your new spreadsheet. Configure Telegram: Ensure your Telegram Bot is connected to the Trigger and Message nodes. Google Sheet Headers To use this workflow, create a Google Sheet with the following header: Resume Details Requirements BrowserAct* account (Template: *AI Resume Replicant**). Google Gemini** account. Telegram** account (Bot Token). CloudConvert** account. Google Sheets** account. How to customize the workflow Refine Design Logic: Modify the system prompt in the "Resume Writer" agent to change how the AI matches industries to design styles (e.g., force specific colors for specific roles). Change Output Format: Replace the Telegram response node with a Google Drive node to save the generated images as PDF or PNG files instead of sending them. Switch Image Model: Update the "Generate an image" node to use a different image generation model if preferred (e.g., OpenAI DALL-E). Need Help? How to Find Your BrowserAct API Key & Workflow ID How to Connect n8n to BrowserAct How to Use & Customize BrowserAct Templates Workflow Guidance and Showcase Video I Built a Resume Bot that CLONES Any Template! 🤖 (BrowserAct + n8n + Gemini Tutorial)
by Facundo Cabrera
Automated Meeting Minutes from Video Recordings This workflow automatically transforms video recordings of meetings into structured, professional meeting minutes in Notion. It uses local AI models (Whisper for transcription and Ollama for summarization) to ensure privacy and cost efficiency, while uploading the original video to Google Drive for safekeeping. Ideal for creative teams, production reviews, or any scenario where visual context is as important as the spoken word. 🔄 How It Works Wait & Detect: The workflow monitors a local folder. When a new .mkv video file is added, it waits until the file has finished copying. Prepare Audio: The video is converted into a .wav audio file optimized for transcription (under 25 MB with high clarity). Transcribe Locally: The local Whisper model generates a timestamped text transcript. Generate Smart Minutes: The transcript is sent to a local Ollama LLM, which produces structured, summarized meeting notes. Store & Share: The original video is uploaded to Google Drive, a new page is created in Notion with the notes and a link to the video, and a completion notification is sent via Discord. ⏱️ Setup Steps Estimated Time**: 10–15 minutes (for technically experienced users). Prerequisites**: Install Python, FFmpeg, and required packages (openai-whisper, ffmpeg-python). Run Ollama locally with a compatible model (e.g., gpt-oss:20b, llama3, mistral). Configure n8n credentials for Google Drive, Notion, and Discord. Workflow Configuration**: Update the file paths for the helper scripts (wait-for-file.ps1, create_wav.py, transcribe_return.py) in the respective "Execute Command" nodes. Change the input folder path (G:\OBS\videos) in the "File" node to your own recording directory. Replace the Google Drive folder ID and Notion database/page ID in their respective nodes. > 💡 Note: Detailed instructions for each step, including error handling and variable setup, are documented in the Sticky Notes within the workflow itself. 📁 Helper Scripts Documentation wait-for-file.ps1 A PowerShell script that checks if a file is still being written to (i.e., locked by another process). It returns 0 if the file is free and 1 if it is still locked. Usage: .\wait-for-file.ps1 -FilePath "C:\path\to\your\file.mkv" create_wav.py A Python script that converts a video file into a .wav audio file. It automatically calculates the necessary audio bitrate to keep the output file under 25 MB—a common requirement for many transcription services. Usage: python create_wav.py "C:\path\to\your\file.mkv" transcribe_return.py A Python script that uses a local Whisper model to transcribe an audio file. It can auto-detect the language or use a language code specified in the filename (e.g., meeting.en.mkv for English, meeting.es.mkv for Spanish). The transcript is printed directly to stdout with timestamps, which is then captured by the n8n workflow. Usage: Auto-detect language python transcribe_return.py "C:\path\to\your\file.mkv" Force language via filename python transcribe_return.py "C:\path\to\your\file.es.mkv" `
by Dr. Firas
💥 Generate product images with NanoBanana Pro to Veo videos and Blotato Who is this for? This workflow is designed for: Content creators and marketers E-commerce and product-based businesses Agencies producing social media visuals and videos Automation builders looking for AI-powered creative pipelines It is ideal for anyone who wants to automate product image and video creation using AI and publish content without manual work. What problem is this workflow solving? / Use case Creating product visuals and marketing videos usually requires multiple tools, manual prompt writing, and repetitive steps. This workflow solves: Manual image and video creation Inconsistent visual quality across assets Time-consuming prompt iteration Manual video publishing to social platforms The workflow automates the entire process from image generation to video publishing using AI. What this workflow does This workflow provides an end-to-end automation pipeline: Generates high-quality product images using NanoBanana Pro Applies Contact Sheet Prompting to explore multiple visual variations Converts selected images into short marketing videos using Veo 3.1 Automatically publishes the final videos via BLOTATO The result is a fully automated creative workflow that turns AI prompts into ready-to-publish video content. Setup To use this workflow, you need the following services and credentials: OpenAI API** Used for image analysis and prompt generation NanoBanana Pro (fal.ai)** Product image generation API: https://fal.ai/models/fal-ai/nano-banana-pro/edit/api Veo 3.1 (fal.ai)** Video generation API: https://fal.ai/models/fal-ai/veo3.1/first-last-frame-to-video Blotato** Video publishing to social platforms Sign up at BLOTATO All credentials must be added in n8n before running the workflow. How to customize this workflow to your needs You can easily adapt this workflow by: Modifying AI prompts to match your brand style Adjusting image composition and realism parameters in NanoBanana Pro Changing video motion, pacing, and aspect ratio in Veo 3.1 Selecting different social platforms or publishing rules in Blotato Replacing or extending individual steps while keeping the same architecture The workflow is modular and can be reused for multiple products or campaigns. 🎥 Watch This Tutorial 👋 Need help or want to customize this? 📩 Contact: LinkedIn 📺 YouTube: @DRFIRASS 🚀 Workshops: Mes Ateliers n8n 📄 Documentation: Notion Guide Need help customizing? Contact me for consulting and support : Linkedin / Youtube / 🚀 Mes Ateliers n8n
by Growth AI
N8N UGC Video Generator - Setup Instructions Transform Product Images into Professional UGC Videos with AI This powerful n8n workflow automatically converts product images into professional User-Generated Content (UGC) videos using cutting-edge AI technologies including Gemini 2.5 Flash, Claude 4 Sonnet, and VEO3 Fast. Who's it for Content creators** looking to scale video production E-commerce businesses** needing authentic product videos Marketing agencies** creating UGC campaigns for clients Social media managers** requiring quick video content How it works The workflow operates in 4 distinct phases: Phase 0: Setup - Configure all required API credentials and services Phase 1: Image Enhancement - AI analyzes and optimizes your product image Phase 2: Script Generation - Creates authentic dialogue scripts based on your input Phase 3: Video Production - Generates and merges professional video segments Requirements Essential Services & APIs Telegram Bot Token** (create via @BotFather) OpenRouter API** with Gemini 2.5 Flash access Anthropic API** for Claude 4 Sonnet KIE.AI Account** with VEO3 Fast access N8N Instance** (cloud or self-hosted) Technical Prerequisites Basic understanding of n8n workflows API key management experience Telegram bot creation knowledge How to set up Step 1: Service Configuration Create Telegram Bot Message @BotFather on Telegram Use /newbot command and follow instructions Save the bot token for later use OpenRouter Setup Sign up at openrouter.ai Purchase credits for Gemini 2.5 Flash access Generate and save API key Anthropic Configuration Create account at console.anthropic.com Add credits to your account Generate Claude API key KIE.AI Access Register at kie.ai Subscribe to VEO3 Fast plan Obtain bearer token Step 2: N8N Credential Setup Configure these credentials in your n8n instance: Telegram API Credential Name: telegramApi Bot Token: Your Telegram bot token OpenRouter API Credential Name: openRouterApi API Key: Your OpenRouter key Anthropic API Credential Name: anthropicApi API Key: Your Anthropic key HTTP Bearer Auth Credential Name: httpBearerAuth Token: Your KIE.AI bearer token Step 3: Workflow Configuration Import the Workflow Copy the provided JSON workflow Import into your n8n instance Update Telegram Token Locate the "Edit Fields" node Replace "Your Telegram Token" with your actual bot token Configure Webhook URLs Ensure all Telegram nodes have proper webhook configurations Test webhook connectivity Step 4: Testing & Validation Test Individual Nodes Verify each API connection Check credential configurations Confirm node responses End-to-End Testing Send a test image to your Telegram bot Follow the complete workflow process Verify final video output How to customize the workflow Modify Image Enhancement Prompts Edit the HTTP Request node for Gemini Adjust the prompt text to match your style preferences Test different aspect ratios (current: 1:1 square format) Customize Script Generation Modify the Basic LLM Chain node prompt Adjust video segment duration (current: 7-8 seconds each) Change dialogue style and tone requirements Video Generation Settings Update VEO3 API parameters in HTTP Request1 node Modify aspect ratio (current: 16:9) Adjust model settings and seeds for consistency Output Customization Change final video format in MediaFX node Modify Telegram message templates Add additional processing steps before delivery Workflow Operation Phase 1: Image Reception and Enhancement User sends product image via Telegram System prompts for enhancement instructions Gemini AI analyzes and optimizes image Enhanced square-format image returned Phase 2: Analysis and Script Creation System requests dialogue concept from user AI analyzes image details and environment Claude generates realistic 2-segment script Scripts respect physical constraints of original image Phase 3: Video Generation Two separate videos generated using VEO3 System monitors generation status Videos merged into single flowing sequence Final video delivered via Telegram Troubleshooting Common Issues API Rate Limits**: Implement delays between requests Webhook Failures**: Verify URL configurations and SSL certificates Video Generation Timeouts**: Increase wait node duration Credential Errors**: Double-check all API keys and permissions Error Handling The workflow includes automatic error detection: Failed video generation triggers error message Status checking prevents infinite loops Alternative outputs for different scenarios Advanced Features Batch Processing Modify trigger to handle multiple images Add queue management for high-volume usage Implement user session tracking Custom Branding Add watermarks or logos to generated videos Customize color schemes and styling Include brand-specific dialogue templates Analytics Integration Track usage metrics and success rates Monitor API costs and optimization opportunities Implement user behavior analytics Cost Optimization API Usage Management Monitor token consumption across services Implement caching for repeated requests Use lower-cost models for testing phases Efficiency Improvements Optimize image sizes before processing Implement smart retry mechanisms Use batch processing where possible This workflow transforms static product images into engaging, professional UGC videos automatically, saving hours of manual video creation while maintaining high quality output perfect for social media platforms.
by Olaf Titel
Setup & Instructions — fluidX: Create Session, Analyze & Notify Goal: This workflow demonstrates the full fluidX THE EYE integration — starting a live session, inviting both the customer (via SMS) and the service agent (via email), and then accessing the media (photos and videos) created during the session. Captured images are automatically analyzed with AI, uploaded to an external storage (such as Google Drive), and a media summary for the session is generated at the end. The agent receives an email with a link to join the live session. The customer receives an SMS with a link to start sharing their camera. Once both are connected, the agent can view the live feed, and the system automatically stores uploaded images and videos in Google Drive. When the session ends, the workflow collects all media and creates a complete AI-powered session summary (stored and updated in Google Drive). Below is an example screenshot from the customer’s phone: Prerequisites Developer account:* https://live.fluidx.digital (activate the *TEST plan**, €0) API docs (Swagger):** fluidX.digital API 🔐 Required Credentials 1️⃣ fluidX API key (HTTP Header Auth) • Credential name in n8n: fluidx API key • Header name: x-api-key • Header value: YOUR_API_KEY 2️⃣ SMTP account (for outbound email) • Credential name in n8n: SMTP account • Configure host, port, username, and password according to your provider • Enable TLS/SSL as required 3️⃣ Google Drive account • Used to store photos, videos, and automatically update the session summary files. 4️⃣ OpenAI API (for AI analysis & summary) •Used in the Analyze Images (AI) and Generate Summary parts of the workflow. • Credential type: OpenAI • Credential name (suggested): OpenAI account • API Key: your OpenAI API key • Model: e.g. gpt-4.1, gpt-4o, or similar (choose in the OpenAI node settings) ⚙️ Configuration (in the “Set Config” node) BASE_URL: https://live.fluidx.digital company / project / billingcode / sku: adjust as needed emailAgent: set before running (empty in template) phoneNumberUser: set before running (empty in template) Flow Overview Form Trigger → Create Session → Set Session Vars → Send SMS (User) → Send Email (Agent) → Monitor Media → Analyze Images (AI) → Upload Files to Google Drive → Generate Summary → Update Summary File The workflow starts automatically when a Form submission is received. Users enter the customer’s phone number and agent’s email, and the system creates a new fluidX THE EYE session. As media is uploaded during the session, the workflow automatically retrieves, stores, analyzes, and summarizes it — providing a complete end-to-end automation example for remote inspection, support, or field-service use cases. Notes Do not store real personal data inside the template. Manage API keys and secrets via n8n Credentials or environment variables. Log out of https://live.fluidx.digital in the agent’s browser before testing, to ensure a clean invite flow and session creation.
by Automate With Marc
🎨 Instagram Carousel & Caption Generator on Autopilot (GPT-5 + Nano Banana + Blotato + Google Sheets) Description Watch the full step-by-step tutorial on YouTube: https://youtu.be/id22R7iBTjo Disclaimer (self-hosted requirement): This template assumes you have valid API credentials for OpenAI, Wavespeed/Nano Banana, Blotato, and Google. If using n8n Self-Hosted, ensure HTTPS access and credentials are set in your instance. How It Works Chat Trigger – Receive a topic/idea (e.g. “5 best podcast tips”). Image Prompt Generator (GPT-5) – Creates 5 prompts using the “Hook → Problem → Insight → Solution → CTA” framework. Structured Output Parser – Formats output into a JSON array. Generate Images (Nano Banana) – Converts prompts into high-quality visuals. Wait for Render – Ensures image generation completes. Fetch Rendered Image URLs – Retrieves image links. Upload to Blotato – Hosts and prepares images for posting. Collect Media URLs – Gathers all uploaded image URLs. Log to Google Sheets – Stores image URLs + timestamps for tracking. Caption Generator (GPT-5) – Writes an SEO-friendly caption. Merge Caption + Images – Combines data. Post Carousel (Blotato) – Publishes directly to Instagram. Step-by-Step Setup Instructions 1) Prerequisites n8n (Cloud or Self-Hosted) OpenAI API Key (GPT-5) Wavespeed API Key (Nano Banana) Blotato API credentials (connected to Instagram) Google Sheets OAuth credentials 2) Add Credentials in n8n OpenAI: Settings → Credentials → Add “OpenAI API” Wavespeed: HTTP Header Auth (e.g. Authorization: Bearer <API_KEY>) Blotato: Add “Blotato API” Google Sheets: Add “Google Sheets OAuth2 API” 3) Configure & Test Run with an idea like “Top 5 design hacks”. Check generated images, caption, and logged sheet entry. Confirm posting works via Blotato. 4) Optional Add a Schedule Trigger for weekly automation. Insert a Slack approval loop before posting. Customization Guide ✏️ Change design style: Modify adjectives in the Image Prompt Generator. 📑 Adjust number of slides: Change Split node loop count. 💬 Tone of captions: Edit Caption Generator’s system prompt. ⏱️ Adjust render wait time: If image generation takes longer, increase the Wait node duration from 30 seconds to 60 seconds or more. 🗂️ Log extra data: Add columns in Google Sheets for campaign or topic. 🔁 Swap posting tool: Replace Blotato with your scheduler or email node. Requirements OpenAI API key (GPT-5 or compatible) Wavespeed API key (Nano Banana) Blotato API credentials Google Sheets OAuth credentials n8n account (Cloud or Self-Hosted)
by SOLOVIEVA ANNA
Who this is for Users who frequently receive images or documents via LINE or email Teams needing automatic OCR + AI summarization Anyone who wants hands-free document processing and structured storage How it works Triggers: LINE Webhook and Gmail IMAP Trigger capture incoming messages or emails. Source Tagging: Inputs are tagged as LINE or EMAIL for later branching. File Handling: Files are uploaded to Google Drive and converted for analysis. OCR: An AI vision model extracts all readable text from the document image. AI Summarization: A text model produces a concise summary. Logging: The summary is appended to Google Sheets for record-keeping. Email Drafting: A Gmail Draft is generated containing the OCR text and summary. How to set up Connect your LINE, Gmail, OpenAI, and Google Drive/Sheets credentials. Update folder IDs, sheet names, and authentication fields as needed. Optional: customize summarization instructions. Customization ideas Add translation or classification steps Modify output format for Slack/Notion Store files in date-based Drive folders
by Marco Florez
Turn your code commits into engaging social media content automatically. This workflow monitors a GitHub repository, uses AI to write a LinkedIn post about your changes, generates a beautiful "Mac-window" style image of your code, and publishes it all to LinkedIn. How it works GitHub Trigger: Watches for new push events in your selected repository. AI Analysis: Passes the code changes to an LLM (via LangChain) to write a professional LinkedIn post and select the best code snippet. Image Generation: Creates a custom HTML view of your code (with syntax highlighting and window controls) and converts it to an image using the HCTI API. Hosting & Posting: Uploads the generated image back to GitHub for hosting, then combines the text and image to publish a live post on LinkedIn. Set up steps Configure Credentials: You will need credentials for: GitHub (OAuth2 or Access Token) LinkedIn (OAuth2) OpenRouter (or swap the model node for OpenAI/Anthropic) HCTI.io (for the HTML-to-Image conversion) Update GitHub Nodes: In the Trigger node: Set your Owner and Repository. In the File Download node: Set the same Owner and Repository. In the Upload Image node: Set the target repo where you want images stored. Update LinkedIn Node: Add your LinkedIn Person URN in the Person field.