by Yaron Been
This workflow provides automated access to the Stability Ai Stable Video Diffusion AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for video generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete video generation process using the Stability Ai Stable Video Diffusion model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model by stability-ai for automated processing tasks. Key Capabilities AI-powered video generation and processing** High-quality video synthesis from inputs** Advanced video manipulation capabilities** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Stability Ai/stable-video-diffusion AI model Stability Ai Stable Video Diffusion**: The core AI model for video generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Video Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Video Content Creation**: Generate videos for social media, marketing, and presentations Animation & Motion Graphics**: Create animated content and visual effects Video Editing**: Enhance and transform existing video content Educational Content**: Produce instructional and explainer videos Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #videogeneration #aivideo #videoai #motion #videoautomation #videocreation #stablediffusion #diffusion #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by WeblineIndia
This smart automation workflow created by the AI development team at WeblineIndia, helps with the daily collection and storage of weather data. Using the OpenWeatherMap API and Airtable, this solution gathers vital weather details such as temperature, humidity, and wind speed. The automation ensures daily updates, creating a dependable historical record of weather patterns for future reference and analysis. Steps: Set Schedule Trigger Configure a Cron node to trigger the workflow daily, for example, at 7 AM. Fetch Weather Data (HTTP Request) Use the HTTP Request node to retrieve weather data from the OpenWeatherMap API. Include your API key and query parameters (e.g., q=London, unit=metric) to specify the city and desired units. Parse Weather Data Utilize a JSON Parse node to extract key weather details, such as temperature, humidity, and wind speed, from the API response. Store Data in Airtable Use the Airtable node to insert the parsed data into the designated Airtable table. Ensure proper mapping of fields like temperature, humidity, and wind speed. Save and Execute Save the workflow and activate it to ensure weather data is fetched and stored automatically every day. Outcome This robust solution, developed by WeblineIndia, reliably collects and archives daily weather data, providing businesses and individuals with an accessible record of weather trends for analysis and decision-making. About WeblineIndia We specialize in creating custom automation solutions and innovative software workflows to help businesses streamline operations and achieve efficiency. This weather data fetcher is just one example of our expertise in delivering value through technology.
by Klaasjan te Voortwis
This n8n workflow template uses community nodes and is only compatible with the self-hosted version of n8n. Export workflows with readable names, tagged for different environments To ensure understandable workflow exports, ease of use in delivery pipelines, and a better developer experience, this workflow helps with exporting workflows. Inner workings First, the workflow ensures that the directory structure for storing the workflows is correct. Exports all workflows. Next, it processes all workflow files and stores them with readable names. Based on tags, it will also export to dev and prod folders for easy commit and usage in a delivery pipeline. Configration No special setup is required for readable exporting. Usage Create a workflow and tag it with 'Auto deploy to dev' Run the workflow, this will create the needed folders and workflows with readable names. Commit these in your version control. Have a CICD pipeline build an n8n container —see the attached Dockerfile. Check our Auto Starter workflow for auto-starting workflows after deployment. CI/CD Bonus: Attached are two nodes with some example configuration on building your own automated n8n deployment. A Dockerfile, to get the new entrypoint and exported workflows packaged in the container. An updated entrypoint to build your own container, import the workflows, and run the Auto Starter. Set the following environment variables: STARTUP_WORKFLOWS_LOAD_LOCATION: to specify the folder to import from and distinguish between environments. STARTUP_WORKFLOW_ID: the ID of the workflow to run after starting n8n. > Note: The 'Instance Started' n8n trigger won't work, as all workflows are disabled upon import.
by Yang
This workflow turns trending news into thoughtful first-person commentary for platforms like LinkedIn. It uses Dumpling AI’s News Search and Scraping APIs to find and extract article content, then feeds the cleaned text to GPT-4o to write personalized insights. The final output is saved back to Google Sheets as a draft for easy review or posting. ✅ What this workflow does Triggers daily using a Schedule node. Fetches a list of content topics from a Google Sheet. Uses Dumpling AI to search for relevant news articles based on each topic. Scrapes the article content with Dumpling AI’s /scrape endpoint. Cleans and aggregates the article content using a Code node. Generates first-person commentary with GPT-4o tailored for LinkedIn. Appends the generated post back to the Google Sheet next to its topic. 🧩 Nodes in this workflow Schedule Trigger**: Starts the workflow daily. Google Sheets (Read Topics)**: Pulls topic rows that don’t have a generated commentary yet. Split In Batches**: Processes each topic one at a time. Wait**: Adds a delay to manage API limits. HTTP Request (Search News)**: Uses Dumpling AI's /search-news to find relevant articles for the topic. Split Out**: Iterates over the list of article results. HTTP Request (Scrape Article)**: Extracts the full article text using Dumpling AI’s /scrape. Aggregate**: Collects and merges article content fields. Code (Clean Article)**: Strips links, formatting, and irrelevant text. OpenAI (GPT-4o)**: Generates a short, first-person LinkedIn post-style commentary using a custom prompt. Google Sheets (Write Back)**: Appends the final output next to the original topic in the sheet. 🧑💼 Who is this for? Founders, content creators, marketers, or **agency teams looking to maintain an active presence on LinkedIn or newsletters by sharing smart takes on industry trends. 💡 What problem does this solve? Most people want to comment on current events but don't have the time to summarize articles or write well-structured posts. This automation saves hours of manual work by: Finding the right article. Extracting and cleaning the content. Writing it in a natural, first-person voice using AI. ⚙️ What you need to use this: A Google Sheet with at least two columns: topic and generated commentary. A Dumpling AI API Key with access to the /search-news and /scrape endpoints. An OpenAI GPT-4o connection.
by Yang
🔎 Who is this for? This workflow is designed for podcast creators, content marketers, and video producers who want to convert YouTube videos into podcast-ready scripts. It's perfect for anyone repurposing long-form content to reach audio-first audiences without manual effort. 🧠 What problem is this workflow solving? Creating podcast scripts from YouTube videos manually is time-consuming. This workflow automates the process by pulling transcripts, cleaning the text, organizing the dialogue, summarizing the key points, and saving everything in one place. It removes the need for manual transcription, formatting, and structuring. ⚙️ What this workflow does This workflow uses Dumpling AI and GPT-4o to automate the transformation of YouTube video transcripts into polished podcast scripts. Here's how it works: RSS Feed Trigger Monitors a YouTube RSS feed for new video uploads. When a new video is detected, the workflow begins automatically. Get YouTube Transcript (Dumpling AI) Uses Dumpling AI's get-youtube-transcript endpoint to extract the full transcript from the video URL. Generate Podcast Script with GPT-4o GPT-4o receives the transcript and generates a structured JSON output including: Cleaned transcript with filler words removed Speaker labels for clarity A short, engaging podcast title A concise summary of the episode Save to Airtable The structured data (title, summary, cleaned transcript) is saved to Airtable for easy review, editing, or publishing. This automation is an ideal workflow for repurposing video content into audio-friendly formats, cutting down production time while increasing content output across platforms.
by Joseph LePage
From PDF to Powerful Blog Posts: AI-Powered Content Transformation Turn complex documents into engaging digital content that drives results. This n8n Workflow uses AI to transforms lengthy PDFs into compelling blog posts that attract and retain readers while you focus on strategic initiatives. Time-Saving Innovation 🚅Lightning-Fast Processing Transform lengthy documents into polished blog content in under 1 minute, eliminating hours of manual work. Our system handles the heavy lifting, delivering up to a 95% reduction in content production time. 📱Intelligent Analysis The AI engine identifies and extracts key insights, organizing information for maximum impact. Each document undergoes comprehensive analysis to ensure no valuable content is overlooked. Advanced Content Optimization ✍️Dynamic Writing Styles Possible Adjust the prompt for multiple tone options: Professional for corporate communications Conversational for engaging blogs Thought leadership for industry authority 📊SEO-Ready Content Potential Adjust the prompt to automatically optimized for search engines, incorporating relevant keywords and semantic structure to improve visibility and drive organic traffic. Ideal Applications 🤼Content Marketing Teams Scale content production without sacrificing quality or consistency. Perfect for teams looking to maintain a robust publishing schedule while maximizing resource efficiency. 🏫Academic Communication Help researchers and institutions share complex findings with broader audiences through accessible, engaging content that maintains academic integrity. 🧑💻Digital Publishers Streamline the content transformation process while ensuring each piece meets modern digital standards and reader expectations. Transform your content strategy with an intelligent system that delivers consistent, high-quality results while dramatically reducing production time.
by David Roberts
This workflow allows you to define multiple tickets/issues in a Notion page, then easily import them into Linear. Why is it useful? We use this workflow internally at n8n for collaboration between Product and Engineering teams: Engineering needs all work to be in our ticketing system (Linear) in order to keep track of it Product prefers to review features in Notion. This is because it and can be used to dump all your thoughts and organise them into themes afterwards, plus it better supports rich content like videos Features Supports rich formatting (bullets, images, videos, links, etc.) Keeps links between the Notion and Linear version, in case you need to refer back Allows you to assign each issue to a team member in the Notion definition Avoids importing the same issues twice if you run it again on the same page (meaning you can issues incrementally) You can see an example of the required format of the Notion page here.
by Mariela Slavenova
This template enriches a lead list by analyzing each contact’s company website and auto-generating a single personalized cold-email opener. Drop a spreadsheet into a Google Drive folder → the workflow parses rows, fetches website content via Jina AI, uses OpenAI to check if the site contains valid business info, then calls Anthropic to craft a one-liner. It writes both the website summary and personalized opener back to Google Sheets, and finally sends you a Telegram confirmation with the file link. What it does Turns a CSV/Google Sheet of leads into tailored cold-email openers. For each lead, the workflow fetches the company website, writes a 300-word business summary, then crafts a one-sentence, emotionally engaging opening line. Results are written back to the same Sheet, and you get a Telegram ping when processing finishes. How it works (high-level) Trigger: Watches a Google Drive folder. When a new Sheet is added, the flow starts. Parse: Reads rows (expects columns like First Name, Last Name, Email, domain). Enrich: An AI Agent calls Jina “r.jina.ai/{url}” to fetch page markdown, then produces a structured website summary. Validate: An OpenAI step checks if the fetched content is a real business page (hasWebsite: true/false). Personalize: • If true → Anthropic crafts a bespoke opener using the summary. • If false → Fallback prompt creates a strong opener using domain + universal lead-gen pains. Update: Writes websiteSummary and personalization back to the Sheet (matching on domain). Notify: Sends a Telegram message with the file name + link when done. What you need • Google Drive (folder to watch) • Google Sheets (the uploaded Sheet to enrich) • Jina HTTP header auth (for the markdown fetch tool) • OpenAI (JSON check for website validity) • Anthropic (Claude Sonnet 4 for copy quality) • Telegram Bot (to receive completion alerts) Inputs & expected schema • A Google Sheet with at least: First Name, Last Name, Email, domain • Optional columns are preserved; rows are processed in batches. Outputs • New/updated columns per row: • websiteSummary — concise, structured business overview • personalization — a single, high-impact opening sentence • Telegram confirmation with file name and link. Customization tips • Tweak the system prompts for tone or length. • Add scoring (e.g., ICP fit) before personalization. • Expand validation (e.g., handle multi-page sites or language detection). • Swap/parallel LLMs to balance quality, cost, and speed. Nodes & key logic • Google Drive Trigger → Google Drive (Download) → Spreadsheet File (parse) → Split in Batches • LangChain Agent with: HTTP Tool (Jina) + Think • OpenAI (JSON validator) → If (website present?) • Anthropic Chat (with + without website branches) • Edit Fields (Set) → Google Sheets (Update) → Telegram Great for Sales teams, SDRs, and founders who want fast, high-quality personalization at scale without manual research. Need help customizing? Contact me for consulting and support: LinkedIn
by Oneclick AI Squad
This automated n8n workflow efficiently manages the setup, creation, and deletion of PostgreSQL and MySQL databases on a Linux server, executing tasks in approximately 10 seconds. It automates installation, configuration, and user management with support for remote access. Core Elements Set Parameters** - Defines server details, database type, action, and credentials Type Check** - Confirms the selected database type PostgreSQL Action Check** - Identifies the action for PostgreSQL PostgreSQL Create Check** - Validates creation conditions for PostgreSQL Install PostgreSQL** - Sets up and configures PostgreSQL Create PostgreSQL DB** - Establishes a new PostgreSQL database with user access Delete PostgreSQL DB** - Removes a PostgreSQL database and user MySQL Action Check** - Identifies the action for MySQL MySQL Create Check** - Validates creation conditions for MySQL Install MySQL** - Sets up and configures MySQL Create MySQL DB** - Establishes a new MySQL database with user access Delete MySQL DB** - Removes a MySQL database and user Format Output** - Structures the final workflow output Getting Started Guide Import the workflow into n8n Adjust parameters in the Set Parameters node Execute the workflow Confirm the database operation on the server Necessary Requirements SSH-enabled Linux server Root-level access rights Customization Options Switch db_type between PostgreSQL and MySQL Select action (install, create, delete) via the action parameter Tailor database_name, db_user, and db_password as needed Features Install Database Server - Deploys PostgreSQL or MySQL with optimal configuration Create Database - Generates new databases with assigned users and permissions Delete Database - Eliminates databases and their associated users Parameters to Configure server_host: Your Linux server IP address server_user: SSH username (typically 'root') server_password: SSH password db_type: Select 'postgresql' or 'mysql' action: Select 'install', 'create', or 'delete' database_name: Name of the database to create or delete db_user: Database username db_password: Database password How to Use Copy the JSON code from the artifact Access your n8n workspace Choose "Import from JSON" or "+" → "From JSON" Insert the JSON code Set parameters in the "Set Parameters" node with your server information Run the workflow Workflow Actions Install: Sets up the database server, enables remote access, and initializes the database Create: Establishes a new database with a specific user Delete: Erases the database and its associated user The workflow automatically manages Ubuntu/Debian package setup Service initialization and configuration Remote access setup User and permission assignments Authentication configuration Update the parameters in the "Set Parameters" node with your server specifics and execute the workflow!
by Dr. Firas
💥 From Drawing to Story: Auto-Publish AI Video to YouTube with Blotato Overview Transform a hand-drawn character sketch into a fully animated, narrated video story — automatically. This 3-part pipeline uses Claude AI, image generation, and video synthesis to go from a simple drawing to a publish-ready video, with no manual editing required. Perfect for: indie creators, educators, storytellers, and anyone who wants to bring hand-drawn characters to life at scale. How It Works Part 1 — From Drawing to Story: Bringing Characters to Life A form submission triggers the workflow with an uploaded drawing The image is analyzed by Claude AI to extract characters and traits Character images are generated via Nano Banana (image generation API) A full story is written by Claude AI, split into scenes, and passed to Part 2 Part 2 — From Characters to Scenes: Rendering the Visual Story Character images are downloaded and converted to Base64 references Scene images are generated using Nano Banana with character consistency Scene image URLs are mapped and the video pipeline is triggered Part 3 — From Scenes to Screen: Video, Narration & Final Render Video prompts and narration context are generated by Claude AI Videos are generated via AtlasCloud (Kling Pro 3.0) with polling loop Narration audio is created with ElevenLabs and uploaded Shotstack assembles the final video with audio sync Final video is published to YouTube (and optionally TikTok) > ⚠️ Important — Workflow Structure > > This template is split into 3 separate workflows. > Each part must be imported and deployed in its own workflow in n8n. > > 📺 Watch the step-by-step tutorial to set everything up correctly: > > @youtube Requirements Credentials needed Blotato** API credentials (YouTube/TikTok publishing) AtlasCloud** API (Kling Pro 3.0 video generation) Anthropic** API key (Claude AI for story & prompts) ElevenLabs** API key (narration audio) Shotstack** API key (video assembly) Nano Banana** API key (image generation) Setup steps Configure credentials for each service above in n8n Set up a form trigger with a file upload field for the drawing Deploy the 3 workflows in order and connect them via webhooks Run a test submission with a simple sketch to validate the full pipeline 🎥 Watch This Tutorial 👋 Need help or want to customize this? 📩 Contact: LinkedIn 📺 YouTube: @DRFIRASS 🚀 Workshops: Mes Ateliers n8n Need help customizing? Contact me for consulting and support : Linkedin / Youtube / 🚀 Mes Ateliers n8n
by @OnePromptMagic
What This Workflow Does Step 1: Analyzes your recent tweets for personality & style Step 2: Generates strategic keywords based on your profile Step 3: Searches for trending tweets in your niche Step 4: Creates personalized responses & original tweets Step 5: Displays results in beautiful HTML format How to Use Execute the workflow Fill out the form with: Your X username (from your URL) Your goals on X (multiple selection) Optional additional info Submit the form Wait for processing (2-3 minutes) Double-click the "Click to show Result" node example clip can be found here
by Kurt Bijl
🎬 Social Media Content Generator Workflow Overview Automated social media content creation from video transcripts 🎯 Trigger: Airtable Webhook Action**: Receives webhook from Airtable automation Data**: RecordId and action type (e.g., "post-ig") Purpose**: Starts the content generation pipeline 📊 Step 1: Fetch Record Node**: Airtable (Get Record) Action**: Retrieves full record data using RecordId Data**: Name, transcript, and other fields 📁 Step 2: Create Google Drive Folder Node**: Google Drive (Create Folder) Action**: Creates blue folder in /tutorials directory Name**: Uses record Name field Updates**: Stores folder ID back to Airtable 🤖 Step 3: AI Content Analysis Node**: AI Agent with Google Gemini 2.5 Flash Input**: Video transcript from Airtable Structured Output**: JSON with all social formats: YouTube title & description YouTube thumbnail text Twitter thread (array) LinkedIn post Instagram caption TikTok caption YouTube Shorts caption Relevant tags 💾 Step 4: Save Transcript File Node**: Google Drive (Create from Text) Action**: Saves transcript as text file Location**: Inside the created folder Name**: Uses record Name field 📋 Step 5: Update Airtable Results Node**: Airtable (Update Record) Data**: All AI-generated social media content Special**: Twitter thread array joined with newlines 🎯 Result: Complete social media content suite ready for multi-platform publishing, organized in Google Drive with all data stored in Airtable.