by Eduard
Create, iterate, and share! Transform a single image through multiple scenes while maintaining consistency. ✨ What this workflow does This template showcases FLUX.1 Kontext - Black Forest Labs' in-context image generation model that excels at maintaining character features across multiple transformations. Combined with the Upload Post community node for effortless multi-platform social media posting, you can create and share compelling visual stories instantly. The workflow demonstrates FLUX Kontext's core strength: character consistency across multiple image generations. Starting with a single input image, it: 🖼️ Loads an initial character image (example: a cute animal mascot) 📝 Defines multiple scene transformation prompts 🔄 Iteratively generates new scenes while preserving exact character features 🎯 Maintains visual consistency by reusing binary data from previous generations 📱 Auto-posts the complete transformation series to multiple social platforms simultaneously 🚀 Key Features: The Consistency Advantage Character Preservation**: FLUX Kontext's signature feature - maintains character features and style across transformations (requires specific prompting techniques) Iterative Context Building**: Each generation uses the previous image as context, creating visual continuity Binary Data Reuse**: Smart workflow design that feeds output from one generation as input to the next Multi-Scene Storytelling**: Transform your character across different environments while keeping them recognizable One-Click Multi-Platform Posting*: Upload Post eliminates the tedious process of posting to each platform individually 📱 Why use Upload Post? Posting the same content to TikTok, Instagram, LinkedIn, YouTube, Facebook, X (Twitter), and Threads individually is time-consuming and error-prone. The Upload Post service* simplifies this process: ✅ Connect once, post everywhere: Link all your social media accounts to Upload Post ✅ Single API call: Post to multiple platforms with one simple node ✅ No more platform juggling: Skip the endless switching between apps and dashboards ✅ Consistent timing: All platforms get your content simultaneously ✅ Trusted by 3,751+ users: Proven solution for content creators and marketers Instead of spending 30+ minutes manually posting to each platform, Upload Post does it all in seconds with a single n8n node! 🛠️ Prerequisites Required Accounts: Black Forest Labs API: Create account at dashboard.bfl.ai Get your API key for FLUX Kontext Pro access Upload Post Account: Sign up at upload-post.com* Connect your social media profiles (TikTok, Instagram, LinkedIn, YouTube, Facebook, X/Twitter, Threads) Get API credentials for automated posting Free tier available: 10 uploads/month 💡 Perfect For: Character Designers** maintaining brand character integrity across scenes Social Media Managers** creating engaging visual story series without manual posting Brand Marketers** ensuring character consistency across campaigns Storytellers** building visual narratives with consistent protagonists Agencies** managing multiple client accounts efficiently 🔧 Customization Options: Modify transformation prompts** to create your own character journey Adjust iteration steps** Change initial character image** Configure social platform targeting** (choose which platforms to post to) Customize post content** and formatting Experiment with different consistency scenarios** \ Affiliate link*
by Don Jayamaha Jr
A professional-grade AI automation system for spot market trading insights on Binance. It analyzes multi-timeframe technical indicators, live price/order data, and crypto sentiment, then delivers fully formatted Telegram-style trading reports. 🎥 Watch Tutorial: 🧩 Required Workflows You must install and activate all of the following workflows for the system to function correctly: | ✅ Workflow Name | 📌 Function Description | | -------------------------------------------------- | -------------------------------------------------------------------------------- | | Binance Spot Market Quant AI Agent | Final AI orchestrator. Parses user prompt and generates Telegram-ready reports. | | Binance SM Financial Analyst Tool | Calls indicator tools and price/order data tools. Synthesizes structured inputs. | | Binance SM News and Sentiment Analyst Webhook Tool | Analyzes crypto sentiment, gives summary and headlines via POST webhook. | | Binance SM Price/24hrStats/OrderBook/Kline Tool | Pulls price, order book, 24h stats, and OHLCV klines for 15m–1d. | | Binance SM 15min Indicators Tool | Calculates 15m RSI, MACD, BBANDS, ADX, SMA/EMA from Binance kline data. | | Binance SM 1hour Indicators Tool | Same as above but for 1h timeframe. | | Binance SM 4hour Indicators Tool | Same as above but for 4h timeframe. | | Binance SM 1day Indicators Tool | Same as above but for 1d timeframe. | | Binance SM Indicators Webhook Tool | Technical backend. Handles all webhook logic for each timeframe tool. | ⚙️ Installation Instructions Step 1: Import Workflows Open your n8n Editor UI Import each workflow JSON file one by one Activate them or ensure they're called via Execute Workflow Step 2: Set Credentials OpenAI API Key** (GPT-4o recommended) Binance endpoints** are public (no auth required) Step 3: Configure Webhook Endpoints Deploy Binance SM Indicators Webhook Tool Ensure the following paths are reachable: /webhook/15m /webhook/1h /webhook/4h /webhook/1d Step 4: Telegram Integration Create a Telegram bot using @BotFather Add your Telegram API token to n8n credentials Replace the Telegram ID placeholder with your own Step 5: Final Trigger Trigger the Binance Spot Market Quant AI Agent manually or from Telegram The agent: Extracts the trading pair (e.g. BTCUSDT) Calls all tools for market data and sentiment Generates a clean, HTML-formatted Telegram report 💬 Telegram Report Output Format BTCUSDT Market Report Spot Strategy • Action: Buy • Entry: $63,800 | SL: $61,200 | TP: $66,500 • Rationale: MACD Crossover (1h) RSI Rebound from Oversold (15m) Sentiment: Bullish Leverage Strategy • Position: Long 3x • Entry: $63,800 • SL/TP zones same as above News Sentiment: Slightly Bullish • "Bitcoin rallies as ETF inflows surge" – CoinDesk • "Whales accumulate BTC at key support" – NewsBTC 🧠 System Overview [Telegram Trigger] → [Session + Auth Logic] → [Binance Spot Market Quant AI Agent] → [Financial Analyst Tool + News Tool] → [All Technical Indicator Tools (15m, 1h, 4h, 1d)] → [OrderBook/Price/Kline Fetcher] → [GPT-4o Reasoning] → [Split & Send Message to Telegram] 🧾 Licensing & Attribution © 2025 Treasurium Capital Limited Company Architecture, prompts, and trade report structure are IP-protected. No unauthorized rebranding or resale permitted. 🔗 For support: LinkedIn – Don Jayamaha
by Harshil Agrawal
Based on your use case, you might want to trigger a workflow if new data gets added to your database. This workflow allows you to send a message to Mattermost when new data gets added in Google Sheets. The Interval node triggers the workflow every 45 minutes. You can modify the timing based on your use case. You can even use the Cron node to trigger the workflow. If you wish to fetch new Tweets from Twitter, replace the Google Sheet node with the respective node. Update the Function node accordingly.
by Harshil Agrawal
This workflow allows you to get analytics of a website and store it Airtable. In this workflow, we get the analytics for the sessions grouped by the country. Based on your use-case, you can select different Dimensions and set different Metrics. You can use the Cron node or the Interval node to trigger the workflow on a particular interval and fetch the analytics data regularly. Based on your use-case, you might want to store the data returned by Google Analytics to a database or a Google Sheet. Replace the Airtable node with the appropriate node.
by Danielle Gomes
This n8n workflow collects and summarizes news from multiple RSS feeds, using OpenAI to generate a concise summary that can be sent to WhatsApp or other destinations. Perfect for automating your daily news digest. 🔁 Workflow Breakdown: Schedule Trigger Start the workflow on your desired schedule (daily, hourly, etc.). 🟨 Note: Set the trigger however you wish. RSS Feeds (My RSS 01–04) Fetches articles from four different RSS sources. 🟨 Note: You can add as many RSS feeds as you want. Edit Fields (Edit Fields1–3) Normalizes RSS fields (title, link, etc.) to ensure consistency across different sources. Merge (append mode) Combines the RSS items into a single unified list. Filter Optionally filter articles by keywords, date, or categories. Limit Limits the analysis to the 10 most recent articles. 🟨 Note: This keeps the result concise and avoids overloading the summary. Aggregate Prepares the selected news for summarization by combining them into a single content block. OpenAI (Message Assistant) Summarizes the aggregated news items in a clean and readable format using AI. Send Summary to WhatsApp Sends the AI-generated summary to a WhatsApp endpoint via webhook (yoururlapi.com). You can replace this with an email service, Google Drive, or any other destination. 🟨 Note: You can send it to your WhatsApp API, email, drive, etc. No Operation (End) Final placeholder to safely close the workflow. You may expand from here if needed.
by Evoort Solutions
AI-Powered Product Research & SEO Content Automation Skip the guesswork and manual effort — this n8n flow automates the entire process of researching your product's online competition and generating high-quality SEO content. Whether you're launching a new product or optimizing existing listings, this workflow leverages real-time web data and AI-driven copywriting to deliver: 📈 Search-optimized metadata (Title, Description, Keywords) 🛍️ Engaging product descriptions tailored for marketing 📊 Auto-organized output ready for use in your content or e-commerce platform All of this happens with just one product title input! 🧠 How It Works • User submits a product title via a form. • The workflow uses Google Custom Search to gather real-time competitor content based on that title. • Titles, snippets, and keywords are extracted from the search results. • This information is sent to a language model (Google Gemini via LangChain) to generate: SEO-optimized metadata (Title, Description, Keywords) A compelling product description tailored for marketing • The AI-generated content is then parsed and organized into two categories: SEO data and product content. • The structured output is saved automatically into a connected Google Sheet for easy access or further automation. 🛠️ What Problems Does This Solve? Manual competitor research and writing SEO content from scratch can be: Time-consuming** Inconsistent in quality** Not optimized for search engines** Hard to scale for multiple products** This workflow automates the entire research + writing + structuring process. ✅ Key Benefits Instant Content Creation**: Generate polished SEO content in seconds. Competitor-Aware**: Pulls in real-time data from the web for relevant, market-aligned content. Scalable**: Easily repeat the process for multiple product titles with minimal effort. Data Centralization**: Stores everything in Google Sheets—great for collaboration or syncing with other tools. Customizable**: Easily extend or modify the workflow to include translations, publishing, or social media automation. ⚙️ Set-Up Steps • Connect Google Custom Search API with a valid API key and search engine ID (CX). • Connect and configure Google Gemini or LangChain with access credentials. • Provide access to a Google Sheet with columns for storing SEO and product data. • Estimated setup time: ~15–25 minutes depending on API access and sheet setup. 🚀 Let’s Get You Started with Automating Your LinkedIn Posts! Create your free n8n account and set up the workflow in just a few minutes using the link below: 👉 Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!
by Ludwig
How It Works: • Scrapes company review data from Glassdoor using ScrapingBee. • Extracts demographic-based ratings using AI-powered text analysis. • Calculates workplace disparities with statistical measures like z-scores, effect sizes, and p-values. • Generates visualizations (scatter plots, bar charts) to highlight patterns of discrimination or bias. Example Visualizations: Set Up Steps: Estimated time: ~20 minutes. • Replace ScrapingBee and OpenAI credentials with your own. • Input the company name you want to analyze (best results with large U.S.-based organizations). • Run the workflow and review the AI-generated insights and visual reports. This workflow empowers users to identify potential workplace discrimination trends, helping advocate for greater equity and accountability. Additional Credit: Wes Medford For algorithms and inspiration
by Shashikanth
Source code, I maintain this worflow here. Usage Guide This workflow backs up all workflows as JSON files named in the [workflow_name].json format. Steps Create GitHub Repository Skip this step if using an existing repository. Add GitHub Credentials In Credentials, add the GitHub credential for the repository owner. Download and Import Workflow Import this workflow into n8n. Set Global Values In the Globals node, set the following: repo.owner: GitHub username of the repository owner. repo.name: Name of the repository for backups. repo.path: Path to the folder within the repository where workflows will be saved. Configure GitHub Nodes Edit each GitHub node in the workflow to use the added credentials. Workflow Logic Each workflow run handles files based on their status: New Workflow If a workflow is new, create a new file in the repository. Unchanged Workflow If the workflow is unchanged, skip to the next item. Changed Workflow If a workflow has changes, update the corresponding file in the repository. Current Limitations / Needs work Name Change of Workflows If a workflow is renamed or deleted in n8n, the old file remains in the repository. Deleted Workflows Deleted workflows in n8n are not removed from the repository.
by Muzaffer AKYIL
Docker Registry Cleanup Template This template is designed to automatically clean up old image tags in the Docker registry and perform garbage collection. Features List all images in the registry Preserve the last 10 tags for each image (latest tag is always preserved) Delete old tags Email notification for Successful/Excused cancellation Registry garbage collection automation Failure notification in error conditions Prerequisites Docker Registry v2 API access Basic Authentication credentials SMTP email settings (for notifications) SSH node installed on n8n (for garbage collection) Installation 1. Identity Information Add the following credentials in n8n: HTTP Basic Auth**: For Registry access SSH Private Key**: For Garbage collection command Email SMTP**: For notifications 2. Set Variables Replace your-registry-url with your actual registry URL on all nodes: ‘url": ‘https://your.registry.com/v2/_catalog’. Customisation Retention Policy: Set the number of tags to be retained by changing the slice(0, 10) value in the Identify Tags to Remove node Schedule: Change the frequency of operation at the Trigger node Notification Content: Customise email templates according to your needs Notes Check DELETE operations before running in a test environment Make sure that the registry is not in read-only mode The registry may need to be put into maintenance mode for garbage collection Step Details: Retrieving image information:** The workflow starts by fetching a list of images and their associated tags from the Docker registry. Filtering and sorting:** The retrieved tags are then filtered and sorted based on specific criteria, such as creation date and tag name. Deleting old tags:** The workflow identifies old or unused tags and attempts to delete them from the registry. Sending notifications:** The workflow sends email notifications to inform the user about the status of the cleanup process, including any errors or successes. Executing additional cleanup tasks:** Finally, the workflow executes an SSH command on the Docker registry server to perform additional cleanup tasks, such as garbage collection. TL;DR In summary, this n8n template provides a robust and automated solution for managing and cleaning up Docker registries. By regularly running this workflow, users can ensure that their registry remains organized and efficient, and avoid running out of storage space.-
by Łukasz
What Is This? This workflow is a comprehensive solution for automating website audits and optimizations, leveraging advanced technologies to boost SEO effectiveness and overall site performance. Who Is It For? Designed for SEO specialists, digital marketers, webmasters, and content teams, this workflow empowers anyone responsible for website performance to automate and scale their audit processes. Agencies managing multiple client sites, in-house SEO teams aiming to save time on routine checks, and developers seeking to integrate data-driven insights into their deployment pipelines will all find this solution invaluable. By combining your site’s sitemap with Google Search Console and Google Analytics data, then applying AI-powered analysis, the workflow continuously uncovers actionable recommendations to boost search visibility, improve user engagement, and accelerate page performance. Whether you manage a single blog or oversee a sprawling e-commerce platform, this automated pipeline delivers precise, prioritized SEO improvements without manual data wrangling. How Does It Work? This end-to-end site analysis automation consists of five main stages: 1. URL Discovery Processes the sitemap.xml using HTTP Request and XML nodes to extract all site URLs. 2. Search Console Performance Analysis Uses the Google Search Console API to fetch detailed metrics for each page, including search position, clicks, impressions, and CTR. 3. Analytics Data Collection Connects to the Google Analytics API to automatically retrieve traffic metrics such as pageviews, average session duration, bounce rate, and conversions. 4. AI Data Processing Employs OpenAI models to perform in-depth analysis of the collected data. The artificial intelligence engine merges insights from all sources, identifies patterns, and produces detailed optimization recommendations. AI analyses website itsefl aswell. Consider testing different models. I do recommend at least trying out o4-mini. 5. Recommendation Generation Creates tailored suggestions for each page, in form of HTML table, that is being sent to your email. How To Set It Up? Accounts: An active n8n account or instance, API keys for Google Search Console and Google Analytics, an OpenAI access token. Enabled Google APIs: You will neeed at least following scopes: Google Search Console API Google Analytics Aadmin API Google Analytics Data API Scheduling: The workflow can run manually for ad hoc audits or be scheduled (daily, weekly) for continuous site monitoring. Testing: There are two nodes that are optional: "Sort for testing purposes" and "Limit for testing purposes" Together they randomly select items from sitemap and limit them to few so you don't need to run hundreds of sitemap.xml items at once, but you can run just a random batch first. Globals: There is node called "Globals- CHANGE ME!". You need to set up proper variables in there, which are: sitemap_url - self exlpainatory search_console_selector - for example "sc-domain:sailingbyte.com" but can be URL aswell- depends on how did you set up your search console analysis_start_date and analysis_end_date - date range for analytics, by default last 30 days analytics_selector_id - ID of Google Analytics setup, it is a large integer, you can find it in analytics url preceeded with letter "p", ex (your number is where there are X's): https://analytics.google.com/analytics/web/#/pXXXXXXXXX/reports/intelligenthome report_receiver - email which will receive report What's More? That's actually it. I hope that this automation will help your website improvement will be much easier! Thank you, perfect! Glad I could help. Visit my profile for other automations for businesses. And if you are looking for dedicated software development, do not hesitate to reach out!
by Davide
Functionality This workflow automates the handling of incoming emails by summarizing their content, generating appropriate responses, and validating the responses through a "Human-in-the-Loop" system. It integrates with IMAP email services (e.g., Gmail, Outlook) and uses AI models to streamline the email response process. The workflow ensures that all AI-generated responses are reviewed by a human before being sent, maintaining a high level of professionalism and accuracy. This approach is particularly useful for businesses that receive a high volume of emails and need to respond quickly while ensuring quality control. How It Works Email Trigger: The workflow starts with the Email Trigger (IMAP) node, which monitors an email inbox for new messages. When a new email arrives, it triggers the workflow. Email Preprocessing: The Markdown node converts the email's HTML content into plain text for easier processing by the AI models. Email Summarization: The Email Summarization Chain node uses an AI model (OpenAI) to generate a concise summary of the email. The summary is limited to 100 words and is written in a professional tone. Email Response Generation: The Write email node uses an AI model (OpenAI) to draft a professional response to the email. The response is based on the email content and is limited to 100 words. Human-in-the-Loop Approval: The Set Email text node prepares the drafted response for approval. The Approve Email node sends the drafted response to a human approver (e.g., an internal email address) for review. The email includes: The original message. The AI-generated response. The Approved? node checks if the response has been approved by the human reviewer. If approved, the workflow proceeds to send the response; otherwise, it stops. Sending the Response: The Send Email node sends the approved response back to the original sender. Key Features Automated Email Summarization**: Summarizes incoming emails to provide a quick overview of the content. AI-Powered Response Generation**: Drafts professional responses to emails using AI. Human-in-the-Loop Approval**: Ensures all AI-generated responses are reviewed and approved by a human before being sent. IMAP Integration**: Works with IMAP email services like Gmail and Outlook. Efficient Email Management**: Reduces the time and effort required to handle incoming emails while maintaining high-quality responses. This workflow is ideal for businesses looking to automate their email response process while maintaining control over the quality of outgoing communications. It leverages AI to handle repetitive tasks and ensures that all responses are reviewed by a human, providing a balance between automation and human oversight. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Cheng Siong Chin
How It Works This workflow automates enterprise compliance governance using a multi-agent AI architecture. It targets compliance officers, legal teams, and risk managers who need continuous, jurisdiction-aware monitoring without manual intervention.The workflow runs on two paths: a scheduled governance check and an event-driven compliance signal receiver. Both converge on a central Governance Agent that orchestrates four specialised sub-agents, Compliance Signal, Jurisdiction Analysis, Remediation Planning, and Audit Documentation, each backed by dedicated AI models. External tools handle API lookups, penalty calculations, escalation notifications, and structured output parsing. Results feed into immutable audit logs, merged with scheduled reports, and dispatched as a daily email summary. Escalation logic ensures critical violations trigger immediate alerts. This eliminates fragmented compliance tracking, reduces human error, and provides auditable, timestamped records across multiple regulatory jurisdictions. Setup Steps Add Claude/OpenAI credentials to all Chat Model nodes. Configure the External Compliance API Tool with your regulatory API key and endpoint. Set the Compliance Data Calculation Tool parameters (jurisdiction codes, penalty thresholds). Connect the Escalation Notification Tool to your email or Slack credentials. Set the Periodic Compliance Check schedule (cron) to your audit cadence. Prerequisites External compliance/regulatory API access Email or Slack account for notifications n8n instance (v1.40+ recommended) Use Cases Automated GDPR/MAS/SOX compliance monitoring across jurisdictions Customization Swap AI models per agent for cost optimisation. Benefits Eliminates manual compliance checks with autonomous multi-agent orchestration.