by Salman Mehboob
Every time you publish a blog post on WordPress, this workflow automatically creates unique, platform-optimised captions for Facebook, Instagram, Twitter and LinkedIn — and publishes them all simultaneously without any manual effort. Built for content creators, digital marketing agencies, SEO agencies, bloggers, and businesses that publish blog content regularly and want to maximise reach across social media without spending hours manually writing different versions for each platform. The Problem This Solves Most businesses publish a blog post and then either copy and paste the same text to every social media platform — which performs poorly because each platform has different audience expectations, character limits, tone and algorithm preferences — or they skip social media promotion entirely because writing four different captions for one post takes too much time. The result is wasted content. Hours of writing a great blog post go unnoticed because the distribution is manual, inconsistent, or nonexistent. This workflow solves that completely. One WordPress publish triggers everything. The AI writes platform-specific captions that actually match how each platform works. Every post gets promoted everywhere automatically. How It Works Step 1 — WordPress Webhook Trigger The workflow starts the instant you publish a post on WordPress. This requires a webhook plugin installed on your WordPress site — the recommended free option is WP Webhooks (available on WordPress.org plugin directory). When you publish or update a post to published status, WP Webhooks sends a POST request to your n8n webhook URL with the full post data including title, content, permalink, thumbnail URL, and post type. No polling. No scheduled checks. The workflow fires in real time the moment your post goes live. Step 2 — Post Type and Status Filter Before any processing happens, the workflow checks two conditions: The post type must be post (not page, product, or custom post type) The post status must be publish (not draft, pending, or scheduled) This ensures the workflow only runs for actual published blog posts and ignores everything else — drafts, page updates, WooCommerce products, and any other WordPress content types. Step 3 — HTML to Markdown Conversion WordPress post content arrives as raw HTML with tags, shortcodes, and formatting markup. Before sending this to an AI model, the workflow converts the HTML content to clean Markdown text using the built-in Markdown node. This removes HTML tags, cleans up formatting, and gives the AI clean, readable text that produces significantly better caption quality than feeding raw HTML. Step 4 — Data Normalization A Set node extracts and organises all the data needed for the rest of the workflow into clean variables: post_title — the blog post title post_content — the cleaned Markdown content post_link — the full permalink URL to the published post post_thumbnail — the featured image URL used for visual posts on Facebook and Instagram facebook_page_id — your Facebook Page ID for the Graph API call instagram_id — your Instagram Business Account ID for the Graph API call Step 5 — AI Caption Generator An AI agent powered by OpenRouter (GPT-4o) reads the post title, link, and the first 800 characters of the post content, then generates four completely different captions — one for each platform — in a single API call. Each caption is written according to strict platform-specific guidelines: Twitter/X caption: Maximum 220 characters. One punchy insight or hook from the post. One to two hashtags only. Ends with the post link. Short, direct, and conversational. Facebook caption: Starts with a bold question or statement. Three to five bullet points of what readers will learn. Ends with the post link. Friendly and conversational tone. Maximum three hashtags. Instagram caption: Opens with a bold hook in the first 125 characters to survive the truncation. Three to five key takeaways as short bullet points. Ends with "Link in bio to read the full guide." Three to five relevant hashtags. Does not include the URL in the body since Instagram does not make links clickable in captions. LinkedIn caption: Professional tone throughout. Hook or surprising statement to open. Three to four short paragraphs covering the problem, what the post covers, the key insight, and a call to action with the link. Two to three professional hashtags. Between 800 and 1200 characters total for maximum LinkedIn algorithmic reach. The output is enforced as strict JSON via a Structured Output Parser — ensuring reliable, parseable data for the downstream publishing nodes every time. Step 6 — Simultaneous Publishing to All Platforms After caption generation, the workflow publishes to all four platforms in parallel: Twitter/X: The Twitter node posts the caption directly using OAuth2 authentication. No additional steps required. Facebook: An HTTP Request node calls the Facebook Graph API /{page_id}/photos endpoint, sending the Facebook caption as the message and the WordPress featured image URL as the photo. This creates a photo post with the caption on your Facebook Page, which performs significantly better than link-only posts in terms of reach and engagement. Instagram: Instagram publishing requires two steps as per the official Meta Graph API requirements. First, an HTTP Request node calls /{instagram_id}/media to create a media container with the caption and featured image URL. Second, another HTTP Request node calls /{instagram_id}/media_publish with the container ID returned from the first step. This is the official two-step publishing flow required by Meta. LinkedIn: The thumbnail image is first downloaded via an HTTP Request node (LinkedIn requires binary image data, not just a URL). Then the LinkedIn node creates a post using the downloaded image binary along with the LinkedIn caption. The post is published to your company organisation page using OAuth2 authentication. WordPress Plugin Required This workflow requires a webhook plugin installed on your WordPress site to send post data to n8n when a post is published. Recommended: WP Webhooks (Free) Available at wordpress.org/plugins/wp-webhooks After installing WP Webhooks on your WordPress site: Go to Settings → WP Webhooks in your WordPress dashboard Click Send Data tab Add a new webhook Select trigger: Post Published or Post Updated Paste your n8n webhook URL as the delivery URL Save and test The plugin will send a POST request to n8n with the full post data every time you publish a blog post. Alternative options: WPBrutha Webhooks Bit Integrations (free plugin with webhook sending) Custom code using WordPress transition_post_status hook if you prefer a code-based approach Payload Structure Your n8n webhook node expects this data structure from WordPress: { "post": { "post_title": "Your Blog Post Title", "post_content": "Your full HTML post content", "post_type": "post", "post_status": "publish" }, "post_permalink": "https://yoursite.com/your-post-slug/", "post_thumbnail": "https://yoursite.com/wp-content/uploads/featured-image.jpg" } The exact field names depend on which WordPress webhook plugin you use. Adjust the Set node field mappings to match the payload structure sent by your chosen plugin. Requirements WordPress website with admin access WP Webhooks plugin (free) or any WordPress webhook plugin installed and configured n8n instance (self-hosted or cloud) with a publicly accessible webhook URL OpenRouter API key (supports GPT-4o, Claude, Gemini, and 200+ other models) Twitter/X Developer account with OAuth2 app credentials Facebook Page with a System User token from Meta Business Portfolio or a long-lived Page Access Token Instagram Business Account connected to your Facebook Page via Meta Business Portfolio LinkedIn Company Page with OAuth2 app credentials Setup Steps Install and activate WP Webhooks on your WordPress site Copy the n8n webhook URL from the Webhook node in this workflow Configure WP Webhooks to send to your n8n webhook URL on post publish Add your OpenRouter API key to n8n credentials Connect your Twitter/X OAuth2 credentials in the Create Tweet node Add your Facebook Page Access Token or System User Token to the HTTP Query Auth credential used in Post on FB node Add the same token to the Instagram nodes — Create Media Container and Post On IG Update the facebook_page_id value in the Clean Data node with your actual Facebook Page ID Update the instagram_id value in the Clean Data node with your actual Instagram Business Account ID Connect your LinkedIn OAuth2 credentials in the Create a post node Update the LinkedIn organisation URN in the Create a post node with your actual company page URN Publish a test post on WordPress and verify all four platforms receive their posts Customization Options Change the AI model** — swap GPT-4o for any OpenRouter model including Claude Sonnet, Gemini Flash, or Mistral to reduce cost per run Adjust caption length and style** — edit the platform guidelines in the system prompt to match your brand voice or preferred posting style Add more platforms** — extend the workflow with TikTok, Pinterest, Threads, or YouTube community posts by adding nodes after the AI Agent Filter by category or tag** — add an IF node after the filter to only trigger for posts in specific WordPress categories Add a delay between platforms** — add Wait nodes between publishing steps if you prefer staggered posting Schedule posts instead of immediate publishing** — replace the direct publishing nodes with a scheduling tool like Postiz or Buffer API Content length control** — the AI currently reads the first 800 characters of post content. Increase or decrease this slice in the user prompt to give the AI more or less context Language support** — the AI will match the language of your blog post automatically. Write in Urdu, Arabic, French, or any other language and captions will be generated in the same language Use Cases SEO and digital marketing agencies** promoting client blog content across all platforms automatically Bloggers and content creators** who publish regularly and want consistent social media presence without manual effort ecommerce businesses** promoting product guides, how-to articles, and SEO content SaaS companies** distributing thought leadership content and product updates Local businesses** sharing informative blog posts about their services News and media sites** distributing articles to social audiences in real time Nodes Used Webhook — receives WordPress post data on publish IF — filters for published blog posts only Markdown — converts HTML post content to clean text Set — normalizes and extracts post data into clean variables AI Agent (LangChain) — generates four platform-specific captions simultaneously Structured Output Parser — enforces JSON output from the AI agent OpenRouter Chat Model — powers the AI caption generation Twitter node — publishes to Twitter/X HTTP Request (Facebook) — posts photo with caption to Facebook Page via Graph API HTTP Request (Instagram container) — creates Instagram media container via Graph API HTTP Request (Instagram publish) — publishes Instagram container to feed HTTP Request (LinkedIn image) — downloads featured image binary for LinkedIn LinkedIn node — publishes image post to LinkedIn company page Built with Meta Graph API v25.0, OpenRouter, Twitter API v2, and LinkedIn API. Compatible with any WordPress site running version 4.7 or above with the WordPress REST API enabled.
by Davide
This workflow automates the creation of long AI-generated videos from prompts, merges the generated clips into a single video, and automatically distributes the final content across multiple platforms. The process starts with a Google Sheet that acts as the control panel for the workflow. Each row in the sheet contains a prompt, the duration of the clip, and a starting frame. The workflow reads this data and generates video clips sequentially. Using the RunPod WAN 2.5 video generation API, the workflow creates individual video segments based on the prompt and input image. Each segment is then stored and tracked in the spreadsheet. Once all clips are generated, the workflow uses the Fal.run FFmpeg API to merge them into a single long video. After merging, the final video is retrieved automatically. The workflow also extracts the last frame of each generated clip to use as the starting frame for the next clip, ensuring smooth visual continuity between scenes. Finally, the completed video is automatically: Uploaded to Google Drive for storage Published to YouTube Uploaded to Postiz, which distributes it to social platforms such as TikTok, Instagram, Facebook, X, and YouTube This creates a fully automated pipeline that transforms prompts in a spreadsheet into a finished long-form video distributed across multiple platforms. Key Advantages 1. ✅ Fully Automated Video Production The workflow automates the entire process of generating, assembling, and publishing videos, eliminating manual editing and upload steps. 2. ✅ Spreadsheet-Based Control Using Google Sheets as the input system makes the workflow easy to manage and scalable. Users can create or modify video scenes simply by editing rows in the sheet. 3. ✅ Scalable AI Video Generation The workflow can generate multiple clips and combine them into longer videos, enabling the creation of long-form content from short AI-generated segments. 4. ✅ Seamless Scene Continuity By extracting the last frame of each clip and using it as the starting frame for the next scene, the workflow maintains visual continuity between segments. 5. ✅ Automatic Video Merging The Fal.run FFmpeg API merges all generated clips into a single final video without requiring external editing tools. 6. ✅ Multi-Platform Distribution Once the video is completed, it is automatically uploaded and published to multiple platforms, significantly reducing the time needed for content distribution. 7. ✅ Centralized Storage The final video is saved to Google Drive, providing organized and secure storage for the generated content. 8. ✅ Error Handling and Status Monitoring The workflow continuously checks the status of generation and processing tasks, waiting and retrying until the job is completed. How it works This workflow automates the creation of long videos by generating multiple clips from a Google Sheet and merging them together. Here's the process: Trigger & Data Loading: When manually executed, the workflow reads a Google Sheet containing video generation parameters (prompts, durations, and starting images). Video Generation Loop: For each row marked for processing, it: Sends the prompt and parameters to RunPod's WAN 2.5 video generation API Waits for completion (with status checking every 60 seconds) Retrieves the generated video URL and updates the Google Sheet Frame Extraction: After each video is generated, it extracts the last frame using Fal.ai's FFmpeg API and updates the next row's starting image (creating visual continuity). Video Merging: Once all individual clips are generated (marked with "x" in the MERGE column), the workflow: Collects all video URLs Sends them to Fal.ai's FFmpeg merge API Polls for completion every 60 seconds Retrieves the final merged video Distribution: The final long video is: Uploaded to Google Drive Posted to YouTube via Upload-Post API Posted to multiple social platforms (TikTok, Instagram, Facebook, X) via Postiz Setup steps Google Sheet Setup: Clone this template sheet Update the sheet ID in all Google Sheets nodes Fill in columns: START (initial image URL), PROMPT, DURATION (4, 6, or 8 seconds) Mark rows to merge with "x" in the MERGE column API Credentials Required: Google Sheets OAuth2: For reading/writing spreadsheet data Google Drive OAuth2: For uploading final videos Fal.ai API Key: For frame extraction and video merging RunPod API Key: For WAN 2.5 video generation Upload-Post API Key: For YouTube uploads Postiz API Key: For social media posting Configure Nodes: Update YOUR_USERNAME in the "Upload to Youtube" node Set channel IDs and titles in the "Upload to Social" node (integrationId, content) Verify folder IDs in Google Drive nodes Test: Run the workflow manually to generate your first long video sequence 👉 Subscribe to my new YouTube channel. Here I’ll share videos and Shorts with practical tutorials and FREE templates for n8n. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Vincent Nguyen
This n8n template is a complete, two-part social media engine: it automates competitor/inspiration research, uses AI to recreate viral formats for your niche, and handles multi-platform distribution with auto-generated visuals. Instead of staring at a blank page, this workflow scrapes top-performing content, breaks down the psychological hooks, rewrites it for your specific audience, and drafts a custom text graphic. Once you approve the draft, n8n handles the heavy lifting of posting it everywhere. You curate and approve. n8n researches, writes, designs, and publishes. How it works Part 1: Research & Recreating A scheduled trigger runs daily to pull a target username from your Airtable database. Apify scrapes the latest tweets from that user, filtering out replies and links, and sorts them by the highest view count. An AI Analysis Agent (Gemini) breaks down the top post's structure and creates a reusable template. An AI Writing Agent (GPT) uses that template to write a fresh, original post tailored to your niche. A Caption Agent (GPT) writes an engaging, long-form caption with hooks and hashtags. The draft is saved into Airtable for your review. Part 2: Auto-Posting & Repurposing When you change a draft's status to "Approved" in Airtable, the second flow triggers. An AI Image generator creates a sleek text-graphic based on your short post. The image is uploaded to ImgBB to generate a public URL. The pure text version is pushed directly to X (Twitter) and Threads (via authenticated HTTP requests). The generated image and long-form caption are published to LinkedIn, Facebook, and Instagram. Finally, the Airtable record is marked as "Posted." How to use Set up an Airtable base with your target usernames and a table for content drafts. Update the AI Agent system prompts (Writing and Caption) to match your specific target audience, expertise, and tone. Connect all your platform credentials in n8n (ensure you use the credential manager for the HTTP nodes, like HTTP Header Auth for Threads and Query Auth for ImgBB—no hardcoded API keys!). Let the schedule scrape and draft content for you. Simply mark a draft as "Approved" in Airtable when you want it to go live. What you get Automated viral content research and curation AI-driven structural analysis and rewriting Auto-generated text-graphics for visual platforms Hands-free publishing to 5 platforms (X, Threads, IG, FB, LinkedIn) A centralized content approval system Requirements Airtable account Apify account (Twitter Scraper actor) OpenAI & Google Gemini API credentials ImgBB account X (Twitter), Threads, LinkedIn, Facebook, and Instagram accounts Need help tweaking the AI agents for your specific funnel? Ask in the n8n Forum or shoot me a DM on LinkedIn Happy automating 🚀
by David P
Curate & post AI news to X, Bluesky, Threads & more via GPT-5 mini & Cue This n8n template automatically curates AI news from RSS feeds and generates platform-tailored social media posts using GPT-5 mini. Posts are saved as drafts in Cue for review before publishing to X, Bluesky, Threads, Mastodon, and Facebook. Use cases include: Daily automated AI/tech news curation Multi-platform social media content creation Building thought leadership with consistent posting Staying on top of industry news without manual effort Who is this for? This workflow is ideal for: Tech content creators who want to share AI news across multiple platforms Social media managers handling multiple accounts Anyone building an audience around AI/tech topics Teams who want consistent daily content without manual curation What problem does this workflow solve? Manually curating news, writing platform-specific posts, and publishing across 5 different social networks is time-consuming. This workflow automates the entire process: Curation** - Pulls from 4 trusted AI/tech RSS feeds daily Deduplication** - Tracks posted articles in Google Sheets so you never share the same story twice Content creation** - GPT-5 mini writes posts tailored to each platform's style and character limits Review workflow** - Creates drafts in Cue so you can review before publishing How it works Schedule Trigger - Runs daily at 9am (configurable) RSS Feeds - Fetches articles from TechCrunch AI, Ars Technica AI, The Verge AI, and MIT Tech Review Filter & Merge - Combines all feeds and filters to articles from the last 7 days Deduplication - Compares against Google Sheets to find unposted articles Random Selection - Picks one random article from available stories AI Generation - GPT-5 mini generates 5 platform-specific posts with appropriate tone and length Save to Cue - Creates a draft post with all 5 platform variations Log to Sheet - Records the article URL to prevent future duplicates Setup Requirements Cue account with connected social accounts OpenAI API key Google account for Sheets Step 1: Install the Cue community node Go to Settings → Community Nodes Click Install Enter @cuehq/n8n-nodes-cue Step 2: Create tracking spreadsheet Create a new Google Sheet named "AI News Tracker" Add these column headers in row 1: article_url title source processed_at Step 3: Configure credentials Google Sheets - Add OAuth2 credentials and connect to the "Get Recent Posts" node OpenAI - Add your API key and connect to the "GPT-5 mini" node Cue - Add your API key from Cue Settings Step 4: Configure the Cue node Open the Create Draft in Cue node Select your Profile For each platform slot, select your social account: Slot 1 → X/Twitter Slot 2 → Bluesky Slot 3 → Threads Slot 4 → Mastodon Slot 5 → Facebook Don't have all 5 platforms? Simply delete the unused slots. Step 5: Publish Save and click Publish to activate the workflow. Customizing this workflow Change the schedule Edit the Daily 9am Trigger node to run at a different time or frequency. Use different RSS feeds Replace the feed URLs with sources relevant to your niche. The workflow handles any standard RSS feed. Keep 3-6 feeds for best results. Auto-publish instead of drafts To publish immediately instead of creating drafts, enable Publish Immediately in the Cue node settings. Adjust the AI tone Modify the system prompt in the Write Social Posts node to match your brand voice or adjust platform-specific guidelines. Good to know Cost** - Each run uses one OpenAI API call. With GPT-5 mini, this costs approximately $0.01-0.02 per execution. Draft review** - Posts are created as drafts in Cue, giving you a chance to review and edit before publishing. Deduplication** - The Google Sheet tracks all posted URLs, so the same article is never shared twice. About Cue Cue is a social media scheduling platform that lets you manage and publish content across X, Bluesky, Threads, Mastodon, Facebook, LinkedIn, TikTok, and Instagram from a single dashboard. Key features: Multi-platform publishing** - Schedule once, publish everywhere Platform-specific content** - Tailor each post for different audiences Draft workflow** - Review and edit before publishing API & integrations** - Connect with n8n, Zapier, Make, and custom apps Get started free · Documentation · n8n Community Node
by Kyriakos Papadopoulos
Auto-Summarize Blog Posts to Social Media with Gemma and Postiz This workflow automates fetching the latest post from a Blogspot RSS feed, summarizes it with an LLM (e.g., Gemma via Ollama), extracts and uploads an image, generates three relevant hashtags, and posts to Facebook, LinkedIn, X (Twitter), and Instagram via the Postiz API. It ensures content fits platform limits (e.g., 280 characters for X) and prevents duplicates using hashing. Pros: Efficient for content creators Local LLM ensures privacy Customizable for any RSS/blog source Cons: Dependent on stable APIs (Postiz/social platforms) LLM outputs may vary in quality without human review Target Audience: Bloggers, content marketers, or social media managers looking to automate cross-platform posting from RSS sources, especially those focused on niches like health, tech, or personal development. Ideal for users with technical setup skills for self-hosting. Customization Options: Adapt prompts in "Generate Summary and Hashtags with LLM" for tone/style (e.g., professional vs. casual). Modify maxChars/hashtag reserve in "Calculate Summary Character Limit" for different platforms. Extend for multiple RSS feeds by adjusting "Calculate Summary Character Limit" array. Add error handling (e.g., IF node after "Create and Post Content via Postiz API") for API failures. Disclaimer: This template is designed for self-hosted n8n instances to leverage local Ollama for privacy. For cloud use, modify as follows: 1) Use an n8n cloud account, 2) Replace Ollama with a cloud API-based LLM like ChatGPT in the "Configure Local LLM Model (Ollama)" node, 3) Switch to cloud-hosted Postiz in the HTTP Request node. Template Image: How it works Set the RSS feed URL in "Set RSS Feed URLs". Fetch the latest post via RSS. Normalize fields and calculate the maximum summary length. Use the LLM to summarize the text, append hashtags, and include the link. Extract and process an image from the post HTML. Validate inputs and post to social platforms via the Postiz API. Setup Instructions Install n8n (self-hosted recommended for Ollama integration). Set up Ollama with the Gemma (or a similar) model using "Ollama Model" credentials. Add Postiz API credentials in the "Create and Post Content via Postiz API" node. Replace placeholders: RSS URL in "Set News RSS Feeds" Integration IDs in the Postiz HTTP body (Optional) Add error handling for API failures. Activate the workflow and test with a sample post. Uncertainties Changes in social media APIs may break posting functionality. LLM output consistency depends on model choice and prompt configuration. Required n8n Version Tested on n8n v1.107.3 (self-hosted). Works with the community node n8n-nodes-langchain. Resources n8n Docs: RSS Feed Read n8n Docs: HTTP Request Ollama Setup Postiz Documentation
by DataMinex
Transform property searches into personalized experiences! This powerful automation delivers dream home matches straight to clients' inboxes with professional CSV reports - all from a simple web form. 🚀 What this workflow does Create a complete real estate search experience that works 24/7: ✨ Smart Web Form - Beautiful property search form captures client preferences 🧠 Dynamic SQL Builder - Intelligently creates optimized queries from user input ⚡ Lightning Database Search - Scans 1000+ properties in milliseconds 📊 Professional CSV Export - Excel-ready reports with complete property details 📧 Automated Email Delivery - Personalized emails with property previews and attachments 🎯 Perfect for: Real Estate Agents** - Generate leads and impress clients with instant service Property Managers** - Automate tenant matching and recommendations Brokerages** - Provide 24/7 self-service property discovery Developers** - Showcase available properties with professional automation 💡 Why this workflow is a game-changer > "From property search to professional report delivery in under 30 seconds!" ⚡ Instant Results: Zero wait time for property matches 🎨 Professional Output: Beautiful emails that showcase your expertise 📱 Mobile Optimized: Works flawlessly on all devices 🧠 Smart Filtering: Only searches criteria clients actually specify 📈 Infinitely Scalable: Handles unlimited searches simultaneously 📊 Real Estate Data Source Built on authentic US market data from the Github: 🏘️ 1000+ Real Properties across all US states 💰 Actual Market Prices from legitimate listings 🏠 Complete Property Details (bedrooms, bathrooms, square footage, lot size) 📍 Verified Locations with accurate cities, states, and ZIP codes 🏢 Broker Information for authentic real estate context 🛠️ Quick Setup Guide Prerequisites Checklist ✅ [ ] SQL Server database (MySQL/PostgreSQL also supported) [ ] Gmail account for automated emails [ ] n8n instance (cloud or self-hosted) [ ] 20 minutes setup time Step 1: Import Real Estate Data 📥 🌟 Download the data 💾 Download CSV file (1000+ properties included) 🗄️ Create SQL Server table with this exact schema: CREATE TABLE [REALTOR].[dbo].[realtor_usa_price] ( brokered_by BIGINT, status NVARCHAR(50), price DECIMAL(12,2), bed INT, bath DECIMAL(3,1), acre_lot DECIMAL(10,8), street BIGINT, city NVARCHAR(100), state NVARCHAR(50), zip_code INT, house_size INT, prev_sold_date NVARCHAR(50) ); 📊 Import your CSV data into this table Step 2: Configure Database Connection 🔗 🔐 Set up Microsoft SQL Server credentials in n8n ✅ Test connection to ensure everything works 🎯 Workflow is pre-configured for the table structure above Step 3: Gmail Setup (The Magic Touch) 📧 🌐 Visit Google Cloud Console 🆕 Create new project (or use existing) 🔓 Enable Gmail API in API Library 🔑 Create OAuth2 credentials (Web Application) ⚙️ Add your n8n callback URL to authorized redirects 🔗 Configure Gmail OAuth2 credentials in n8n ✨ Authorize your Google account Step 4: Launch Your Property Search Portal 🚀 📋 Import this workflow template (form is pre-configured) 🌍 Copy your webhook URL from the Property Search Form node 🔍 Test with a sample property search 📨 Check email delivery with CSV attachment 🎉 Go live and start impressing clients! 🎨 Customization Playground 🏷️ Personalize Your Brand // Customize email subjects in the Gmail node "🏠 Exclusive Properties Curated Just for You - ${results.length} Perfect Matches!" "✨ Your Dream Home Portfolio - Handpicked by Our Experts" "🎯 Hot Market Alert - ${results.length} Premium Properties Inside!" 🔧 Advanced Enhancements 🎨 HTML Email Templates**: Create stunning visual emails with property images 📊 Analytics Dashboard**: Track popular searches and user engagement 🔔 Smart Alerts**: Set up automated price drop notifications 📱 Mobile Integration**: Connect to React Native or Flutter apps 🤖 AI Descriptions**: Add ChatGPT for compelling property descriptions 🌍 Multi-Database Flexibility // Easy database switching // MySQL: Replace Microsoft SQL node → MySQL node // PostgreSQL: Swap for PostgreSQL node // MongoDB: Use MongoDB node with JSON queries // Even CSV files: Use CSV reading nodes for smaller datasets 🚀 Advanced Features & Extensions 🔥 Pro Tips for Power Users 🔄 Bulk Processing**: Handle multiple searches simultaneously 💾 Smart Caching**: Store popular searches for lightning-fast results 📈 Lead Scoring**: Track which properties generate most interest 📅 Follow-up Automation**: Schedule nurturing email sequences 🎯 Integration Possibilities 🏢 CRM Connection**: Auto-add qualified leads to your CRM 📅 Calendar Integration**: Add property viewing scheduling 📊 Price Monitoring**: Track market trends and price changes 📱 Social Media**: Auto-share featured properties to social platforms 💬 Chat Integration**: Connect to WhatsApp or SMS for instant alerts 🔗 Expand Your Real Estate Automation 🌟 Related Workflow Ideas 🤖 AI Property Valuation - Add machine learning for price predictions 📊 Market Analysis Reports - Generate comprehensive market insights 📱 SMS Property Alerts - Instant text notifications for hot properties 🏢 Commercial Property Search - Adapt for office and retail spaces 💹 Investment ROI Calculator - Add financial analysis for investors 🏘️ Neighborhood Analytics - Include school ratings and demographics 🛠️ Technical Extensions 📷 Image Processing: Auto-resize and optimize property photos 🗺️ Map Integration: Add interactive property location maps 📱 Progressive Web App: Create mobile app experience 🔔 Push Notifications: Real-time alerts for saved searches 🚀 Get Started Now Import this workflow template Configure your database and Gmail Customize branding and messaging Launch your professional property search portal Watch client satisfaction soar!
by Dr. Firas
💥 Automate AI Video Creation & Multi-Platform Publishing with Veo 3.1 & Blotato 🎯 Who is this for? This workflow is designed for content creators, marketers, and automation enthusiasts who want to produce professional AI-generated videos and publish them automatically on social media — without editing or manual uploads. Perfect for those using Veo 3.1, GPT-4, and Blotato to scale video creation. 💡 What problem is this workflow solving? Creating short-form content (TikTok, Instagram Reels, YouTube Shorts) is time-consuming — from writing scripts to video editing and posting. This workflow eliminates the manual steps by combining AI storytelling + video generation + automated publishing, letting you focus on creativity while your system handles production and distribution. ⚙️ What this workflow does Reads new ideas from Google Sheets Generates story scripts using GPT-4 Creates cinematic videos using Veo 3.1 (fal.ai/veo3.1/reference-to-video) with 3 input reference images Uploads the final video automatically to Google Drive Publishes the video across multiple platforms (TikTok, Instagram, Facebook, X, LinkedIn, YouTube) via Blotato Updates Google Sheets with video URL and status (Completed / Failed) 🧩 Setup Required accounts: OpenAI → GPT-4 API key fal.ai → Veo 3.1 API key Google Cloud Console → Sheets & Drive connection Blotato → API key for social media publishing Configuration steps: Copy the Google Sheets structure: A: id_video B: niche C: idea D: url_1 E: url_2 F: url_3 G: url_final H: status Add your API keys to the Workflow Configuration node. Insert three image URLs and a short idea into your sheet. Wait for the automation to process and generate your video. 🧠 How to customize this workflow Change duration or aspect ratio** → Edit the Veo 3.1 node JSON body (duration, aspect_ratio) Modify prompt style** → Adjust the “Optimize Prompt for Veo” node for your desired tone or cinematic look Add more platforms** → Extend Blotato integration to publish on Pinterest, Reddit, or Threads Enable Telegram Trigger** → Allow users to submit ideas and images directly via Telegram 🚀 Expected Outcome Within 2–3 minutes, your idea is transformed into a full cinematic AI video — complete with storytelling, visuals, and automatic posting to your social media channels. Save hours of editing and focus on strategy, creativity, and growth. 👋 Need help or want to customize this? 📩 Contact: LinkedIn 📺 YouTube: @DRFIRASS 🚀 Workshops: Mes Ateliers n8n 📄 Documentation: Notion Guide Need help customizing? Contact me for consulting and support : Linkedin / Youtube / 🚀 Mes Ateliers n8n
by isaWOW
Description Automate Facebook post scheduling from a Google Sheets content calendar. Runs 4 times daily, reads approved posts scheduled for today, downloads images from Google Drive, schedules via Facebook Graph API, and updates tracking sheet with published URLs—perfect for social media managers and agencies. What this workflow does This workflow eliminates manual Facebook posting by automating the entire scheduling process from a centralized Google Sheets content calendar. It runs four times daily (9:35 AM, 10:35 AM, 11:35 AM, 12:35 PM) to catch posts scheduled at different times throughout the morning. The workflow reads your Google Sheet, filters posts marked with Approval Status = "Good" and Platform = "Facebook", then checks which posts are scheduled for today. For each approved post, it intelligently determines if it's a text-only post or a photo post—if there's a Media URL, it downloads the image from Google Drive; otherwise, it schedules just the text. Both types are scheduled via Facebook Graph API with future publishing times (not posted immediately), and once successfully scheduled, the workflow updates your Google Sheet with the published post URL and changes the Approval Status to "Published". This creates a complete audit trail of all scheduled content while supporting team collaboration through the approval workflow. Perfect for social media managers handling multiple Facebook pages, marketing agencies scheduling client content with approval checkpoints, content creators batch-planning posts in Google Sheets, and teams needing collaborative content calendars with centralized image management. Key features Google Sheets content calendar: Manage all Facebook posts in a familiar spreadsheet with columns for Scheduled On, Platform, Post Type, Caption, Media URL, and Approval Status—no complex social media management tools needed. Built-in approval workflow: Only posts marked "Good" in the Approval Status column are published. Team members can review, approve, or reject posts directly in Google Sheets before they go live. Dual post type support: Handles both text-only posts (scheduled via /feed endpoint) and photo posts (scheduled via /photos endpoint with binary image data)—automatically detects which type based on Media URL presence. Google Drive image integration: Stores all images in Google Drive (centralized, shared storage), then automatically downloads them when scheduling photo posts—no manual file management needed. Runs 4 times daily: Schedule trigger fires at 9:35 AM, 10:35 AM, 11:35 AM, and 12:35 PM to catch posts scheduled at different morning times—handles busy posting schedules without missing slots. Facebook Graph API scheduling: Uses official Facebook Graph API v24.0 with scheduled_publish_time parameter (published: false) to schedule posts for future times—not immediate posting, actual scheduling. Post URL tracking: After successfully scheduling, updates Google Sheet with the published Facebook post URL—creates complete audit trail and enables easy post performance tracking. Multi-platform ready: Uses "Platform" column to filter Facebook posts only—same Google Sheet can manage Instagram, LinkedIn, Twitter content by adding more platform-specific workflows. Story post filtering: Automatically skips posts where Post Type = "Story" (Facebook Stories scheduling not supported by this workflow)—only processes Feed and Photo posts. How it works 1. Scheduled trigger fires 4 times daily A cron trigger runs at 9:35 AM, 10:35 AM, 11:35 AM, and 12:35 PM every day. This catches posts scheduled at different times throughout the morning without needing to run the workflow every minute. 2. Load Facebook credentials The workflow reads a separate ".env" sheet in your Google Sheets document containing: Facebook Page ID:** Your Facebook Page's unique ID Facebook Page Access Token:** Long-lived access token with pages_manage_posts and pages_read_engagement permissions These credentials are used for all Facebook Graph API calls later in the workflow. 3. Read approved Facebook posts The workflow reads your main "Post URL" sheet and applies two filters: Approval Status = "Good":** Only processes approved posts Platform = "Facebook":** Filters out Instagram, LinkedIn, etc. This returns all approved Facebook posts regardless of scheduled date. 4. Filter posts scheduled for today A Code node compares the "Scheduled On" column value against today's date (ignores time, just checks the date part). Posts scheduled for today pass through; others are filtered out. Supported date formats: "2025-10-30 10:00" "2025-10-30 06-42" Any format with YYYY-MM-DD at the beginning 5. Loop through each post The Split in Batches node processes one post at a time, preventing API rate limits and ensuring each post is handled individually. If there are 5 approved posts for today, it loops 5 times. 6. Platform verification A Switch node double-checks that Platform = "Facebook" (redundant but ensures accuracy). This allows the same workflow structure to be copied for other platforms. 7. Story post filtering An If node checks if Post Type != "Story". Facebook Stories scheduling is not supported in this workflow, so Story posts are skipped and merged back into the loop to continue with the next post. 8. Determine post type (text-only vs. photo) An If node checks if the "Media URL" column is empty: Empty → Text-only post** (routes to Branch A) Has value → Photo post** (routes to Branch B) Branch A: Text-Only Post 9a. Schedule Facebook text post HTTP POST request to Facebook Graph API: https://graph.facebook.com/v24.0/{page-id}/feed Parameters: message: Caption text from Google Sheet access_token: From credentials sheet published: false (schedules instead of posting immediately) scheduled_publish_time: Unix timestamp converted from "Scheduled On" field Example: If Scheduled On = "2025-10-30 14:00", the workflow converts this to Unix timestamp (1730296800) and Facebook schedules the post for that exact time. 10a. Update sheet with text post URL After successful API response, the workflow constructs the Facebook post URL from the response ID: https://www.facebook.com/{page-id}/posts/{post-id} Then updates the Google Sheet row: Approval Status:** "Published" Post URL:** Constructed Facebook URL This marks the post as published and provides a clickable link to view it on Facebook. Branch B: Photo Post 9b. Download image from Google Drive Uses the Media URL (Google Drive sharing link) to download the image file. Supports: Direct Google Drive file URLs Shared Drive files Public or private files (as long as the OAuth account has access) The image is downloaded as binary data and passed to the next node. 10b. Schedule Facebook photo post HTTP POST request to Facebook Graph API: https://graph.facebook.com/v24.0/{page-id}/photos Content-Type: multipart/form-data Parameters: source: Binary image data (from Google Drive download) caption: Caption text from Google Sheet access_token: From credentials sheet published: false (schedules instead of posting immediately) scheduled_publish_time: Unix timestamp + 15 minute buffer Note: Photo posts get an extra 15-minute buffer in the scheduled time to account for image processing delays on Facebook's side. 11b. Update sheet with photo post URL After successful API response, constructs the Facebook photo URL: https://www.facebook.com/photo/?fbid={photo-id} Then updates the Google Sheet row: Approval Status:** "Published" Post URL:** Constructed Facebook photo URL 12. Merge and loop All three branches (text posts, photo posts, skipped stories) merge back together. The loop then proceeds to the next post until all approved posts for today are processed. Setup requirements Tools you'll need: Active n8n instance (self-hosted or n8n Cloud) Google Sheets with OAuth access Google Drive with OAuth access Facebook Page (not personal profile) Facebook Page Access Token with proper permissions Estimated setup time: 30–35 minutes Configuration steps 1. Create Facebook Page Access Token Go to Facebook Developer Console Create an app (or use existing) Add "Facebook Login" product Under Tools → Graph API Explorer: Select your Page Request permissions: pages_manage_posts, pages_read_engagement, publish_to_groups Generate long-lived access token (follow Facebook's token extension process) Save the Page ID and Access Token 2. Set up Google Sheets Create two sheets in one Google Sheets document: Sheet 1: ".env" (credentials) | Facebook Page ID | Facebook Page Access Token | |---|---| | 123456789 | EAAxxxxxxx... | Sheet 2: "Post URL" (content calendar) | Scheduled On | Platform | Post Type | Caption | Media URL | Approval Status | Post URL | row_number | |---|---|---|---|---|---|---|---| | 2025-10-30 10:00 | Facebook | Photo | Check out our new product! | https://drive.google.com/file/d/xxx | Good | | 1 | | 2025-10-30 14:00 | Facebook | Feed | Happy Monday everyone! | | Good | | 2 | Important column details: Scheduled On:** Format must be YYYY-MM-DD HH-MM (24-hour format) Platform:** Must be "Facebook" (case-sensitive) Post Type:** "Feed" (text-only), "Photo" (with image), or "Story" (skipped) Media URL:** Google Drive sharing link (leave empty for text-only posts) Approval Status:** "Good" (publish), "Pending" (hold), "Rejected" (skip) Post URL:** Leave empty (auto-filled after publishing) row_number:** Auto-generated by Google Sheets 3. Connect Google Sheets OAuth In n8n: Credentials → Add credential → Google Sheets OAuth2 API Complete OAuth authentication Open these nodes and select your credential: "Load Facebook Credentials from Sheet" "Read Approved Facebook Posts" "Update Sheet with Photo Post URL" "Update Sheet with Text Post URL" 4. Connect Google Drive OAuth In n8n: Credentials → Add credential → Google Drive OAuth2 API Complete OAuth authentication Open "Download Image from Google Drive" node Select your Google Drive credential 5. Update sheet URLs Open the following nodes and update the documentId value with your Google Sheets URL: "Load Facebook Credentials from Sheet"** → Point to your .env sheet "Read Approved Facebook Posts"** → Point to your Post URL sheet "Update Sheet with Photo Post URL"** → Point to your Post URL sheet "Update Sheet with Text Post URL"** → Point to your Post URL sheet 6. Test with sample posts Add 2 test rows in your Google Sheet: Row 1: Text-only post (no Media URL) scheduled for today Row 2: Photo post (with Google Drive URL) scheduled for today Set both Approval Status to "Good" Manually trigger the workflow (or wait for the next scheduled run) Verify: Posts appear in Facebook's Publishing Tools as scheduled Google Sheet updated with Post URLs Approval Status changed to "Published" 7. Activate the workflow Toggle the workflow to Active The workflow will now run automatically at 9:35 AM, 10:35 AM, 11:35 AM, and 12:35 PM daily Monitor the first few days to ensure posts are scheduling correctly Use cases Social media managers: Schedule 20-30 Facebook posts per week from a centralized Google Sheets calendar. Team members add content, you approve in the sheet, workflow handles publishing—no manual Facebook Business Suite logins. Marketing agencies: Manage 10+ client Facebook Pages from one Google Sheet. Each client gets their own rows, separate Facebook credentials loaded per page, automated scheduling with URL tracking for client reporting. Content creators: Batch-create a month of posts in one sitting (captions + images in Google Drive), mark them "Good" when ready, let the workflow publish them at scheduled times—focus on creation, not distribution. Small businesses: Schedule promotional posts, event announcements, and product launches without paying for Buffer, Hootsuite, or Later. Free automation with Google Sheets as the interface. E-commerce stores: Schedule new product announcements with product images from Google Drive. Workflow downloads images, posts to Facebook with captions, tracks URLs for performance analysis. Agencies with approval workflows: Content team creates posts, marks "Pending". Manager reviews, changes to "Good" or "Rejected". Only approved posts publish—built-in quality control without third-party tools. Resources n8n documentation Facebook Graph API Facebook Page Access Tokens Google Sheets API Google Drive API n8n Schedule Trigger n8n Google Sheets node Support Need help or custom development? 📧 Email: info@isawow.com 🌐 Website: https://isawow.com/
by vinci-king-01
Customer Support Analysis Dashboard with AI and Automated Insights 🎯 Target Audience Customer support managers and team leads Customer success teams monitoring satisfaction Product managers analyzing user feedback Business analysts measuring support metrics Operations managers optimizing support processes Quality assurance teams monitoring support quality Customer experience (CX) professionals 🚀 Problem Statement Manual analysis of customer support tickets and feedback is time-consuming and often misses critical patterns or emerging issues. This template solves the challenge of automatically collecting, analyzing, and visualizing customer support data to identify trends, improve response times, and enhance overall customer satisfaction. 🔧 How it Works This workflow automatically monitors customer support channels using AI-powered analysis, processes tickets and feedback, and provides actionable insights for improving customer support operations. Key Components Scheduled Trigger - Runs the workflow at specified intervals to maintain real-time monitoring AI-Powered Ticket Analysis - Uses advanced NLP to categorize, prioritize, and analyze support tickets Multi-Channel Integration - Monitors email, chat, help desk systems, and social media Automated Insights - Generates reports on trends, response times, and satisfaction scores Dashboard Integration - Stores all data in Google Sheets for comprehensive analysis and reporting 📊 Google Sheets Column Specifications The template creates the following columns in your Google Sheets: | Column | Data Type | Description | Example | |--------|-----------|-------------|---------| | timestamp | DateTime | When the ticket was processed | "2024-01-15T10:30:00Z" | | ticket_id | String | Unique ticket identifier | "SUP-2024-001234" | | customer_email | String | Customer contact information | "john@example.com" | | subject | String | Ticket subject line | "Login issues with new app" | | description | String | Full ticket description | "I can't log into the mobile app..." | | category | String | AI-categorized ticket type | "Technical Issue" | | priority | String | Calculated priority level | "High" | | sentiment_score | Number | Customer sentiment (-1 to 1) | -0.3 | | urgency_indicator | String | Urgency classification | "Immediate" | | response_time | Number | Time to first response (hours) | 2.5 | | resolution_time | Number | Time to resolution (hours) | 8.0 | | satisfaction_score | Number | Customer satisfaction rating | 4.2 | | agent_assigned | String | Support agent name | "Sarah Johnson" | | status | String | Current ticket status | "Resolved" | 🛠️ Setup Instructions Estimated setup time: 20-25 minutes Prerequisites n8n instance with community nodes enabled ScrapeGraphAI API account and credentials Google Sheets account with API access Help desk system API access (Zendesk, Freshdesk, etc.) Email service integration (optional) Step-by-Step Configuration 1. Install Community Nodes Install required community nodes npm install n8n-nodes-scrapegraphai npm install n8n-nodes-slack 2. Configure ScrapeGraphAI Credentials Navigate to Credentials in your n8n instance Add new ScrapeGraphAI API credentials Enter your API key from ScrapeGraphAI dashboard Test the connection to ensure it's working 3. Set up Google Sheets Connection Add Google Sheets OAuth2 credentials Grant necessary permissions for spreadsheet access Create a new spreadsheet for customer support analysis Configure the sheet name (default: "Support Analysis") 4. Configure Support System Integration Update the websiteUrl parameters in ScrapeGraphAI nodes Add URLs for your help desk system or support portal Customize the user prompt to extract specific ticket data Set up categories and priority thresholds 5. Set up Notification Channels Configure Slack webhook or API credentials for alerts Set up email service credentials for critical issues Define alert thresholds for different priority levels Test notification delivery 6. Configure Schedule Trigger Set analysis frequency (hourly, daily, etc.) Choose appropriate time zones for your business hours Consider support system rate limits 7. Test and Validate Run the workflow manually to verify all connections Check Google Sheets for proper data formatting Test ticket analysis with sample data 🔄 Workflow Customization Options Modify Analysis Targets Add or remove support channels (email, chat, social media) Change ticket categories and priority criteria Adjust analysis frequency based on ticket volume Extend Analysis Capabilities Add more sophisticated sentiment analysis Implement customer churn prediction models Include agent performance analytics Add automated response suggestions Customize Alert System Set different thresholds for different ticket types Create tiered alert systems (info, warning, critical) Add SLA breach notifications Include trend analysis alerts Output Customization Add data visualization and reporting features Implement support trend charts and graphs Create executive dashboards with key metrics Add customer satisfaction trend analysis 📈 Use Cases Support Ticket Management**: Automatically categorize and prioritize tickets Response Time Optimization**: Identify bottlenecks in support processes Customer Satisfaction Monitoring**: Track and improve satisfaction scores Agent Performance Analysis**: Monitor and improve agent productivity Product Issue Detection**: Identify recurring problems and feature requests SLA Compliance**: Ensure support teams meet service level agreements 🚨 Important Notes Respect support system API rate limits and terms of service Implement appropriate delays between requests to avoid rate limiting Regularly review and update your analysis parameters Monitor API usage to manage costs effectively Keep your credentials secure and rotate them regularly Consider data privacy and GDPR compliance for customer data 🔧 Troubleshooting Common Issues: ScrapeGraphAI connection errors: Verify API key and account status Google Sheets permission errors: Check OAuth2 scope and permissions Ticket parsing errors: Review the Code node's JavaScript logic Rate limiting: Adjust analysis frequency and implement delays Alert delivery failures: Check notification service credentials Support Resources: ScrapeGraphAI documentation and API reference n8n community forums for workflow assistance Google Sheets API documentation for advanced configurations Help desk system API documentation Customer support analytics best practices
by vinci-king-01
Competitor Price Monitoring Dashboard with AI and Real-time Alerts 🎯 Target Audience E-commerce managers and pricing analysts Retail business owners monitoring competitor pricing Marketing teams tracking market positioning Product managers analyzing competitive landscape Data analysts conducting pricing intelligence Business strategists making pricing decisions 🚀 Problem Statement Manual competitor price monitoring is inefficient and often leads to missed opportunities or delayed responses to market changes. This template solves the challenge of automatically tracking competitor prices, detecting significant changes, and providing actionable insights for strategic pricing decisions. 🔧 How it Works This workflow automatically monitors competitor product prices using AI-powered web scraping, analyzes price trends, and sends real-time alerts when significant changes are detected. Key Components Scheduled Trigger - Runs the workflow at specified intervals to maintain up-to-date price data AI-Powered Scraping - Uses ScrapeGraphAI to intelligently extract pricing information from competitor websites Price Analysis Engine - Processes historical data to detect trends and anomalies Alert System - Sends notifications via Slack and email when price changes exceed thresholds Dashboard Integration - Stores all data in Google Sheets for comprehensive analysis and reporting 📊 Google Sheets Column Specifications The template creates the following columns in your Google Sheets: | Column | Data Type | Description | Example | |--------|-----------|-------------|---------| | timestamp | DateTime | When the price was recorded | "2024-01-15T10:30:00Z" | | competitor_name | String | Name of the competitor | "Amazon" | | product_name | String | Product name and model | "iPhone 15 Pro 128GB" | | current_price | Number | Current price in USD | 999.00 | | previous_price | Number | Previous recorded price | 1099.00 | | price_change | Number | Absolute price difference | -100.00 | | price_change_percent | Number | Percentage change | -9.09 | | product_url | URL | Direct link to product page | "https://amazon.com/iphone15" | | alert_triggered | Boolean | Whether alert was sent | true | | trend_direction | String | Price trend analysis | "Decreasing" | 🛠️ Setup Instructions Estimated setup time: 15-20 minutes Prerequisites n8n instance with community nodes enabled ScrapeGraphAI API account and credentials Google Sheets account with API access Slack workspace for notifications (optional) Email service for alerts (optional) Step-by-Step Configuration 1. Install Community Nodes Install required community nodes npm install n8n-nodes-scrapegraphai npm install n8n-nodes-slack 2. Configure ScrapeGraphAI Credentials Navigate to Credentials in your n8n instance Add new ScrapeGraphAI API credentials Enter your API key from ScrapeGraphAI dashboard Test the connection to ensure it's working 3. Set up Google Sheets Connection Add Google Sheets OAuth2 credentials Grant necessary permissions for spreadsheet access Create a new spreadsheet for price monitoring data Configure the sheet name (default: "Price Monitoring") 4. Configure Competitor URLs Update the websiteUrl parameters in ScrapeGraphAI nodes Add URLs for each competitor you want to monitor Customize the user prompt to extract specific pricing data Set appropriate price thresholds for alerts 5. Set up Notification Channels Configure Slack webhook or API credentials Set up email service credentials (SendGrid, SMTP, etc.) Define alert thresholds and notification preferences Test notification delivery 6. Configure Schedule Trigger Set monitoring frequency (hourly, daily, etc.) Choose appropriate time zones for your business hours Consider competitor website rate limits 7. Test and Validate Run the workflow manually to verify all connections Check Google Sheets for proper data formatting Test alert notifications with sample data 🔄 Workflow Customization Options Modify Monitoring Targets Add or remove competitor websites Change product categories or specific products Adjust monitoring frequency based on market volatility Extend Price Analysis Add more sophisticated trend analysis algorithms Implement price prediction models Include competitor inventory and availability tracking Customize Alert System Set different thresholds for different product categories Create tiered alert systems (info, warning, critical) Add SMS notifications for urgent price changes Output Customization Add data visualization and reporting features Implement price history charts and graphs Create executive dashboards with key metrics 📈 Use Cases Dynamic Pricing**: Adjust your prices based on competitor movements Market Intelligence**: Understand competitor pricing strategies Promotion Planning**: Time your promotions based on competitor actions Inventory Management**: Optimize stock levels based on market conditions Customer Communication**: Proactively inform customers about price changes 🚨 Important Notes Respect competitor websites' terms of service and robots.txt Implement appropriate delays between requests to avoid rate limiting Regularly review and update your monitoring parameters Monitor API usage to manage costs effectively Keep your credentials secure and rotate them regularly Consider legal implications of automated price monitoring 🔧 Troubleshooting Common Issues: ScrapeGraphAI connection errors: Verify API key and account status Google Sheets permission errors: Check OAuth2 scope and permissions Price parsing errors: Review the Code node's JavaScript logic Rate limiting: Adjust monitoring frequency and implement delays Alert delivery failures: Check notification service credentials Support Resources: ScrapeGraphAI documentation and API reference n8n community forums for workflow assistance Google Sheets API documentation for advanced configurations Slack API documentation for notification setup
by Max aka Mosheh
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. How it works • Publishes content to 9 social platforms (Instagram, YouTube, TikTok, Facebook, LinkedIn, Threads, Twitter/X, Bluesky, Pinterest) from a single Airtable base • Automatically uploads media to Blotato, handles platform-specific requirements (YouTube titles, Pinterest boards), and tracks success/failure for each post • Includes smart features like GPT-powered YouTube title optimization, Pinterest Board ID finder tool, and random delays to avoid rate limits Set up steps • Takes ~20–35 minutes to configure all 9 platforms (or less if you only need specific ones) • Requires Airtable personal access token, Blotato API key, and connecting your social accounts in Blotato dashboard • Workflow includes comprehensive sticky notes with step-by-step Airtable base setup, credential configuration, platform ID locations, and quick debugging links for each social network Pro tip: The workflow is modular - you can disable any platforms you don't use by deactivating their respective nodes, making it flexible for any social media strategy from single-platform to full omnichannel publishing.
by Ronalds Palacis
🚀 AI-Powered LinkedIn Post Automation with Figma Templates 🧩 How It Works This workflow automatically generates professional, branded LinkedIn posts using your custom Figma designs. Perfect for marketers, agencies, content creators, and businesses who want to maintain consistent branding while automating social media content creation. Key Features: Design-first approach using Figma templates AI-powered content generation (optional) High-quality image generation from templates Automatic LinkedIn publishing Telegram notifications for success/failure tracking High-Level Workflow: Template Selection: Fetch your pre-designed Figma templates from Templated Content Preparation: Set static content or configure the prompts in the agents to generate with AI (ChatGPT/Claude) Image Generation: Create branded images with dynamic content via Templated MCP server LinkedIn Publishing: Automatically post text and image to your LinkedIn profile or company page (single image or carousel post) Notification: Receive Telegram alert on success/failure (optional) ⚙️ Set Up Steps (Quick Overview) 🕐 Estimated Setup Time: ~15 minutes Create Templated Account: Sign up at templated.cometai.eu, import Figma designs, generate API key Configure LinkedIn OAuth: Set up LinkedIn Developer app with OAuth2 credentials for automatic posting Connect Templated MCP: Add API key authentication to MCP server nodes for template and image generation Set Up Telegram (Optional): Create bot for workflow notifications Customize Content: Add static fields or configure AI nodes for dynamic content generation Schedule & Activate: Set posting schedule (daily, weekly, etc.) and activate workflow 💡 Important Notes Figma Integration**: Uses real Figma files as templates - maintain professional design quality without manual recreation Template Variables**: Supports dynamic placeholders in Figma text layers for content replacement Character Limits**: Respects maxLength settings to ensure text fits your design Rate Limits**: LinkedIn allows 25 posts/day (personal), 100/day (company pages) AI-Ready**: Easily integrate ChatGPT, Claude, or other AI models for content generation Batch Generation**: Generate multiple posts at once with different templates 🛠 Detailed Node Breakdown 1. Schedule Trigger Action**: Triggers the workflow on a schedule (daily, weekly, custom cron) Configuration**: Set your desired posting frequency Alternative**: Use manual trigger for on-demand posts 2. LinkedIn Post Writer (OpenAI Chat Model) Action**: AI generates engaging LinkedIn post content based on your topic/prompt Tools**: Simple Memory, Think, Date & Time, Search latest news Output**: Professional post text ready for publication 3. Carousel Ideator (Templated MCP Client) Action**: Connects to Templated MCP server to fetch available templates Configuration**: Uses your Templated API key for authentication Output**: Available template IDs and configurations for carousel generation 4. Generate the Carousel (POST to Templated MCP) Action**: Sends content to Templated server with template ID and field data Input**: Template selection, content fields (title, subtitle, etc.) Output**: Encoded carousel images generated from Figma templates 5. Extract from File Action**: Extracts generated image data from the MCP response Process**: Parses the encoded image string for LinkedIn upload 6. Get LinkedIn User Info (HTTP Request) Action**: Fetches your LinkedIn profile URN for post attribution Authentication**: OAuth2 LinkedIn credentials Output**: User ID required for posting 7. Initialize Upload URN Action**: Requests upload URL from LinkedIn for carousel document upload Process**: Prepares LinkedIn's upload mechanism for multi-image posts 8. Edit Fields Action**: Maps and formats data for LinkedIn API requirements Process**: Structures image data and post metadata correctly 9. Convert to Binary Action**: Converts image data to binary format Key**: Required format for LinkedIn document upload API 10. Upload Posts as Binary Action**: Uploads the carousel document to LinkedIn's servers Process**: Multi-part upload of generated images 11. Get Uploaded File URN Action**: Retrieves LinkedIn's asset ID for the uploaded content Output**: Asset URN needed for post creation 12. Switch (Conditional Logic) Action**: Handles success/error routing Routes**: Directs to LinkedIn post creation on success, error notification on failure 13. Create LinkedIn Post (HTTP Request) Action**: Creates the final LinkedIn post with carousel and text Authentication**: OAuth2 with LinkedIn posting permissions Result**: Published carousel post on your LinkedIn feed 14. Success/Error Notifications (Telegram) Action**: Sends notification about workflow execution status Success**: Confirms post published with details Error**: Alerts you to failures with error message and debugging info ⏱ Execution Time Breakdown Total Estimated Execution Time: ~10–30 seconds per workflow run Template Fetch: ~1–2 seconds Content Preparation: ~1–2 seconds Image Generation: ~5–15 seconds LinkedIn Post Upload: ~2–5 seconds Telegram Notification: ~1–2 seconds Note: AI content generation (if added) adds ~5-10 seconds 🚀 Ready to Get Started? What You'll Need: Free Templated account at templated.cometai.eu Figma designs with placeholder text LinkedIn Developer app (free) n8n instance (cloud or self-hosted) Quick Start: Import this workflow to your n8n instance Follow the setup guide in the workflow notes Test with a single post Schedule for automatic posting Sit back and watch your LinkedIn feed grow! 🎉 📝 Notes & Customizations Template Customization Create multiple templates in Figma for different content types (quotes, announcements, tips, etc.) Rotate templates for visual variety Use template descriptions to guide AI content generation Set character limits per placeholder to prevent overflow Content Generation Options Static Mode**: Define fields manually for recurring post types AI Mode**: Connect ChatGPT/Claude for dynamic, topic-based content Hybrid Mode**: Mix static brand elements with AI-generated copy Scheduled Variety**: Rotate between templates and content styles Advanced Features Multi-Platform**: Duplicate workflow for Twitter, Instagram, Facebook A/B Testing**: Track performance across different templates Content Calendar**: Pull scheduled posts from Notion/Airtable Analytics Integration**: Log post IDs for engagement tracking