by bangank36
This workflow automates the Mark as Fulfilled action in Shopify for each order, ensuring a seamless fulfillment process without manual intervention. How It Works This workflow retrieves all unfulfilled orders and processes their fulfillment automatically. Since Shopify requires a Fulfillment Order ID (not Order ID) to trigger fulfillment, the workflow follows these steps: 1️⃣ Get all unfulfilled orders using the Shopify node. 2️⃣ Retrieve the Fulfillment Order ID using the "List Fulfillment Orders" action. 3️⃣ Create a fulfillment request using "Mark fulfillment order as fulfilled." 4️⃣ Handle edge cases, such as partially fulfilled orders or API errors. This ensures that every valid order is marked as fulfilled efficiently. 🔗 Ongoing discussions on this topic: Relevant Shopify API Discussion Step-by-step The workflow can be run as requested or on schedule You can adjust these parameters within the Shopify and filter nodes: Shopify Admin URL – A Global node to customize the Shopify store URL. To find your Shopify store ID, login into your Shopify admin, then look at the URL in your browser's address bar, the subdomain portion (e.g., example_store_id.myshopify.com) is your store ID (in this case: example_store_id) Order Filtering – Ensures only valid orders are fulfilled, particularly useful if: You sell digital downloads or gift cards exclusively. You use third-party fulfillment services for all products. Credentials To run this workflow, you'll need: Shopify API Key – Required for authentication. Who Is This For? Shopify store owners looking to automate their fulfillment process. Merchants selling digital or personalized products who need instant fulfillment. Explore More Templates 👉 Check out my other n8n templates
by Miquel Colomer
Disclaimer: This template contains a community node and therefore only works for n8n self-hosted users== This is Miquel from Aprende n8n and Automate with n8n. We have created a new community node Document Generator that generates dynamic content using templates. You can compose complex content with no SETs or FUNCTION ITEMs nodes using this node, like: Send one email with a list of items in the body (i.e., one email with the last entries of an RSS feed). Send one email per item (i.e., one invoice per email). Emails are just a sample. You can create complex dynamic content to: Send messages to Telegram/Slack. Create WordPress entries. Create HTML pages for your website. Create tickets. And more! The sky is your limit ;) If you want to use this workflow, install the community node n8n-nodes-document-generator from Settings > Community nodes. Type "n8n-nodes-document-generator", check "I understand the risks..." and click on "Install": Later, copy and paste this workflow into your n8n. You will get this workflow: This workflow uses the Customer Datastore node to generate sample input items. You can render one template with all items (enable "Render All Items with One Template"): or one template per input item: Visit the oficial NPM page to see more samples. Learning n8n by yourself is nice, but a bit tricky :) We offer n8n training video courses at Aprende n8n. If you need custom trainings, let us know. Additionally, you can contact us at Automate with n8n if you need the next services: Custom installations. Custom nodes. Monitor and alarms. Delegate 12/5 or 24/7 workflow issue resolutions. Automated backups of your workflows. HTTP integrations of non-supported APIs. Complex workflows. I hope you will enjoy this new node and this workflow. Automate your life! Automate it with n8n!
by Yaron Been
Description This workflow automatically searches multiple flight booking websites to find the cheapest flights for your desired routes. It leverages web scraping to compare prices across platforms, helping you save money on air travel. Overview This workflow automatically searches multiple flight booking websites to find the cheapest flights for your desired routes. It uses Bright Data to scrape flight prices and can notify you when prices drop below your target threshold. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping flight prices from booking websites. Notification Services:** Email, SMS, or other messaging platforms. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Notifications: Configure your preferred notification method. Customize: Set your routes, date ranges, and price thresholds. Use Cases Frequent Travelers:** Find the best deals for your regular routes. Travel Agencies:** Monitor flight prices for client bookings. Budget Travelers:** Get notified when flights to your dream destination become affordable. Connect with Me YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #travel #flights #brightdata #dealalerts #webscraping #flightdeals #cheapflights #travelhacks #budgettravel #travelplanning #airfare #flightprices #travelautomation #n8nworkflow #workflow #nocode #traveltech #flightbooking #savemoney #traveltools #flightcomparison #bestflightdeals #travelsmarter #automatedtravel
by n8n Team
This workflow automatically imports data from a CSV file located at a specific URL and then updates the Google Sheets document with the provided data. Below is a step-by-step description of what this workflow does: The workflow is started manually using the "When you click 'Execute Workflow'" node. The CSV file is then uploaded from the specified URL "https://opendata.ecdc.europa.eu/covid19/testing/csv/data.csv" using the "Upload CSV" node. The "Import CSV" node accepts the uploaded CSV file and converts it into JSON formatted data. The "Add Unique Field" node generates a unique key by combining the 'country_code' and 'year_week' fields from the JSON data, which will be further used in the Google Sheets document. The 'Keep only DACH in 2023' node filters the data to keep only records where 'country_code' is either 'DE', 'AT', or 'CH' and 'year_week' starts with '2023'. Google's API has limitations on the speed of read and write operations, so only a subset of the data is taken. The filtered data is loaded into the specified Google Sheets document via the 'Load to Spreadsheet' node. The operation is set to 'appendOrUpdate' and the document ID and sheet name are specified. Also, the previously generated 'unique_key' key is set as the key to match the columns.
by Yaron Been
Aihilums Sehatsanjha AI Generator Description None Overview This n8n workflow integrates with the Replicate API to use the aihilums/sehatsanjha model. This powerful AI model can generate high-quality other content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Optional Parameters cookie** (string, default: None): Cookie to be returned unchanged user_id** (string, default: ): Unique session identifier audio_file** (string, default: None): Audio file user_state** (string, default: None): User state end_session** (boolean, default: False): End the current recording new_session** (boolean, default: False): Start a new recording How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate other content Access the generated output from the final node API Reference Model: aihilums/sehatsanjha API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of other generation parameters
by Tom
This workflow parses content from a website (for this example, Baserow's release page) and creates an RSS feed based on the extracted data. Prerequisites Some familiarity with HTML and CSS selectors Nodes Webhook node triggers the workflow when new content (a new Baserow release) is published on a website. Set nodes set the required URLs and links for the RSS feed. HTTP Request node fetches data from a specified website page. HTML Extract nodes extract the posts and their fields (such as date, title, description, and link) from the website. Item Lists node iterates over each post on the page. Date & Time node converts the date of the post to a different format. Function Item node creates RSS items for each post. Function node creates the response code for the RSS feed. Respond to Webhook node returns the RSS feed in response to the Webhook node. The result of this workflow would look like this:
by Yaron Been
Description This workflow automatically monitors and tracks trending topics across multiple platforms and websites. It helps content creators and marketers stay ahead of the curve by identifying emerging trends before they go mainstream. Overview This workflow automatically monitors and tracks trending topics across multiple platforms and websites. It uses Bright Data to scrape trend data from social media, news sites, and other sources, then compiles the information into a structured format. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping trend data from various websites without getting blocked. Spreadsheets/Databases:** For storing and analyzing trend information. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Data Storage: Configure where you want to store the trend data. Customize: Specify which platforms to monitor and what topics to focus on. Use Cases Content Creators:** Stay on top of trending topics for content ideas. Marketers:** Identify emerging trends for timely campaigns. Researchers:** Track the evolution of topics and conversations over time. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #trends #trendtracking #brightdata #contentmarketing #trendanalysis #trendalerts #markettrends #trendmonitoring #n8nworkflow #workflow #nocode #trendresearch #emergingtrends #socialmediatrends #trendscraping #trenddata #contentideas #digitalmarketing #marketresearch #trendforecasting #trendspotting #dataanalysis #marketintelligence #trendautomation
by Yaron Been
Fire Part Crafter Image Generator Description PartCrafter is a structured 3D mesh generation model that creates multiple parts and objects from a single RGB image. Overview This n8n workflow integrates with the Replicate API to use the fire/part-crafter model. This powerful AI model can generate high-quality image content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters image** (string): Input image for 3D mesh generation Optional Parameters seed** (integer, default: 0): Random seed for reproducibility. Use 0 for random seed num_parts** (integer, default: 16): Number of parts to generate num_tokens** (string, default: 2048): Number of tokens for generation guidance_scale** (number, default: 7): Guidance scale for generation remove_background** (boolean, default: False): Remove background from input image use_flash_decoder** (boolean, default: False): Use flash decoder for faster inference (Tempermental?) num_inference_steps** (integer, default: 50): Number of inference steps How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate image content Access the generated output from the final node API Reference Model: fire/part-crafter API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of image generation parameters
by Jacob
Tired of manually watching long YouTube videos just to extract the main points? With our YouTube Transcript & Summary Automation, you can instantly turn any video into an actionable, AI-generated summary — all from a simple Google Sheet. What this automation does: Reads video URLs from Google Sheets (just add your links!) Generates accurate transcripts using Supadata.ai — with 100 free uses/month Creates a smart summary using DeepSeek AI: 🔹 Short summary of the video 🔹 Key points 🔹 Main topics Youtube tutorial: https://www.youtube.com/watch?v=uj7KaFSh95Y Automatically updates your Google Sheet with the transcript and the AI-generated summary How to set it up: Create a simple Google Sheet with these columns: Url – link to the YouTube video Status – set to pending to trigger the automation, updated to done after completion Transcript – filled automatically Summary – filled automatically Once your sheet is ready, the automation takes care of the rest — no technical skills needed. Why you'll love it: This is the perfect tool for content creators, researchers, marketers, and educators who want to save time, extract insights faster, and never miss key ideas from long videos. Want something custom? Got feedback or want to build your own custom automation or workflow? We’d love to hear from you! Reach out to us at jacobmarketingservice@gmail.com — let’s bring your idea to life.
by Paul Taylor
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 📄 Post New Articles from Feeds to Slack Channel 🧠 What This Workflow Does This workflow automates the discovery and sharing of fresh articles from a curated list of RSS feeds. It performs the following steps: Reads a list of RSS feed URLs from a Google Sheet (Feeds tab). Fetches the latest articles from each feed. Checks for duplicates against previously published links stored in another sheet (Posted Articles tab). Filters out already shared articles. Posts the new articles to a designated Slack channel with formatted titles and links. Logs the newly shared articles back into the Google Sheet to prevent duplicates. 🛠️ Prerequisites To use this workflow, you must have: ✅ Google Sheets OAuth2 credentials set up in n8n (Used to access and update the RSS feed and post history sheets) ✅ Slack OAuth2 credentials (Used to post messages to a specific Slack channel) ✅ A Google Spreadsheet with: Feeds tab – Columns: title, link Posted Articles tab – Columns: title, link, pubDate 🔧 Environment Variables or Custom Values You will need to set the following n8n variable or replace with direct input: {{$vars.Daily_Industry_News_Automation_Google_Sheet}}: Reference to the Google Sheet Document ID (you can use a static ID if preferred) Also update: Slack channelId: Replace with your actual Slack channel ID if not dynamically referenced ⏰ Trigger & Scheduling Trigger type**: Cron node Default schedule: Every day at **7:00 AM You can modify this in the “Trigger Workflow” node to suit your own schedule. 🎯 Intended Use Case This workflow is ideal for: Marketing teams curating daily or weekly news digests Founders or industry professionals monitoring sector updates Automating internal Slack news updates Avoiding duplicate content when sourcing from multiple feeds
by Yaron Been
Fire Flux Image Generator Description The image generation model tailored for local development and personal use Overview This n8n workflow integrates with the Replicate API to use the fire/flux model. This powerful AI model can generate high-quality image content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Prompt for generated image Optional Parameters seed** (integer, default: 0): Random seed. Set for reproducible generation go_fast** (boolean, default: True): Run faster predictions with model optimized for speed (currently fp8 quantized); disable to run in original bf16 megapixels** (string, default: 1): Approximate number of megapixels for generated image num_outputs** (integer, default: 1): Number of outputs to generate aspect_ratio** (string, default: 2:1): Aspect ratio for the generated image output_format** (string, default: png): Format of the output images output_quality** (integer, default: 80): Quality when saving the output images, from 0 to 100. 100 is best quality, 0 is lowest quality. Not relevant for .png outputs num_inference_steps** (integer, default: 4): Number of denoising steps. 4 is recommended, and lower number of steps produce lower quality outputs, faster. disable_safety_checker** (boolean, default: False): Disable safety checker for generated images. How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate image content Access the generated output from the final node API Reference Model: fire/flux API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of image generation parameters
by Kyle Morse
Takes your raw, unpolished voice transcripts and transforms them into well-structured LinkedIn posts using AI. Perfect for when you have good ideas but they come out as rambling thoughts. The Problem: You record voice memos with great ideas, but when you read the transcript, it's full of "ums," incomplete sentences, and scattered thoughts. Turning that into a professional LinkedIn post takes forever. The Solution: Email your raw transcript to this workflow. It combines your unpolished content with examples from your inspiration document (posts you've saved that match your desired style), then uses AI to create a clean, engaging LinkedIn post. What actually happens: You email a raw voice transcript to your workflow email -The workflow pulls style examples from your Google Doc AI reformats your scattered thoughts into a coherent 150-300 word LinkedIn post You get an email back with the polished content + suggested image description Copy, paste, and post to LinkedIn You provide: The raw transcript (from your phone's voice recorder or any transcription tool), a Google Doc with LinkedIn posts you admire for style reference. You get: Professional LinkedIn content that sounds like you, but organized and polished. Technical requirements: Anthropic API, email account, Google Doc with example posts. This is basically having an AI writing assistant that knows your voice and preferred style, turning your brain dumps into professional content.