by Emmanuel Bernard
🎉 Do you want to master AI automation, so you can save time and build cool stuff? I’ve created a welcoming Skool community for non-technical yet resourceful learners. 👉🏻 Join the AI Atelier 👈🏻 Keeping your YouTube video descriptions updated and consistent across your channel can be a daunting task. Manually editing each video is not only time-consuming but also prone to errors. 📋 Blog post 📺 Youtube Video This workflow streamlines this process, allowing you to maintain a shared section in all your video descriptions and effortlessly update them all at once. By incorporating a unique identifier, you can automate updates across your entire channel, keeping your content fresh and relevant with minimal effort. How it Works Define Your Unique Delimiter:** Choose your unique delimiter (e.g., "---n8ninja---"). It will be visible, so select something appropriate for your audience. Automate Updates:** Anything below the delimiter can be automatically updated by this workflow. Configure Text Updates:** Set the text you wish to add to every video description in the configuration node. Getting Started Integrate Google (YouTube) Credentials:** Securely add your credentials to enable API access. Set Up the Configuration Node:** Define your delimiter and the text for the shared section you wish to append to your video descriptions. Prepare Your Videos:** Add the chosen delimiter to all videos you want to update automatically. Execute the Workflow:** Run the workflow whenever you wish to batch update the descriptions of your videos. Created by the n8ninja ✨ follow on X 📺 follow on YT
by Yaron Been
Description This workflow automatically searches multiple flight booking websites to find the cheapest flights for your desired routes. It leverages web scraping to compare prices across platforms, helping you save money on air travel. Overview This workflow automatically searches multiple flight booking websites to find the cheapest flights for your desired routes. It uses Bright Data to scrape flight prices and can notify you when prices drop below your target threshold. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping flight prices from booking websites. Notification Services:** Email, SMS, or other messaging platforms. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Notifications: Configure your preferred notification method. Customize: Set your routes, date ranges, and price thresholds. Use Cases Frequent Travelers:** Find the best deals for your regular routes. Travel Agencies:** Monitor flight prices for client bookings. Budget Travelers:** Get notified when flights to your dream destination become affordable. Connect with Me YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #travel #flights #brightdata #dealalerts #webscraping #flightdeals #cheapflights #travelhacks #budgettravel #travelplanning #airfare #flightprices #travelautomation #n8nworkflow #workflow #nocode #traveltech #flightbooking #savemoney #traveltools #flightcomparison #bestflightdeals #travelsmarter #automatedtravel
by n8n Team
This workflow automatically imports data from a CSV file located at a specific URL and then updates the Google Sheets document with the provided data. Below is a step-by-step description of what this workflow does: The workflow is started manually using the "When you click 'Execute Workflow'" node. The CSV file is then uploaded from the specified URL "https://opendata.ecdc.europa.eu/covid19/testing/csv/data.csv" using the "Upload CSV" node. The "Import CSV" node accepts the uploaded CSV file and converts it into JSON formatted data. The "Add Unique Field" node generates a unique key by combining the 'country_code' and 'year_week' fields from the JSON data, which will be further used in the Google Sheets document. The 'Keep only DACH in 2023' node filters the data to keep only records where 'country_code' is either 'DE', 'AT', or 'CH' and 'year_week' starts with '2023'. Google's API has limitations on the speed of read and write operations, so only a subset of the data is taken. The filtered data is loaded into the specified Google Sheets document via the 'Load to Spreadsheet' node. The operation is set to 'appendOrUpdate' and the document ID and sheet name are specified. Also, the previously generated 'unique_key' key is set as the key to match the columns.
by Yaron Been
Aihilums Sehatsanjha AI Generator Description None Overview This n8n workflow integrates with the Replicate API to use the aihilums/sehatsanjha model. This powerful AI model can generate high-quality other content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Optional Parameters cookie** (string, default: None): Cookie to be returned unchanged user_id** (string, default: ): Unique session identifier audio_file** (string, default: None): Audio file user_state** (string, default: None): User state end_session** (boolean, default: False): End the current recording new_session** (boolean, default: False): Start a new recording How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate other content Access the generated output from the final node API Reference Model: aihilums/sehatsanjha API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of other generation parameters
by Yaron Been
Description This workflow automatically generates comprehensive property market reports by scraping real estate listings and market data from multiple sources. It helps real estate professionals save time and provide data-driven insights to clients without manual research. Overview This workflow automatically generates property market reports by scraping real estate listings and market data. It uses Bright Data to access multiple real estate websites and compiles the data into comprehensive reports. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping real estate websites and property data without getting blocked. Spreadsheets/Databases:** For storing and analyzing property data. Document Generation:** For creating professional PDF reports. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Data Storage: Configure where you want to store the property data. Customize: Specify locations, property types, and report format. Use Cases Real Estate Agents:** Generate market reports for clients. Property Investors:** Track market trends in target areas. Market Analysts:** Automate data collection for property market analysis. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #realestate #propertymarket #brightdata #marketreports #propertyanalysis #realestatedata #markettrends #propertyinvestment #n8nworkflow #workflow #nocode #realestateanalysis #propertyreports #realestateintelligence #marketresearch #propertyscraping #realestateautomation #investmentanalysis #propertytrends #datadriven #realestatetech #propertyinsights #marketanalysis #realestateinvesting
by Yaron Been
Description This workflow automatically monitors and tracks trending topics across multiple platforms and websites. It helps content creators and marketers stay ahead of the curve by identifying emerging trends before they go mainstream. Overview This workflow automatically monitors and tracks trending topics across multiple platforms and websites. It uses Bright Data to scrape trend data from social media, news sites, and other sources, then compiles the information into a structured format. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping trend data from various websites without getting blocked. Spreadsheets/Databases:** For storing and analyzing trend information. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Data Storage: Configure where you want to store the trend data. Customize: Specify which platforms to monitor and what topics to focus on. Use Cases Content Creators:** Stay on top of trending topics for content ideas. Marketers:** Identify emerging trends for timely campaigns. Researchers:** Track the evolution of topics and conversations over time. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #trends #trendtracking #brightdata #contentmarketing #trendanalysis #trendalerts #markettrends #trendmonitoring #n8nworkflow #workflow #nocode #trendresearch #emergingtrends #socialmediatrends #trendscraping #trenddata #contentideas #digitalmarketing #marketresearch #trendforecasting #trendspotting #dataanalysis #marketintelligence #trendautomation
by Yaron Been
Description This workflow automatically collects weather data from multiple sources and compiles it into comprehensive reports. It helps you make informed decisions based on accurate weather forecasts without manually checking multiple weather services. Overview This workflow automatically scrapes weather data from multiple sources and compiles it into a comprehensive report. It uses Bright Data to access weather websites and can be configured to send you regular weather updates for your locations of interest. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping weather websites and forecast data without getting blocked. Notification Services:** Email, messaging apps, or other platforms. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Notifications: Configure how you want to receive weather reports. Customize: Add your locations of interest and reporting frequency. Use Cases Event Planners:** Get weather forecasts for upcoming outdoor events. Farmers:** Monitor weather conditions for agricultural planning. Travelers:** Check weather forecasts for destinations before trips. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #weather #weatherforecasts #brightdata #webscraping #weatherreports #weatheralerts #weatherdata #weathermonitoring #n8nworkflow #workflow #nocode #weatherautomation #weatherscraping #weathertracking #weathernotifications #weatherupdates #forecastdata #weatherplanning #weatherservice #outdoorevents #weatherapi #weatherinformation #climatedata #weathertech
by Yaron Been
Fire Part Crafter Image Generator Description PartCrafter is a structured 3D mesh generation model that creates multiple parts and objects from a single RGB image. Overview This n8n workflow integrates with the Replicate API to use the fire/part-crafter model. This powerful AI model can generate high-quality image content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters image** (string): Input image for 3D mesh generation Optional Parameters seed** (integer, default: 0): Random seed for reproducibility. Use 0 for random seed num_parts** (integer, default: 16): Number of parts to generate num_tokens** (string, default: 2048): Number of tokens for generation guidance_scale** (number, default: 7): Guidance scale for generation remove_background** (boolean, default: False): Remove background from input image use_flash_decoder** (boolean, default: False): Use flash decoder for faster inference (Tempermental?) num_inference_steps** (integer, default: 50): Number of inference steps How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate image content Access the generated output from the final node API Reference Model: fire/part-crafter API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of image generation parameters
by Harshil Agrawal
This workflow demonstrates how to use noItemsLeft to check if there are items left to be processed by the SplitInBatches node. Function node: This node generates mock data for the workflow. Replace it with the node whose data you want to split into batches. SplitInBatches node: This node splits the data with the batch size equal to 1. Based on your use-case, set the value of the Batch Size. IF node: This node check if all the data by the SplitInBatches are not processed or not. It uses the expression {{$node["SplitInBatches"].context["noItemsLeft"]}} which returns a boolean value. If there is data yet to be processed, the expression will return false, otherwise true. Set node: This node prints a message No Items Left. Based on your use-case, connect the false output of the IF node to the input of the node you want to execute, after the data is processed by the SplitInBatches node.
by Open Paws
This sub-workflow uses two custom Hugging Face regression models from Open Paws to evaluate and predict the real-world performance and advocacy alignment of text content. It’s designed to support animal advocacy organizations in optimizing their messaging across platforms like social media, email campaigns, and more. 🛠️ What It Does Sends input text to two deployed Hugging Face endpoints: Predicted Performance Model – Estimates real-world content success (e.g., engagement, shares, opens) based on patterns from real online data. Advocate Preference Model – Predicts how well the content will resonate with animal advocates (emotional impact, relevance, rationality, etc.) Outputs structured scores for both models Can be integrated into larger workflows for automated content review, filtering, or revision 📊 About the Models Text Performance Prediction Model** Trained on real-world data from 30+ animal advocacy organizations, this model predicts actual online performance of content—including social media, email marketing, and other outreach channels. Advocate Preference Prediction Model** Trained on ratings from animal advocates to evaluate how well a piece of text aligns with advocacy goals and values. Model Repositories: open-paws/text_performance_prediction_longform open-paws/animal_advocate_preference_prediction_longform > 📌 You must deploy each model as an inference endpoint on Hugging Face. Click "Deploy" on each model’s repo, then add the endpoint URL and your Hugging Face access token using n8n credentials. 📦 Use Cases Advocacy content review before publishing Automated scoring of outreach messages Filtering or flagging content with low predicted impact A/B testing support for message optimization
by Yaron Been
Google Veo 3 Video Generator Description Sound on: Google’s flagship Veo 3 text to video model, with audio Overview This n8n workflow integrates with the Replicate API to use the google/veo-3 model. This powerful AI model can generate high-quality video content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Text prompt for video generation Optional Parameters seed** (integer, default: None): Random seed. Omit for random generations resolution** (string, default: 720p): Resolution of the generated video negative_prompt** (string, default: None): Description of what to discourage in the generated video How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate video content Access the generated output from the final node API Reference Model: google/veo-3 API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of video generation parameters
by Jacob
Tired of manually watching long YouTube videos just to extract the main points? With our YouTube Transcript & Summary Automation, you can instantly turn any video into an actionable, AI-generated summary — all from a simple Google Sheet. What this automation does: Reads video URLs from Google Sheets (just add your links!) Generates accurate transcripts using Supadata.ai — with 100 free uses/month Creates a smart summary using DeepSeek AI: 🔹 Short summary of the video 🔹 Key points 🔹 Main topics Youtube tutorial: https://www.youtube.com/watch?v=uj7KaFSh95Y Automatically updates your Google Sheet with the transcript and the AI-generated summary How to set it up: Create a simple Google Sheet with these columns: Url – link to the YouTube video Status – set to pending to trigger the automation, updated to done after completion Transcript – filled automatically Summary – filled automatically Once your sheet is ready, the automation takes care of the rest — no technical skills needed. Why you'll love it: This is the perfect tool for content creators, researchers, marketers, and educators who want to save time, extract insights faster, and never miss key ideas from long videos. Want something custom? Got feedback or want to build your own custom automation or workflow? We’d love to hear from you! Reach out to us at jacobmarketingservice@gmail.com — let’s bring your idea to life.