by n8n Team
This workflow demonstrates how to export SQL to XML and present the data nicely formatted using an XSL Template. The upper part of the workflow starts with a webhook. Then it gets several random records from the SQL table and converts them into an XML string. Then a final XML file is created that contains a link to the XML Stylesheet file. The lower part of the workflow contains a helper Webhook that reads an XSL Template from the GitHub gist and serves it back via the Respond to Webhook node. This is required to comply with the CORS rules of modern browsers. These rules dictate that both XML data and a stylesheet file should come from the same domain.
by Yaron Been
Bytedance Seededit 3.0 Image Generator Description Text-guided image editing model that preserves original details while making targeted modifications like lighting changes, object removal, and style conversion Overview This n8n workflow integrates with the Replicate API to use the bytedance/seededit-3.0 model. This powerful AI model can generate high-quality image content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Text prompt for image generation image** (string): Input image to edit Optional Parameters seed** (integer, default: None): Random seed. Set for reproducible generation guidance_scale** (number, default: 5.5): Prompt adherence. Higher = more literal. How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate image content Access the generated output from the final node API Reference Model: bytedance/seededit-3.0 API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of image generation parameters
by Jonathan | NEX
Stop manually checking suspicious links. This free n8n workflow provides the foundation for a powerful, automated URL analysis pipeline. Using the NixGuard AI engine, you can instantly analyze suspicious URLs from emails, logs, or tickets to uncover phishing attempts, malware hosting sites, and malicious redirects. What You Will Automate: 🤖 Instant Threat Triage: Get an immediate AI-powered summary of why a URL is malicious, saving you critical investigation time. 🎯 Actionable IOC Extraction: Automatically extract the final redirected URL, malicious domains, and IPs to fuel your threat hunting and blocking rules. 🚀 SOAR-Ready Foundation: This workflow is the perfect starting point for your security playbooks. Use the output to: Alert: Send instant notifications to Slack or Teams. Respond: Create tickets in Jira or TheHive. Block: Add malicious domains to your firewall or DNS filter. Download this free template and automate your first line of defense against web-based threats in minutes! Don't have the main workflow yet? Get it HERE! 🔗 Learn more about NixGuard: thenex.world 🔗 Get started with a free security subscription: thenex.world/security/subscribe For search: URL Scanning, Phishing, Threat Intelligence, SOAR, SOC Automation, NixGuard, Free, AI, Incident Response, Cybersecurity, Automation, Link Analysis, MTTR, Malware, VirusTotal
by n8n Team
This workflow is designed for dynamic and intelligent conversational capabilities. It incorporates OpenAI's GPT-4o model for natural language understanding and generation. Additional tools include SerpAPI and Wikipedia for enriched, data-driven responses. The workflow is triggered manually, and utilizes a 'Window Buffer Memory' to maintain the context of the last 20 interactions for better conversational continuity. All these components are orchestrated through n8n nodes, ensuring seamless interconnectivity. To use this template, you need to be on n8n version 1.50.0 or later.
by Belmont Digital
This n8n workflow verifies the deliverability of mailing addresses stored in HighLevel by integrating with Lob's address verification service. Who is this for? This template is designed for HighLevel users who need to ensure the accuracy of mailing addresses stored in their CRM systems. What problem is this workflow solving? / Use Case This workflow addresses the challenge of maintaining accurate mailing addresses in CRM databases by verifying the deliverability of addresses. What this workflow does A new contact is created in HighLevel Webhook sent to n8n Verify if the address is deliverable via LOB Report back to HighLevel Set Up Steps Watch this setup video: https://www.youtube.com/watch?v=T7Baopubc-0 Takes 10-30 minutes to set up Accounts Needed: HighLevel LOB Account (https://www.lob.com $0.00/mo 300 US addresses Verifications) n8n Before using this template, ensure you have API keys for your HighLevel app and Lob. Set up authentication for both services within n8n. How to customize this workflow to your needs You can customize this workflow by adjusting the trigger settings to match HighLevel's workflow configuration. Additionally, you can modify the actions taken based on the deliverability outcome, such as updating custom fields or sending notifications.
by LuisBetancourt.co
Description Whenever a Zoom “Meeting assets” email arrives in your Gmail inbox, this workflow will: 1) Trigger on new Gmail messages filtered by the subject “Meeting assets”. 2) Extract from the email (HTML or plain text): 3) Type of session (e.g. “1 hour”, “2 hours”, or “exploratory call”). Client’s full name. Session date & time (from the GMT… timestamp). Duration (HH:MM:SS). Recording link. Quick summary. Detailed summary. List of next steps. 4) Lookup the client in your Master Airtable base, table People, by full name. 5) Send a personalized Gmail to the client with all extracted details. 6) Create a new record in your Sessions table in Airtable, linking back to that client. Quick Start Import this JSON into n8n as a new workflow. Connect your Gmail credentials (OAuth2). Connect your Airtable credentials (Personal Access Token). In the Search Records node: Base → your Master base ID. Table → “Your people table”. Filter By Formula → ={Full Name} = '{{ $json.clientName }}'. In the Create Record node: Table → “Sessions”. Map each field (dateTime, duration, summaries, next steps, client link). Activate the workflow. Prerequisites n8n v1.50 or higher A Gmail account with OAuth2 credentials configured An Airtable base containing: Table People with a Full Name field (and email). Table Sessions with fields: DateTime, Duration, Quick Summary, Detailed Summary, Next Steps, and a Linked Record to People. An Airtable Personal Access Token with read/write access to that base. Tips & Extensions Timezone conversion: Use a Function node with moment-timezone to convert UTC if needed. Error handling: Add a catch node to log or notify if any field fails to parse. Alternate notifications: Swap the Gmail node for Slack, Microsoft Teams, or SMS integrations. With this documentation, your team can import and deploy the workflow in minutes. Enjoy!
by Samir Saci
Tags: Productivity, Education, Learning, Language Context I’m a Supply Chain Data Scientist from Paris who lived six years in China — and yes, learning Mandarin while working full-time was tough. Learning Mandarin as an adult can be very difficult, especially if you have a full-time job. With AI, you can now have a Chinese tutor available 24/7 on your phone — no excuses left! It is with this spirit that I designed this workflow to support fellow Mandarin learners with a Chinese Teacher powered by GPT-4o. >Boost your language skills with AI using N8N! 📬 For business inquiries, you can add me on LinkedIn Who is this template for? This workflow template is designed for language learners and educators who need support to learn a vocabulary list in Mandarin (or any other language) using Open AI GPT-4o, an AI agent and a Telegram Bot to interact with users. For the vocabulary list, you can use another template shared in my profile 🉑 Generate Anki Flash Cards for Language Learning with Google Translate and GPT-4o to generate the Google Sheet needed in this workflow. How does it work? The workflow loads a vocabulary list stored in your Google Sheet. The bot will: 📥 Load your vocabulary list from Google Sheets 🧠 Generate multiple-choice questions with GPT-4o ✅ Evaluate your answer and give instant feedback 🔁 Loop to the next word until you're fluent These fields will be automatically added to a Google Sheet, ready to be loaded in Anki to create flash cards. What do I need to start? This workflow does not require any advanced programming skills. Prerequisite A Google Drive Account with a folder including a Google Sheet filled with the vocabulary list you want to learn. API Credentials: Open AI API for the chat model, Google Drive API and Google Sheets API activated with OAuth2 credentials A Telegram Bot with its token recorded in the Telegram Node Credentials A Google Sheet** with two columns (initialText: words in your own language, translatedText: words in the target language) Next Steps Follow the sticky notes to set up the parameters inside each node and get ready to pump your learning skills. 🎥 Watch My Tutorial 🚀 Curious how N8N can supercharge learning or supply chain? 📬 Let’s connect on LinkedIn Notes This workflow can be used for any language. In the AI Agent prompt, you just need to replace Chinese with your language. This workflow has been created with N8N 1.82.1 Submitted: March 23th, 2025
by Niko
Capture URL Screenshots Automatically from Google Sheets & Drive with ScreenshotOne & Gmail Alerts Summary This automation template streamlines the process of capturing screenshots for multiple URLs. Instead of manually visiting each URL, taking a screenshot, and organizing the results, this workflow automates everything. When a spreadsheet is added to a designated Google Drive folder, the template extracts URLs from the column named "Url." These URLs are then processed through ScreenshotOne to capture screenshots, which are saved back to the same folder. Finally, an email notification is sent via Gmail with a link to the folder containing the screenshots. Problem Solved This template addresses the challenge of manual screenshot capture for multiple URLs. Without this automation, a user would need to: Open each URL from a spreadsheet. Take a screenshot manually. Save each screenshot with an appropriate name. Organize the screenshots in a folder. Notify stakeholders when the process is complete. These steps are not only time-consuming but also repetitive, especially when handling a large number of URLs. Who Can Benefit: Digital Marketers:** Monitor website appearances for competitive analysis or to track campaign landing pages. Web Developers/Designers:** Capture screenshots of multiple websites for inspiration or reference. QA Teams:** Document the visual state of web pages during various stages of development. SEO Specialists:** Track visual changes to websites they are optimizing. Content Managers:** Monitor how content appears across various web properties. Prerequisites Google Drive Node:** Must have appropriate permissions to create and access folders. Connected Google Sheets Node:** To extract URLs from the spreadsheet. Authenticated Gmail Node:** For sending notifications. ScreenshotOne Account:* Either a free or paid plan depending on volume needs, along with an *Access key**. Ensure you replace the placeholder --YOUR ACCESS KEY-- with your generated access key in the "Get Screenshots" node. Workflow Details Step 1: Google Drive Integration Trigger Node:** Monitors a specific folder in Google Drive. When a spreadsheet is added, the workflow is initiated. Step 2: Google Sheets Processing Google Sheets Node:** Extracts URLs from the column named "Url." Step 3: Screenshot Capture Get Screenshots Node:** Sends each extracted URL to ScreenshotOne to capture screenshots. Step 4: Saving Screenshots and Notifications Google Drive Node:** Saves the captured screenshots back into the same folder. Gmail Node:** Sends an email notification with a link to the folder, alerting stakeholders that the screenshots are ready. Customization Guidance Folder Monitoring: The workflow is set to monitor a specific Google Drive folder. It can be customized by selecting a different folder in the node settings. Spreadsheet Structure: While the template expects a spreadsheet with a column named "Url." for extracting URLs, users can add additional columns (e.g., titles, categories, or tags) and modify the workflow to utilize them as needed. Email Settings: Customize the recipient, subject, and body of the notification email to suit your needs. If required, enable optional notifications for different stakeholders. ScreenshotOne Access Key & Configurations: A valid ScreenshotOne Access key is required to capture screenshots. Users can further refine screenshot settings (e.g., viewport size, device emulation, or delay timing) by exploring the available options in the ScreenshotOne API documentation.
by Abbas Ali
This automation fetches the latest article from a WordPress blog, summarizes it using OpenAI, and sends the summary to a list of subscribers via email. Ideal for content creators and bloggers who want to distribute digestible content without manual effort. Use Case Perfect for: • Newsletter creators • Content marketers • Bloggers • Knowledge managers Nodes Used • Schedule Trigger • HTTP Request • Set • OpenAI • Google Sheets • Email (Gmail/SMTP) • IF • SplitInBatches Workflow Steps Trigger: Starts on a schedule (e.g., daily at 9:00 AM). Fetch Blog Post: Retrieves the most recent post from a WordPress blog via HTTP Request. Extract Fields: A Set node extracts the title, link, and content. Summarize Article: OpenAI processes the article and returns a 3-point summary. Fetch Subscribers: Google Sheets reads email addresses from a subscriber list. Loop Emails: SplitInBatches and Send Email nodes loop through subscribers. Conditional Logic: IF node skips articles shorter than 300 words. Credentials Required • OpenAI API Key (for content summarization) • Google Sheets OAuth2 (to read subscriber emails) • Gmail or SMTP (for sending emails) Test Instructions Replace blog URL in HTTP Request node. Connect OpenAI API key. Link your Google Sheet with a column named Email. Set up Gmail or SMTP credentials. Run manually for testing, then activate schedule.
by Oneclick AI Squad
This n8n workflow monitors email alerts for disk utilization exceeding 80%, extracts the server IP, logs into the server, and purges logs from Nginx, PM2, Docker, and system files to clear disk space. Key Insights Ensure email alerts are consistently formatted with server IP details. SSH access must be properly configured to avoid authentication failures. Workflow Process Initiate the workflow with the Check Disk Alert Emails node when an email triggers on high disk usage. Parse the email to extract the server IP using the Extract Server IP from Email node. Set up SSH credentials and paths manually with the Prepare SSH Variables node. Execute cleanup commands to delete logs from Nginx, PM2, Docker, and system files using the Run LogCleanup Commands via SSH node. Usage Guide Import the workflow into n8n and configure email and SSH credentials. Test with a sample email alert to verify IP extraction and log deletion. Prerequisites Email service (e.g., IMAP or API) for alert monitoring SSH access with valid credentials Customization Options Modify the Prepare SSH Variables node to target specific log directories or adjust cleanup commands for different server setups.
by Yaron Been
Description This workflow automatically generates comprehensive property market reports by scraping real estate listings and market data from multiple sources. It helps real estate professionals save time and provide data-driven insights to clients without manual research. Overview This workflow automatically generates property market reports by scraping real estate listings and market data. It uses Bright Data to access multiple real estate websites and compiles the data into comprehensive reports. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping real estate websites and property data without getting blocked. Spreadsheets/Databases:** For storing and analyzing property data. Document Generation:** For creating professional PDF reports. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Data Storage: Configure where you want to store the property data. Customize: Specify locations, property types, and report format. Use Cases Real Estate Agents:** Generate market reports for clients. Property Investors:** Track market trends in target areas. Market Analysts:** Automate data collection for property market analysis. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #realestate #propertymarket #brightdata #marketreports #propertyanalysis #realestatedata #markettrends #propertyinvestment #n8nworkflow #workflow #nocode #realestateanalysis #propertyreports #realestateintelligence #marketresearch #propertyscraping #realestateautomation #investmentanalysis #propertytrends #datadriven #realestatetech #propertyinsights #marketanalysis #realestateinvesting
by Emmanuel Bernard
🎉 Do you want to master AI automation, so you can save time and build cool stuff? I’ve created a welcoming Skool community for non-technical yet resourceful learners. 👉🏻 Join the AI Atelier 👈🏻 Keeping your YouTube video descriptions updated and consistent across your channel can be a daunting task. Manually editing each video is not only time-consuming but also prone to errors. 📋 Blog post 📺 Youtube Video This workflow streamlines this process, allowing you to maintain a shared section in all your video descriptions and effortlessly update them all at once. By incorporating a unique identifier, you can automate updates across your entire channel, keeping your content fresh and relevant with minimal effort. How it Works Define Your Unique Delimiter:** Choose your unique delimiter (e.g., "---n8ninja---"). It will be visible, so select something appropriate for your audience. Automate Updates:** Anything below the delimiter can be automatically updated by this workflow. Configure Text Updates:** Set the text you wish to add to every video description in the configuration node. Getting Started Integrate Google (YouTube) Credentials:** Securely add your credentials to enable API access. Set Up the Configuration Node:** Define your delimiter and the text for the shared section you wish to append to your video descriptions. Prepare Your Videos:** Add the chosen delimiter to all videos you want to update automatically. Execute the Workflow:** Run the workflow whenever you wish to batch update the descriptions of your videos. Created by the n8ninja ✨ follow on X 📺 follow on YT