by Oriol Seguí
This fun workflow automates the generation and delivery of personalized jokes by email based on the names or objects entered in the form. The process works as follows: On form submission The workflow starts when someone submits a form with the required names or objects to create the joke. You can modify the form fields to make the jokes more creative. Set the output language Manually define the language in which you want to receive the joke. OpenAI Message Model Uses the OpenAI model to generate the joke based on the prompt and in the chosen language. (The response is limited to around 200 tokens.) Gmail: send message The generated joke is automatically sent to the specified email address via Gmail.
by Emmanuel Bernard
🎉 Do you want to master AI automation, so you can save time and build cool stuff? I’ve created a welcoming Skool community for non-technical yet resourceful learners. 👉🏻 Join the AI Atelier 👈🏻 Accepting payments via credit card online is a crucial component for the majority of businesses. Stripe provides a robust suite of tools for processing payments, yet many people still find it challenging to create a simple payment page and distribute it to their customers. 📋 Blog post 📺 Youtube Video This n8n workflow aims to offer the simplest and most direct method for generating a Stripe payment link. Features Quick Stripe Payment Link Creation:** Simply enter a title and select a price to create a Stripe payment link in seconds. Set Up Steps Connect your Stripe credentials. Fill the config node (currency). This n8n workflow template is crafted to significantly reduce the creation time of a Stripe Payment link. Created by the n8ninja.
by Mobder
This workflow automatically connects to a Cloudflare R2 bucket (via S3-compatible API), filters out files older than 14 days, deletes them, and then sends a Telegram notification for each deletion. It runs on a daily schedule. 🕘 Schedule Trigger Executes the workflow once a day at a specified hour (e.g., 9 AM). 📦 S3 Node – List Files Retrieves all objects from a specific folder (prefix) in a Cloudflare R2 bucket using the S3 API. 🔎 Code Node – Filter Files Older Than 2 Weeks Filters the retrieved files by comparing their LastModified timestamps to the current date. Files older than 14 days (2 weeks) are selected for deletion. 🗑️ S3 Node – Delete File Deletes each filtered file from the R2 bucket. 📨 Telegram Node – Notify Deletion Sends a Telegram message with the name of the deleted file to a specified chat ID. The message includes:
by Open Paws
This sub-workflow uses two custom Hugging Face regression models from Open Paws to evaluate and predict the real-world performance and advocacy alignment of text content. It’s designed to support animal advocacy organizations in optimizing their messaging across platforms like social media, email campaigns, and more. 🛠️ What It Does Sends input text to two deployed Hugging Face endpoints: Predicted Performance Model – Estimates real-world content success (e.g., engagement, shares, opens) based on patterns from real online data. Advocate Preference Model – Predicts how well the content will resonate with animal advocates (emotional impact, relevance, rationality, etc.) Outputs structured scores for both models Can be integrated into larger workflows for automated content review, filtering, or revision 📊 About the Models Text Performance Prediction Model** Trained on real-world data from 30+ animal advocacy organizations, this model predicts actual online performance of content—including social media, email marketing, and other outreach channels. Advocate Preference Prediction Model** Trained on ratings from animal advocates to evaluate how well a piece of text aligns with advocacy goals and values. Model Repositories: open-paws/text_performance_prediction_longform open-paws/animal_advocate_preference_prediction_longform > 📌 You must deploy each model as an inference endpoint on Hugging Face. Click "Deploy" on each model’s repo, then add the endpoint URL and your Hugging Face access token using n8n credentials. 📦 Use Cases Advocacy content review before publishing Automated scoring of outreach messages Filtering or flagging content with low predicted impact A/B testing support for message optimization
by Tom
n8n does not currently offer a way to retrieve emails from arbritrary providers via a regular node. Unless you're using Gmail or Outlook, you can only use the email trigger to start a workflow when a new email arrives. This currently limits the possible use cases you can cover in your n8n workflows, as you cannot (for example) get an idea of how many unread messages there are in an inbox, or search for specific messages when an event occurs. But fear not, there's a new sheriff in town! The JMAP standard allows you to interact with your mailboxes, calendars and contacts through single HTTP requests whenever needed. This n8n workflow demonstrates how to retrieve the total number of unread messages from a JMAP server and also retrieve details for the first 3 messages. It can easily be adapted to search for messages other than unread, or to return details for more than the first 3 messages. Screenshots FAQ Which n8n version do I need? The workflow was built using n8n 1.20 and should work here out of the box. HTTP requests are also supported on older n8n versions, so the workflow can be backported as an alternative. Which credentials do I need? The JMAP standard does not limit the available authentication options. Fastmail (the sponsor of the standard) supports Bearer authentication as well as OAuth2. In n8n you can implement the Fastmail Bearer authentication by creating Header Auth credentials with a name of Authorization and a value of Bearer $apiToken (replacing $apiToken with your actual API token from Fastmail). For other services you'd need to check the respective API documentation for more details on the support authentication methods. What even is JMAP? It's an official Internet Engineering Task Force (IETF) standard, sponsored by Fastmail, that will hopefully replace the legacy standards CalDAV, CardDAV, and IMAP soon. The full specs are available here. How can I use JMAP? If you're a Fastmail customer or if you're hosting your own Stalwart mail server you can use JMAP today. If your email provider doesn't yet support JMAP, you might want to contact them and let them know you're interested in this functionality.
by Belmont Digital
Description This n8n workflow verifies the deliverability of mailing addresses stored in Keap/Infusionsoft by integrating with Lob’s address verification service. Who is this for? This template is designed for Keap/Infusionsoft users who need to ensure the accuracy of mailing addresses stored in their CRM systems. What problem is this workflow solving? / Use Case This workflow addresses the challenge of maintaining accurate mailing addresses in CRM databases by verifying the deliverability of addresses. What this workflow does A new contact is created in Keap/Infusionsoft Webhook sent to n8n Verify if the address is deliverable via LOB Report back to Keap/Infusionsoft Set Up Steps Watch this setup video: https://www.youtube.com/watch?v=T7Baopubc-0 Takes 10-30 minutes to set up Accounts Needed: Keap/Infusionsoft LOB Account (https://www.lob.com $0.00/mo 300 US addresses Verifications) n8n Before using this template, ensure you have API keys for your Keap/Infusionsoft app and Lob. Set up authentication for both services within n8n. How to customize this workflow to your needs You can customize this workflow by adjusting the trigger settings to match Keap/Infusionsoft’s workflow configuration. Additionally, you can modify the actions taken based on the deliverability outcome, such as updating custom fields or sending notifications.
by Jonathan | NEX
Stop manually checking suspicious links. This free n8n workflow provides the foundation for a powerful, automated URL analysis pipeline. Using the NixGuard AI engine, you can instantly analyze suspicious URLs from emails, logs, or tickets to uncover phishing attempts, malware hosting sites, and malicious redirects. What You Will Automate: 🤖 Instant Threat Triage: Get an immediate AI-powered summary of why a URL is malicious, saving you critical investigation time. 🎯 Actionable IOC Extraction: Automatically extract the final redirected URL, malicious domains, and IPs to fuel your threat hunting and blocking rules. 🚀 SOAR-Ready Foundation: This workflow is the perfect starting point for your security playbooks. Use the output to: Alert: Send instant notifications to Slack or Teams. Respond: Create tickets in Jira or TheHive. Block: Add malicious domains to your firewall or DNS filter. Download this free template and automate your first line of defense against web-based threats in minutes! Don't have the main workflow yet? Get it HERE! 🔗 Learn more about NixGuard: thenex.world 🔗 Get started with a free security subscription: thenex.world/security/subscribe For search: URL Scanning, Phishing, Threat Intelligence, SOAR, SOC Automation, NixGuard, Free, AI, Incident Response, Cybersecurity, Automation, Link Analysis, MTTR, Malware, VirusTotal
by Belmont Digital
This n8n workflow verifies the deliverability of mailing addresses stored in HighLevel by integrating with Lob's address verification service. Who is this for? This template is designed for HighLevel users who need to ensure the accuracy of mailing addresses stored in their CRM systems. What problem is this workflow solving? / Use Case This workflow addresses the challenge of maintaining accurate mailing addresses in CRM databases by verifying the deliverability of addresses. What this workflow does A new contact is created in HighLevel Webhook sent to n8n Verify if the address is deliverable via LOB Report back to HighLevel Set Up Steps Watch this setup video: https://www.youtube.com/watch?v=T7Baopubc-0 Takes 10-30 minutes to set up Accounts Needed: HighLevel LOB Account (https://www.lob.com $0.00/mo 300 US addresses Verifications) n8n Before using this template, ensure you have API keys for your HighLevel app and Lob. Set up authentication for both services within n8n. How to customize this workflow to your needs You can customize this workflow by adjusting the trigger settings to match HighLevel's workflow configuration. Additionally, you can modify the actions taken based on the deliverability outcome, such as updating custom fields or sending notifications.
by LuisBetancourt.co
Description Whenever a Zoom “Meeting assets” email arrives in your Gmail inbox, this workflow will: 1) Trigger on new Gmail messages filtered by the subject “Meeting assets”. 2) Extract from the email (HTML or plain text): 3) Type of session (e.g. “1 hour”, “2 hours”, or “exploratory call”). Client’s full name. Session date & time (from the GMT… timestamp). Duration (HH:MM:SS). Recording link. Quick summary. Detailed summary. List of next steps. 4) Lookup the client in your Master Airtable base, table People, by full name. 5) Send a personalized Gmail to the client with all extracted details. 6) Create a new record in your Sessions table in Airtable, linking back to that client. Quick Start Import this JSON into n8n as a new workflow. Connect your Gmail credentials (OAuth2). Connect your Airtable credentials (Personal Access Token). In the Search Records node: Base → your Master base ID. Table → “Your people table”. Filter By Formula → ={Full Name} = '{{ $json.clientName }}'. In the Create Record node: Table → “Sessions”. Map each field (dateTime, duration, summaries, next steps, client link). Activate the workflow. Prerequisites n8n v1.50 or higher A Gmail account with OAuth2 credentials configured An Airtable base containing: Table People with a Full Name field (and email). Table Sessions with fields: DateTime, Duration, Quick Summary, Detailed Summary, Next Steps, and a Linked Record to People. An Airtable Personal Access Token with read/write access to that base. Tips & Extensions Timezone conversion: Use a Function node with moment-timezone to convert UTC if needed. Error handling: Add a catch node to log or notify if any field fails to parse. Alternate notifications: Swap the Gmail node for Slack, Microsoft Teams, or SMS integrations. With this documentation, your team can import and deploy the workflow in minutes. Enjoy!
by Niko
Capture URL Screenshots Automatically from Google Sheets & Drive with ScreenshotOne & Gmail Alerts Summary This automation template streamlines the process of capturing screenshots for multiple URLs. Instead of manually visiting each URL, taking a screenshot, and organizing the results, this workflow automates everything. When a spreadsheet is added to a designated Google Drive folder, the template extracts URLs from the column named "Url." These URLs are then processed through ScreenshotOne to capture screenshots, which are saved back to the same folder. Finally, an email notification is sent via Gmail with a link to the folder containing the screenshots. Problem Solved This template addresses the challenge of manual screenshot capture for multiple URLs. Without this automation, a user would need to: Open each URL from a spreadsheet. Take a screenshot manually. Save each screenshot with an appropriate name. Organize the screenshots in a folder. Notify stakeholders when the process is complete. These steps are not only time-consuming but also repetitive, especially when handling a large number of URLs. Who Can Benefit: Digital Marketers:** Monitor website appearances for competitive analysis or to track campaign landing pages. Web Developers/Designers:** Capture screenshots of multiple websites for inspiration or reference. QA Teams:** Document the visual state of web pages during various stages of development. SEO Specialists:** Track visual changes to websites they are optimizing. Content Managers:** Monitor how content appears across various web properties. Prerequisites Google Drive Node:** Must have appropriate permissions to create and access folders. Connected Google Sheets Node:** To extract URLs from the spreadsheet. Authenticated Gmail Node:** For sending notifications. ScreenshotOne Account:* Either a free or paid plan depending on volume needs, along with an *Access key**. Ensure you replace the placeholder --YOUR ACCESS KEY-- with your generated access key in the "Get Screenshots" node. Workflow Details Step 1: Google Drive Integration Trigger Node:** Monitors a specific folder in Google Drive. When a spreadsheet is added, the workflow is initiated. Step 2: Google Sheets Processing Google Sheets Node:** Extracts URLs from the column named "Url." Step 3: Screenshot Capture Get Screenshots Node:** Sends each extracted URL to ScreenshotOne to capture screenshots. Step 4: Saving Screenshots and Notifications Google Drive Node:** Saves the captured screenshots back into the same folder. Gmail Node:** Sends an email notification with a link to the folder, alerting stakeholders that the screenshots are ready. Customization Guidance Folder Monitoring: The workflow is set to monitor a specific Google Drive folder. It can be customized by selecting a different folder in the node settings. Spreadsheet Structure: While the template expects a spreadsheet with a column named "Url." for extracting URLs, users can add additional columns (e.g., titles, categories, or tags) and modify the workflow to utilize them as needed. Email Settings: Customize the recipient, subject, and body of the notification email to suit your needs. If required, enable optional notifications for different stakeholders. ScreenshotOne Access Key & Configurations: A valid ScreenshotOne Access key is required to capture screenshots. Users can further refine screenshot settings (e.g., viewport size, device emulation, or delay timing) by exploring the available options in the ScreenshotOne API documentation.
by Abbas Ali
This automation fetches the latest article from a WordPress blog, summarizes it using OpenAI, and sends the summary to a list of subscribers via email. Ideal for content creators and bloggers who want to distribute digestible content without manual effort. Use Case Perfect for: • Newsletter creators • Content marketers • Bloggers • Knowledge managers Nodes Used • Schedule Trigger • HTTP Request • Set • OpenAI • Google Sheets • Email (Gmail/SMTP) • IF • SplitInBatches Workflow Steps Trigger: Starts on a schedule (e.g., daily at 9:00 AM). Fetch Blog Post: Retrieves the most recent post from a WordPress blog via HTTP Request. Extract Fields: A Set node extracts the title, link, and content. Summarize Article: OpenAI processes the article and returns a 3-point summary. Fetch Subscribers: Google Sheets reads email addresses from a subscriber list. Loop Emails: SplitInBatches and Send Email nodes loop through subscribers. Conditional Logic: IF node skips articles shorter than 300 words. Credentials Required • OpenAI API Key (for content summarization) • Google Sheets OAuth2 (to read subscriber emails) • Gmail or SMTP (for sending emails) Test Instructions Replace blog URL in HTTP Request node. Connect OpenAI API key. Link your Google Sheet with a column named Email. Set up Gmail or SMTP credentials. Run manually for testing, then activate schedule.
by Oneclick AI Squad
This n8n workflow monitors email alerts for disk utilization exceeding 80%, extracts the server IP, logs into the server, and purges logs from Nginx, PM2, Docker, and system files to clear disk space. Key Insights Ensure email alerts are consistently formatted with server IP details. SSH access must be properly configured to avoid authentication failures. Workflow Process Initiate the workflow with the Check Disk Alert Emails node when an email triggers on high disk usage. Parse the email to extract the server IP using the Extract Server IP from Email node. Set up SSH credentials and paths manually with the Prepare SSH Variables node. Execute cleanup commands to delete logs from Nginx, PM2, Docker, and system files using the Run LogCleanup Commands via SSH node. Usage Guide Import the workflow into n8n and configure email and SSH credentials. Test with a sample email alert to verify IP extraction and log deletion. Prerequisites Email service (e.g., IMAP or API) for alert monitoring SSH access with valid credentials Customization Options Modify the Prepare SSH Variables node to target specific log directories or adjust cleanup commands for different server setups.