by Yaron Been
This workflow provides automated access to the Stability Ai Sdxl AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for image generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete image generation process using the Stability Ai Sdxl model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model by stability-ai for automated processing tasks. Key Capabilities High-quality image generation from text prompts** Advanced AI-powered visual content creation** Customizable image parameters and styles** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Stability Ai/sdxl AI model Stability Ai Sdxl**: The core AI model for image generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Image Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Content Creation**: Generate unique images for blogs, social media, and marketing materials Design Prototyping**: Create visual concepts and mockups for design projects Art & Creativity**: Produce artistic images for personal or commercial use Marketing Materials**: Generate eye-catching visuals for campaigns and advertisements Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #imagegeneration #aiart #texttoimage #visualcontent #aiimages #generativeart #sdxl #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by victor de coster
The template allows to make Dropcontact batch requests up to 250 requests every 10 minutes (1500/hour). Valuable if high volume email enrichment is expected. Dropcontact will look for email & basic email qualification if first_name, last_name, company_name is provided. +++++++++++++++++++++++++++++++++++++++++ Step 1: Node "Profiles Query" Connect your own source (Airtable, Google Sheets, Supabase,...) the template is using Postgres by default. Note I: Be careful your source is only returning a maximum of 250 items. Note II: The next node uses the next variables, make sure you can map these from your source file: first_name last_name website (company_name would work too) full_name (see note) Note III: This template is using the Dropcontact Batch API, which works in a POST & GET setup. Not a GET request only to retrieve data, as Dropcontact needs to process the batch data load properly. +++++++++++++++++++++++++++++++++++++++++ Step 2: Node "Data Transformation" Will transform the input variables in the proper json format. This json format is expected from the Dropcontact API to make a batch request. "full_name" is being used as a custom identifier to update the returned email to the proper contact in your source database. To make things easy, use a unique identiefer in the full_name variable. +++++++++++++++++++++++++++++++++++++++++ Step3: Node: "Bulk Dropcontact Requests". Enter your Dropcontact credentials in the node: Bulk Dropcontact Requests. +++++++++++++++++++++++++++++++++++++++++ Step4: Connect your output source by mapping the data you like to use. +++++++++++++++++++++++++++++++++++++++++ Step5: Node: "Slack" (OPTIONAL) Connect your slack account, if an error occur, you will be notified. TIP: Try to run the workflow with a batch of 10 (not 250) as it might need to run initially before you will be able to map the data to your final destination. Once the data fields are properly mapped, adjust back to 250.
by Akhil Varma Gadiraju
Workflow: HubSpot Contact Email Validation with Hunter.io Overall Goal This workflow retrieves contacts from HubSpot that have an email address but haven't yet had their email validated by Hunter. It then iterates through each of these contacts, uses Hunter.io to verify their email, updates the contact record in HubSpot with the validation status and date, and finally sends a summary email notification upon completion. How it Works (Step-by-Step Breakdown) Node: "When clicking ‘Test workflow’" (Manual Trigger) Type:** n8n-nodes-base.manualTrigger Purpose:** Start the workflow manually via the n8n interface. Output:** Triggers workflow execution. Node: "HubSpot" (HubSpot) Type:** n8n-nodes-base.hubspot Purpose:** Fetch contacts from HubSpot. Configuration:** Authentication: App Token Operation: Search for contacts Return All: True Filter Groups: Contact HAS_PROPERTY email Contact NOT_HAS_PROPERTY hunter_email_validation_status Output:** List of contact objects. Node: "Loop Over Items" (SplitInBatches) Type:** n8n-nodes-base.splitInBatches Purpose:** Process each contact one-by-one. Configuration:** Options > Reset: false Output:** Output 1 to "Hunter" Output 2 to "Send Email" Node: "Hunter" (Inside the loop) Type:** n8n-nodes-base.hunter Purpose:** Verify email with Hunter.io Configuration:** Operation: Email Verifier Email: {{ $json.properties.email }} Node: "Add Hunter Details (Contact)" (HTTP Request - Inside the loop) Type:** n8n-nodes-base.httpRequest Purpose:** Update HubSpot contact. Configuration:** Method: PATCH URL: https://api.hubapi.com/crm/v3/objects/contacts/{{ $('Loop Over Items').item.json.id }} Headers: Content-Type: application/json Body (JSON): { "properties": { "hunter_email_validation_status": "{{ $json.status }}", "hunter_verification_date": "{{ $now.format('yyyy-MM-dd') }}" } } Node: "Wait" (Inside the loop) Type:** n8n-nodes-base.wait Purpose:** Avoid API rate limits. Configuration:** Wait for 1 second. Node: "Replace Me" (NoOp - Inside the loop) Type:** n8n-nodes-base.noOp Purpose:** Junction node to complete the loop. Node: "Send Email" (After the loop completes) Type:** n8n-nodes-base.emailSend Purpose:** Send summary notification. Configuration:** From Email: test@gmail.com To Email: akhilgadiraju@gmail.com Subject: "Email Verification Completed for Your HubSpot Contacts" HTML: Formatted confirmation message Sticky Notes "HubSpot": Create custom properties (hunter_email_validation_status, hunter_verification_date). "Add Hunter Details": Ensure field names match HubSpot properties. "Wait": Prevent API rate limits. How to Customize It Trigger Replace Manual Trigger with Schedule Trigger (Cron) for automation. Optionally use HubSpot Trigger for new contact events. HubSpot Node Create matching custom properties. Adjust filters and returned properties as needed. Hunter Node Minimal customization needed. HTTP Request Node Update JSON property names if renaming in HubSpot. Customize date format as needed. Wait Node Adjust wait time to balance speed and API safety. Email Node Customize email addresses, subject, and body. Add dynamic contact count with a Set or Function node. Error Handling Add Error Trigger nodes. Use If nodes inside loop to act on certain statuses. Use Cases Clean your email list. Enrich CRM data. Prep verified lists for campaigns. Automate contact hygiene on a schedule. Required Credentials HubSpot App Token Used by: HubSpot node and HTTP Request node Create a Private App in HubSpot with required scopes. Hunter API Used by: Hunter node SMTP Used by: Email Send node Configure host, port, username, and password. Made with ❤️ using n8n by Akhil.
by Akhil Varma Gadiraju
AI-Powered GitHub Commit Reviewer Overview Workflow Name: AI-Powered GitHub Commit Reviewer Author: Akhil Purpose: This n8n workflow triggers on a GitHub push event, fetches commit diffs, formats them into HTML, runs an AI-powered code review using Groq LLM, and sends a detailed review via email. How It Works (Step-by-Step) 1. GitHub Trigger Node Type**: n8n-nodes-base.githubTrigger Purpose**: Initiates the workflow on GitHub push events. Repo**: akhilv77/relevance Output**: JSON with commit and repo details. 2. Parser Node Type**: n8n-nodes-base.set Purpose**: Extracts key info (repo ID, name, commit SHA, file changes). 3. HTTP Request Node Type**: n8n-nodes-base.httpRequest Purpose**: Fetches commit diff details using GitHub API. Auth**: GitHub OAuth2 API. 4. Code (HTML Formatter) Node Type**: n8n-nodes-base.code Purpose**: Formats commit info and diffs into styled HTML. Output**: HTML report of commit details. 5. Groq Chat Model Node Type**: @n8n/n8n-nodes-langchain.lmChatGroq Purpose**: Provides the AI model (llama-3.1-8b-instant). 6. Simple Memory Node Type**: @n8n/n8n-nodes-langchain.memoryBufferWindow Purpose**: Maintains memory context for AI agent. 7. AI Agent Node Type**: @n8n/n8n-nodes-langchain.agent Purpose**: Executes AI-based code review. Prompt**: Reviews for bugs, style, grammar, and security. Outputs styled HTML. 8. Output Parser Node Type**: n8n-nodes-base.code Purpose**: Combines commit HTML with AI review into one HTML block. 9. Gmail Node Type**: n8n-nodes-base.gmail Purpose**: Sends review report via email. Recipient**: akhilgadiraju@gmail.com 10. End Workflow Node Type**: n8n-nodes-base.noOp Purpose**: Marks the end. Customization Tips GitHub Trigger**: Change repo/owner or trigger events. HTTP Request**: Modify endpoint to get specific data. AI Agent**: Update the prompt to focus on different review aspects. Groq Model**: Swap for other supported LLMs if needed. Memory**: Use dynamic session key for per-commit reviews. Email**: Change recipient or email styling. Error Handling Use Error Trigger nodes to handle failures in: GitHub API requests LLM generation Email delivery Use Cases Instant AI-powered feedback on code pushes. Pre-human review suggestions. Security and standards enforcement. Developer onboarding assistance. Required Credentials | Credential | Used By | Notes | |-----------|---------|-------| | GitHub API (ID PSygiwMjdjFDImYb) | GitHub Trigger | PAT with repo and admin:repo_hook | | GitHub OAuth2 API | HTTP Request | OAuth2 token with repo scope | | Groq - Akhil (ID HJl5cdJzjhf727zW) | Groq Chat Model | API Key from GroqCloud | | Gmail OAuth2 - Akhil (ID wqFUFuFpF5eRAp4E) | Gmail | Gmail OAuth2 for sending email | Final Note Made with ❤️ using n8n by Akhil.
by LEDGERS
Before Using this Template first search LEDGERS in the nodes list and Update! The workflow automates contact creation in LEDGERS from a Google Sheets document, while performing validation checks on contact data to ensure data quality. Key Features: ✅ Google Sheets Trigger – Polls the sheet every minute to detect new or updated rows. ✅ Contact Name Validation – Checks if the contact name is provided; sends an email notification if missing. ✅ Email & Mobile Validation – Uses custom code to validate the format of email and mobile fields. ✅ Dynamic Routing – If both email and mobile are valid → proceeds to create the contact in LEDGERS. If either is invalid → sends an error notification via Gmail. ✅ Mobile Number Formatter – Extracts country code and mobile number for consistent formatting before creating the contact. ✅ LEDGERS API Integration – Creates the contact using the LEDGERS custom node. ✅ Error Handling – If contact creation fails (e.g., duplicate email), sends a failure email with details. ✅ Audit Trail – Appends results (including timestamps) back to Google Sheets for record-keeping. Extended Field Support This workflow is designed to be flexible. 👉 Along with core fields (Name, Email, Mobile), users can optionally map and include additional fields when creating a contact in LEDGERS, such as: Mobile Country Code GSTIN Business Name Billing Address 1 Billing Address 2 City State Country ➡ The workflow’s LEDGERS node can be configured to map these fields dynamically from Google Sheets if desired. Customization ✅ Users can extend or modify the workflow to handle these additional fields without needing structural changes — just by updating the mapping in the LEDGERS node parameters. Requirements: A connected Google Sheets document with at least Name column Configured Gmail credentials for sending email notifications A LEDGERS account with API access and credentials set up in n8n
by mail
Listen on new emails on a given IMAP mailbox. Converts the binary XML attachment to a stringified xmlproperty on the $data object and converts it to JSON. The Setnode could be replaced with Function nodes for example to do a conversion/mapping/transformation. The JSON is ultimately sent to a HTTP Endpoint via POST. This could be a starter for EDI purposes like receiving or transmitting XML data and conversion via JSON object property mapping.
by Harshil Agrawal
This workflow executes every minute and fetches the recent tweets from Twitter with the search query n8n_io. The workflow is built on the concept of polling. Cron node: The Cron node triggers the workflow every minute. Based on your use-case you can configure the time. You can even use the Interval node to trigger the workflow at a certain time interval. Twitter node: The Twitter node searches for the tweets that contain n8n_io and returns the most recent tweets. You can specify a different search query based on your use-case. Set node: The Set node sets the data that we pass on to the next nodes in the workflow. You can set only the values that you require in your workflow. Function node: All the magic happens in this node. The Twitter node returns all the recent tweets, including the ones which were returned earlier. The Function node, using the getWorkflowStaticData() method, only returns the tweets that are new, i.e., the tweets that were not returned in the previous workflow. Mattermost node: The Mattermost node sends the tweets from the Function node to the Twitter notifications channel. If you don't use Mattermost and want to share this data on a different platform, replace this node with the appropriate node.
by Muhammad Zeeshan Ahmad
Platform: n8n (Telegram Bot Integration) Purpose: Let users fetch top meme coin prices in real-time using a simple /memecoin Telegram command How It Works (Logic Breakdown) This flow listens for a Telegram command and fetches data from the CoinGecko API to respond with live memecoin prices. 🔹 1. Telegram Trigger Node Listens for incoming Telegram messages from users. Activated when a message is sent in a Telegram chat connected to the bot. Passes the raw message (e.g., /memecoin) to the next node. 🔹 2. IF Node – Check if Message is /memecoin Condition: {{$json"message"}} === "/memecoin" If true ➝ continue to fetch data from CoinGecko. If false ➝ nothing happens. 🔹 3. HTTP Request – Fetch Meme Coins from CoinGecko API: https://api.coingecko.com/api/v3/coins/markets?...category=meme-token Fetches top 5 meme tokens by market cap. Data includes: Name Symbol Current price (USD) Coin ID (for URL linking) 🔹 4. Function Node – Format the Message Parses the JSON response from CoinGecko. Builds a clean message like: ruby Copy Edit 🚀 Dogecoin (DOGE) 💰 Price: $0.123 🔗 More: https://www.coingecko.com/en/coins/dogecoin Loops through top 5 meme coins and adds line breaks. 🔹 5. Telegram Send Node – Reply to User Sends the formatted message to the original chat. Uses chat_id from the trigger to ensure correct user receives it. 🖼 Sample User Flow 👤 User types /memecoin in Telegram bot 🤖 Bot fetches meme coin prices 📬 Bot replies with live prices + links
by Michael Gullo
Automate Drafts From Google Drive This workflow automates the end-to-end process of extracting and summarizing information from PDFs stored in a specific Google Drive folder. When a new PDF or any binary data is added, the workflow is triggered and begins by downloading and processing the PDF to extract all available text. If multiple PDFs are detected, their content is aggregated into a single, combined dataset. This automation eliminates the time consuming task of manually reading, taking notes, and drafting documents. By removing this burden, users can focus on more meaningful tasks while the workflow handles the repetitive, tedious work. The extracted content is then passed through an AI-powered information extractor that identifies key details such as names, dates, addresses, and any other structured data points the user wants to extract from the PDF. This step is highly customizable, allowing the user to define exactly what type of information should be extracted. While the workflow is designed to extract all available content from the PDF, specifying additional structured data points ensures that critical details are accurately captured. A second OpenAI Node uses the extracted information to draft a professional, formal summary suitable for documentation. This is the most important part of the workflow and can be fully customized to meet the user's specific needs. By editing the prompts, users can tailor the workflow to generate a wide variety of draft formats based on the extracted content. The workflow then generates a new Google Document containing the full draft and composes an email summarizing the key points in 3 to 5 bullet points. This email is automatically sent to the designated recipient along with a direct link to the Google Doc. This solution is ideal for insurance, legal, or administrative use cases where timely, accurate extraction and reporting from incoming PDFs is essential. How To Use The Workflow Step 1 - Place any binary data (e.g., PDF files) into the designated Google Drive folder. Step 2 - The workflow will automatically download each PDF, extract the text, and if multiple PDFs are present combine them into a single dataset for analysis. Step 3 - The OpenAI Draft Agent will analyze the extracted information, generate a formal draft, and create a Google Document. This document will be updated with the draft content and saved back into the same Google Drive folder. Step 4 - An email will be sent to the designated recipient(s), including a summary of the draft and key extracted information, along with a link to view the Google Document. Need Help? Have Questions? For consulting and support, or if you have questions, please feel free to connect with me on LinkedIn or email michael.gullo@outlook.com.
by Oneclick AI Squad
This AI-powered workflow reads emails, understands the request using an LLM, and creates structured Jira issues. Key Insights Poll for new emails every 5 minutes; ensure Gmail/IMAP is properly configured. AI analysis requires a reliable LLM model (e.g., Chat Model or AI Tool). Workflow Process Trigger the workflow with the Check for New Emails Gmail Trigger node. Fetch full email content using the Fetch Full Email Content get message node. Analyze email content with the Analyze Email & Extract Tasks node using AI. Parse the AI-generated JSON output into tasks with the Parse JSON Output from AI node. Create the main Jira issue with the Jira - Create Main Issue create: issue node. Split subtasks from JSON and create them with the Split Subtasks JSON Items and Create Subtasks create: issue nodes. Usage Guide Import the workflow into n8n and configure Gmail and Jira credentials. Test with a sample email to ensure ticket creation and subtask assignment. Prerequisites Gmail/IMAP credentials for email polling Jira API credentials with issue creation permissions Customization Options Adjust the Analyze Email & Extract Tasks node to refine AI task extraction or modify the polling frequency in the trigger node.
by Oneclick AI Squad
This n8n workflow transforms simple chat requests into professional Center of Excellence (COE) blog posts using AI, then automatically publishes them to Google Drive. What Is This? An intelligent blog creation system that takes a topic from chat and produces executive-level blog posts. Uses three AI stages to ensure high-quality, professional content suitable for business consumption. Why Use This? Professional Content**: Creates sophisticated blogs with business insights Save Time**: Topic to published blog in 2-3 minutes No Writing Skills Needed**: AI handles all the writing and formatting Auto-Publishing**: Creates and shares Google Docs automatically Easy to Use**: Just chat your topic, get your blog How It Works 1. Blog Request & Planning Start Blog Request**: Chat interface receives your blog topic Create Blog Outline**: AI generates structured outline with sections AI Brain for Outline**: Powers the initial content planning 2. Content Review & Writing Review & Fix Outline**: AI improves outline structure and flow Write Full Blog Post**: Transforms outline into complete professional blog AI Brain for Review/Writing**: Advanced AI models handle content creation 3. Publish & Share Blog Clean Up Text Format**: Removes formatting issues for clean presentation Save Blog to Google Drive**: Creates properly formatted Google Doc Email Blog to Stakeholder**: Shares with specified team members Make Blog Public**: Creates shareable public link Send Blog Link to User**: Returns document URL via chat How to Use Start Chat: Access the chat interface and type your blog topic Wait: AI automatically processes (2-3 minutes) Get Your Blog: Receive Google Drive link to your published blog Good to Know Processing Time**: 2-3 minutes per blog Content Quality**: Uses advanced Gemini AI models for professional output Auto-Formatting**: Creates clean Google Docs ready for sharing Instant Sharing**: Stakeholders get email notifications automatically Public Access**: Generates shareable links for broader distribution Customizing This Workflow Content Style Modify AI prompts to match your company's writing tone Adjust content evaluation criteria for different audiences Change blog structure templates Publishing & Sharing Update stakeholder email addresses Change Google Drive folder destinations Modify sharing permissions (public/private) Add more distribution channels AI Enhancement Switch between different AI models for speed vs quality Add more review stages for specialized content Include company-specific knowledge sources
by Jan Oberhauser
Triggers every day at 1pm Gets the current content from Hacker News Gets all the different submission items Extracts the rank, title and url Checks if it is a "Show HN" submission Combines the items into a simple email text Sends an email with the email text