by Akram Kadri
Who is this for? This workflow template is ideal for marketers, designers, content creators, and developers who need to generate custom text-based images dynamically. Whether you want to create social media graphics, placeholder images, or text-based LinkedIn carousels, this workflow provides a simple, no-code solution using an API that requires no authentication. What problem does this workflow solve? Creating text-based images often requires design software or complex integrations with graphic tools. This workflow eliminates that hassle by allowing users to generate images with custom text, font styles, colors, and background colors using a simple HTTP request. It’s perfect for automating image generation without relying on external tools or manual effort. What this workflow does This workflow leverages an HTTP request to a free API that generates text-based images dynamically. Here's what it enables you to do: Define custom image text Set image dimensions (width x height) Choose a background color and text color using hex codes Select a font family and font size Specify the image format (PNG, JPG, or WebP) The generated image can be used immediately, making it ideal for automating content creation workflows. Setup Open the workflow in n8n. Modify the Set node to define your preferred image properties: text: The message displayed on the image. size: Image dimensions (e.g., 500x300 pixels). backgroundColor: Hex color code for the background. textColor: Hex color code for the text. fontFamily: Select from available font options (e.g., Pacifico, Ubuntu). fontSize: Define the text size. type: Choose the image format (PNG, JPG, or WebP). Execute the workflow to generate an image. The HTTP request returns the generated image, ready for use. How to customize this workflow 1. Adjust the Set node values to match your desired design. 2. Use dynamic data for text, allowing personalized images based on user input. 3. Automate image delivery by adding email or social media posting nodes. 4. Integrate this workflow into larger automation sequences, such as content marketing pipelines.
by David Ashby
Complete MCP server exposing all Mandrill Tool operations to AI agents. Zero configuration needed - all 2 operations pre-built. ⚡ Quick Setup Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community Import this workflow into your n8n instance Activate the workflow to start your MCP server Copy the webhook URL from the MCP trigger node Connect AI agents using the MCP URL 🔧 How it Works • MCP Trigger: Serves as your server endpoint for AI agent requests • Tool Nodes: Pre-configured for every Mandrill Tool operation • AI Expressions: Automatically populate parameters via $fromAI() placeholders • Native Integration: Uses official n8n Mandrill Tool tool with full error handling 📋 Available Operations (2 total) Every possible Mandrill Tool operation is included: 💬 Message (2 operations) • Send a message based on a template • Send a message based on HTML 🤖 AI Integration Parameter Handling: AI agents automatically provide values for: • Resource IDs and identifiers • Search queries and filters • Content and data payloads • Configuration options Response Format: Native Mandrill Tool API responses with full data structure Error Handling: Built-in n8n error management and retry logic 💡 Usage Examples Connect this MCP server to any AI agent or workflow: • Claude Desktop: Add MCP server URL to configuration • Custom AI Apps: Use MCP URL as tool endpoint • Other n8n Workflows: Call MCP tools from any workflow • API Integration: Direct HTTP calls to MCP endpoints ✨ Benefits • Complete Coverage: Every Mandrill Tool operation available • Zero Setup: No parameter mapping or configuration needed • AI-Ready: Built-in $fromAI() expressions for all parameters • Production Ready: Native n8n error handling and logging • Extensible: Easily modify or add custom logic > 🆓 Free for community use! Ready to deploy in under 2 minutes.
by Mark Shcherbakov
Video Guide I prepared a detailed guide that showed the whole process of integrating the Binance API and storing data in Airtable to manage funding statements associated with tokens in a wallet. Youtube Link Who is this for? This workflow is ideal for developers, financial analysts, and cryptocurrency enthusiasts who want to automate the process of managing funding statements and token prices. It’s particularly useful for those who need a systematic approach to track and report funding fees associated with tokens in their wallets. What problem does this workflow solve? Managing funding statements and token prices across multiple platforms can be cumbersome and error-prone. This workflow automates the process, allowing users to seamlessly fetch funding fees from Binance and record them alongside token prices in Airtable, minimizing manual data entry and potential discrepancies. What this workflow does This workflow integrates the Binance API with an Airtable database, facilitating the storage and management of funding statements linked to tokens in a wallet. The agent can: Fetch funding fees and current positions from Binance. Aggregate data to create structured funding statements. Insert records into Airtable, ensuring proper linkage between funding data and tokens. API Authentication: The workflow establishes authentication with the Binance API using a Crypto Node to handle API keys and signatures, ensuring secure and verified requests. Data Collection: It retrieves necessary data, including funding fees and current positions with properly formatted API requests to ensure seamless communication with Binance. Airtable Integration: The workflow inserts aggregated funding statements and token data into the corresponding Airtable records, managing token existence checks to avoid duplicate entries. Setup Set Up Airtable Database: Create an Airtable base with tables for Funding Statements and Tokens. Generate Binance API Key: Log in and create an API key with appropriate permissions. Set Up Authentication in N8N: Utilize a Crypto Node for Binance API authentication. Configure API Request to Binance: Set request method and headers for communication with the Binance API. Fetch Funding Fees and Current Positions: Retrieve funding data and current positions efficiently. Aggregate and Create Statements: Aggregate data to create detailed funding statements. Insert Data into Airtable: Input the structured data into Airtable and manage token records. Using Get Price Node: Implement a Get Price Node to maintain current token price tracking without additional setup.
by PUQcloud
Overview The Docker NextCloud WHMCS module leverages a sophisticated workflow for n8n, designed to automate the comprehensive deployment, configuration, and management processes for NextCloud and NextCloud Office services. Through its intuitive API interface, the workflow securely receives commands and orchestrates predefined tasks via SSH on your Docker-hosted server, ensuring streamlined operations and efficient management. Prerequisites You must deploy your own dedicated n8n server to manage workflows effectively. Alternatively, you may opt for the official n8n cloud-based solutions accessible via: n8n Official Site Your Docker server must be accessible via SSH with necessary permissions. Installation Steps Install the Required Workflow on n8n You can select from two convenient installation options: Option 1: Use the Latest Version from the n8n Marketplace The latest workflow templates are continuously updated and available on the n8n marketplace. Explore all templates provided by PUQcloud directly here: PUQcloud on n8n Option 2: Manual Installation Each module version includes a bundled workflow template file. Import this workflow file directly into your n8n server manually. n8n Workflow API Backend Setup for WHMCS Configure API Webhook and SSH Access Create a secure Basic Auth Credential for Webhook API interactions within n8n. Create an SSH Credential within n8n to securely communicate with the Docker host. Modify Template Parameters Adjust and update the following critical parameters to match your deployment specifics: server_domain – Set this to the domain of your WHMCS Docker server. clients_dir – Specify the directory where user data and related resources will be stored. mount_dir – The standard mount point for container storage (recommended to remain unchanged). Do not alter the following technical parameters to avoid workflow disruption: screen_left, screen_right. Deploy-docker-compose Configuration Fine-tune Docker Compose configurations tailored specifically for these critical operational scenarios: Initial service provisioning and setup Service suspension and subsequent unlocking Service configuration updates Routine service maintenance tasks nginx Configuration Management Enhance and customize proxy server configurations using the dedicated nginx workflow element: main**: Define specialized parameters within the server configuration block. main_location**: Set custom headers, caching policies, and routing rules for the root location. Bash Script Automation Automate Docker container management and related server tasks through dynamically generated Bash scripts within n8n. Scripts execute securely via SSH and provide responses in JSON or plain text formats for easy parsing and logging. Scripts are conveniently linked directly to the SSH action elements. You retain complete flexibility to adapt or extend these scripts as necessary to meet your precise operational requirements.
by Lucas Peyrin
How it works This workflow automates the process of checking for and applying updates to a self-hosted n8n instance running on Docker. It runs on a schedule, checks for new versions, summarizes the release notes with AI, and asks for your approval via Telegram before updating. Scheduled Check: The workflow runs hourly, triggered by a Schedule node. Version Discovery: It first confirms it's running in a Docker environment. It uses SSH to connect to the host machine and inspects the running n8n container to find its current version tag (e.g., latest or next). It then queries the Docker Hub API to compare the image digest (a unique ID for an image version) of the running version against the latest available version for that tag. Update Detection: If the digests do not match, it means a new image has been pushed for your version tag (e.g., a new latest image is available), and an update is needed. AI-Powered Release Notes: It fetches the official release notes for the new version from the GitHub API. An AI model (LLM) summarizes these technical notes into a concise, human-readable overview of the key features and fixes. Manual Approval: It sends a message to a Telegram chat with the AI-generated summary and two buttons: "✅ Update" and "❌ Ignore". The workflow then pauses and waits for your response. Execute Update: If you approve the update, the workflow uses SSH to run a docker compose command on your server, which pulls the new image, stops the old containers, and starts the new ones. Set up steps Setup time: ~5-10 minutes SSH Credentials: Go to Credentials and create a new SSH credential with the username, host, and password/private key for the server where your n8n Docker instance is running. Select this credential in the Get n8n Current Version and Update Docker nodes. Telegram Bot Credentials: Create a Telegram Bot and get its API token. Go to Credentials and create a new Telegram credential with your bot's token. Select this credential in the Send a text message node. AI Model Credentials: Ensure you have credentials for an AI provider (like Google AI, OpenAI, etc.) set up. Select your desired credential in the Google Gemini Chat Model node (or replace it with your preferred LLM node). Configure Paths and Commands: Open the Docker Path node. Set the docker_path to the absolute path of your docker-compose.yml file on the server (e.g., /root/n8n). If you use workers, adjust the worker_command to include the correct --scale argument for your setup. If not, you can leave it blank. Set Your Chat ID: Open the Approve Update Telegram node and enter your personal Telegram Chat ID in the Chat ID field. This ensures the approval message is sent to you. Activate the workflow. It will now check for updates every hour. To enable fully automatic updates (without manual approval): Delete the nodes from Get n8n Releases to Approved ? and connect the Needs Update ? node directly to the Update Docker node.
by Reza Gholizade
This n8n workflow template uses community nodes and is only compatible with the self-hosted version of n8n. Conversational Kubernetes Management with GPT-4o and MCP Integration This workflow enables you to manage Kubernetes clusters conversationally using OpenAI’s GPT-4o and a secure MCP (Model Context Protocol) server. It transforms natural language queries into actionable Kubernetes commands via a lightweight MCP API gateway — perfect for developers and platform engineers seeking to simplify cluster interaction. 🚀 Setup Instructions Import the Workflow Upload this template to your n8n instance. Configure Required Credentials OpenAI API Key: Add your GPT-4o API key in the credentials. MCP Client Node: Set the URL and auth for your MCP server. Test Kubernetes Access Ensure your MCP server is correctly configured and has access to the target Kubernetes cluster. 🧩 Prerequisites n8n version 0.240.0 or later Access to GPT-4o via OpenAI A running MCP server Kubernetes cluster credentials configured in your MCP backend ⚠️ Community Nodes Disclaimer This workflow uses custom community nodes (e.g., MCP Client). Make sure to review and trust these nodes before running in production. 🛠️ How It Works A webhook or chat input triggers the conversation. GPT-4o interprets the message and generates structured Kubernetes queries. MCP Client securely sends requests to your cluster. The result is returned and formatted for easy reading. 🔧 Customization Tips Tweak the GPT-4o prompt to match your tone or technical level. Extend MCP endpoints to support new Kubernetes actions. Add alerting or monitoring integrations (e.g., Slack, Prometheus). 🖼️ Template Screenshot 🧠 Example Prompts Show me all pods in the default namespace. Get logs for nginx pod in kube-system. List all deployments in staging. 📎 Additional Resources MCP Server on GitHub OpenAI Documentation n8n Docs Build smarter Kubernetes workflows with the power of AI !
by Juan Carlos Cavero Gracia
Description This n8n automation template provides an end-to-end solution for generating a series of themed images for Instagram and TikTok carousels using OpenAI's GPT Image (via the image generation API) and automatically publishing them to both platforms. It uses a sequence of prompts to create a narrative or themed carousel, generating each image based on the previous one, and then posts them with an AI-generated caption. Who Is This For? Social Media Managers:** Quickly create and schedule engaging image carousels for Instagram and TikTok. Content Creators:** Automate the visual content creation process for thematic posts or storytelling carousels. Digital Marketers:** Efficiently produce visual assets for campaigns that require sequential imagery. Small Businesses:** Generate unique promotional content for social media without needing advanced design skills. What Problem Does This Workflow Solve? Manually creating a series of related images for a carousel and then publishing them across multiple platforms can be repetitive and time-consuming. This workflow addresses these issues by: Automating Image Generation:** Uses OpenAI to generate a sequence of 5 images, where each new image is an evolution based on the previous one and a new prompt. Automating Caption Generation:** Leverages OpenAI (GPT) to create a suitable description/caption for the carousel based on the image prompts. Streamlining Multi-Platform Publishing:** Automatically uploads the generated image carousel and caption to both Instagram and TikTok. Reducing Manual Effort:** Significantly cuts down the time spent on designing individual images and manually uploading them. Ensuring Visual Cohesion:** The sequential image generation method (editing the previous image) helps maintain a consistent style or narrative across the carousel. How It Works Trigger: The workflow is initiated manually (can be adapted to a schedule or webhook). Define Prompts: Five distinct prompts are pre-set within the workflow to guide the generation of each image in the carousel. AI Caption Generation: OpenAI (GPT-4.1) generates a concise (≤ 90 characters for TikTok) description for the social media posts based on all five image prompts. Sequential AI Image Generation: Image 1: OpenAI's image generation API (specified as gpt-image-1) creates the first image based on prompt1. Image 2-5: For each subsequent image, the workflow uses the OpenAI image edits API. It takes the previously generated image and a new prompt (prompt2 for image 2, prompt3 for image 3, and so on) to create the next image in the sequence. Images are converted from base64 JSON to binary format. Content Aggregation: The five generated binary image files (named photo1 through photo5) are merged. Multi-Platform Distribution: The merged images and the AI-generated description are sent to api.upload-post.com for publishing as a carousel to Instagram. The same content is sent to api.upload-post.com for publishing as a carousel to TikTok, with an option to automatically add music. The TikTok description is truncated if it exceeds 90 characters. Setup Accounts & API Keys: You will need: An n8n instance. An OpenAI API key. An API key for upload-post.com. Configure Credentials: Add your OpenAI API key to the "OpenAI" credentials in n8n. This will be used by the "Generate Description for Tiktok and Instagram" node and the HTTP Request nodes calling the OpenAI image generation/edit APIs. In the "POST TO INSTAGRAM" and "POST TO TIKTOK" nodes, replace "Apikey add_api_key_here" with your actual upload-post.com API key. Update the user field in the "POST TO INSTAGRAM" and "POST TO TIKTOK" nodes if "upload_post" is not your user identifier for that service. Customize Prompts: Modify the five prompts (prompt1 to prompt5) in the "Set All Prompts" node to define the story or theme of your image carousel. Review Image Generation Parameters: In the "Set API Variables" node, you can adjust: size_of_image (e.g., "1024x1536" for vertical carousels). openai_image_model (ensure this matches a valid OpenAI model identifier for image generation/edits, like dall-e-2 or dall-e-3 if gpt-image-1 is a placeholder). response_format_image (should generally remain b64_json for this workflow). (Optional) TikTok Auto Music: The "POST TO TIKTOK" node has an auto_add_music parameter set to true. Change this to false if you prefer to add music manually or not at all. Requirements Accounts:** n8n, OpenAI, upload-post.com. API Keys & Credentials:** API Keys for OpenAI and https://upload-post.com. (Potentially) Paid Plans:** OpenAI and upload-post.com usage may incur costs depending on your volume and their respective pricing models. This template empowers you to automate the creation and distribution of visually consistent image carousels, saving time and enhancing your social media presence.
by PUQcloud
Overview The Docker InfluxDB WHMCS module uses a specially designed workflow for n8n to automate deployment processes. The workflow provides an API interface for the module, receives specific commands, and connects via SSH to a server with Docker installed to perform predefined actions. Prerequisites You must have your own n8n server. Alternatively, you can use the official n8n cloud installations available at: n8n Official Site Installation Steps Install the Required Workflow on n8n You have two options: Option 1: Use the Latest Version from the n8n Marketplace The latest workflow templates for our modules are available on the official n8n marketplace. Visit our profile: PUQcloud on n8n Option 2: Manual Installation Each module version comes with a workflow template file. You need to manually import this template into your n8n server. n8n Workflow API Backend Setup for WHMCS/WISECP 1. Configure API Webhook and SSH Access Create a Basic Auth Credential for the Webhook API block in n8n. Create an SSH Credential for accessing a server with Docker installed. 2. Modify Template Parameters In the Parameters block of the template, update the following settings: server_domain – Must match the domain of the WHMCS/WISECP Docker server. clients_dir – Directory where user data related to Docker and disks will be stored. mount_dir – Default mount point for the container disk (recommended not to change). Do not modify the following technical parameters: screen_left screen_right Deploy-docker-compose In the Deploy-docker-compose element, you can modify the Docker Compose configuration. This is generated in the following scenarios: When the service is created When the service is unlocked When the service is updated nginx In the nginx element, you can modify configuration parameters of the web interface proxy server. The main section allows you to add custom parameters to the server block in the proxy server configuration file. The main_location section contains settings that will be added to the location / block of the configuration. Here, you can define custom headers and parameters. Bash Scripts Management of Docker containers and related procedures is done by executing Bash scripts generated in n8n. These scripts return either JSON or plain strings. All scripts are located in elements directly connected to the SSH element. You have full control over any script and can modify or execute it as needed.
by Davide
Voiceflow is a no-code platform that allows you to design, prototype, and deploy conversational assistants across multiple channels—such as chat, voice, and phone—with advanced logic and natural language understanding. It supports integration with APIs, webhooks, and even tools like Twilio for phone agents. It's perfect for building customer support agents, voice bots, or intelligent assistants. This workflow connects n8n and Voiceflow with tools like Google Calendar, Qdrant (vector database), OpenAI, and an order tracking API to power a smart, multi-channel conversational agent. There are 3 main webhook endpoints in n8n that Voiceflow interacts with: n8n_order – receives user input related to order tracking, queries an API, and responds with tracking status. n8n_appointment – processes appointment booking, reformats date input using OpenAI, and creates a Google Calendar event. n8n_rag – handles general product/service questions using a RAG (Retrieval-Augmented Generation) system backed by: Google Drive document ingestion, Qdrant vector store for search, and OpenAI models for context-based answers. Each webhook is connected to a corresponding "Capture" block inside Voiceflow, which sends data to n8n and waits for the response. How It Works This n8n workflow integrates Voiceflow for chatbot/voice interactions, Google Calendar for appointment scheduling, and RAG (Retrieval-Augmented Generation) for knowledge-based responses. Here’s the flow: Trigger**: Three webhooks (n8n_order, n8n_appointment, n8n_rag) receive inputs from Voiceflow (chat, voice, or phone calls). Each webhook routes requests to specific functions: Order Tracking: Fetches order status via an external API. Appointment Scheduling: Uses OpenAI to parse dates, creates Google Calendar events, and confirms via WhatsApp. RAG System: Queries a Qdrant vector store (populated with Google Drive documents) to answer customer questions using GPT-4. AI Processing**: OpenAI Chains: Convert natural language dates to Google Calendar formats and generate responses. RAG Pipeline: Embeds documents (via OpenAI), stores them in Qdrant, and retrieves context-aware answers. Voiceflow Integration: Routes responses back to Voiceflow for multi-channel delivery (chat, voice, or phone). Outputs**: Confirmation messages (e.g., "Event created successfully"). Dynamic responses for orders, appointments, or product support. Setup Steps Prerequisites: APIs**: Google Calendar & Drive OAuth credentials. Qdrant vector database (hosted or cloud). OpenAI API key (for GPT-4 and embeddings). Configuration: Qdrant Setup: Run the "Create collection" and "Refresh collection" nodes to initialize the vector store. Populate it with documents using the Google Drive → Qdrant pipeline (embeddings generated via OpenAI). Voiceflow Webhooks: Link Voiceflow’s "Captures" to n8n’s webhook URLs (n8n_order, n8n_appointment, n8n_rag). Google Calendar: Authenticate the Google Calendar node and set event templates (e.g., summary, description). RAG System: Configure the Qdrant vector store and OpenAI embeddings nodes. Adjust the Retrieve Agent’s system prompt for domain-specific queries (e.g., electronics store support). Optional: Add Twilio for phone-agent capabilities. Customize OpenAI prompts for tone/accuracy. PS. You can import a Twilio number to assign it to your agent for becoming a Phone Agent Need help customizing? Contact me for consulting and support or add me on Linkedin
by Dina Lev
Automate Legal Document Generation with n8n, Apify, Google Drive, and AI This tutorial details an end-to-end automation solution for streamlining the lien filing process for Homeowners Associations (HOAs) using an n8n workflow. It significantly reduces manual effort and potential errors for legal professionals by automating document retrieval, information extraction, and document generation. Who's it for This template is ideal for legal professionals, law firms, and property management companies that frequently handle lien filings for Homeowners Associations. If you're looking to reduce manual document processing time, minimize errors, and improve efficiency in your legal operations, this workflow is for you. The Problem Legal professionals often allocate a significant portion of their time—up to 40%—to manual document processing tasks. The traditional process for filing a lien is particularly time-consuming (e.g., 15 minutes per case) and error-prone, involving steps like manual searching, downloading, extracting, and populating legal documents. The Automation Solution Overview This automation leverages an n8n workflow in conjunction with external services like Playwright (via Apify), Google Drive, Google Sheets, Gmail, and the Gemini API. The primary objective is to automate the legal document generation process—from initial data submission to final document generation and notification. Requirements Before importing and running the n8n workflow, you need the following: n8n Instance:** A running n8n instance (self-hosted or cloud). Google Account:** With access to Google Sheets, Google Drive, and Gmail. Google Sheets:** An Input Sheet to receive form responses (e.g., "Legal Automation Input Form (Responses)"). An Output/Review Sheet for extracted data and approval (e.g., "Automation Output data Sheet") with specific columns like "Timestamp", "Legal Description", "Association Name", "Debt", "Parcel", "Owner", "Doc link", "Approval", and "Created". Google Drive:** A main folder for n8n outputs (e.g., "N8N Folder"). A Google Docs Lien Template with placeholders (e.g., {{ASSOCIATION}}, {{DEBT}}, {{PROPERTY}}, {{MONTH}}, {{YEAR}}, {{DAY}}, {{PARCEL}}, {{OWNER}}). Google Gemini API Key:** For text and image processing. Apify Account & Playwright Actor:** An Apify account with access to a Playwright actor capable of scraping property information from your target county's website. Setup Steps n8n Credentials: Add Google Sheets, Google Drive, and Gmail credentials in your n8n instance. Add an HTTP Query Auth credential for your Gemini API key (named "Query Auth account" in the template). Ensure your Apify API token is configured within the Apify Playwright script to find property info node. Google Sheets Configuration: Link the Google Sheets Trigger node to your Input Sheet. Link the Google Sheets node (for appending data) and the Intermediate data received trigger to your Output/Review Sheet. Google Drive Configuration: Update the Create folder to output node with the ID of your "N8N Folder". Update the Make Copy of Template node with the ID of your Google Docs Lien Template. Email Addresses: Update the recipient email addresses in the Approve Through Email and Notify complete nodes to your desired notification email. Detailed Tutorial Steps and n8n Workflow Breakdown Summary This n8n workflow, "Legal Document Generator E2E", automates the process of generating legal lien documents, from initial data input to final document creation and notification. Initiate Workflow: The workflow starts with a Google Sheets Trigger node, which listens for new lien requests submitted via a form that populates a Google Sheet. Gather Property Data: An Apify Playwright script to find property info node fetches property details from county websites, and a Get file for property node downloads associated legal documents. Process and Store Document: The downloaded document is transformed to base64 using Transform to base64 and then uploaded to Google Drive via Upload legal doc for storage and further processing. Extract Information with AI: Call Gemini API for legal desc and Property metadata nodes leverage the Gemini API to extract the precise legal description, parcel number, and owner's name from the document. This extracted data is then structured by the Property Information Extractor. Review and Approve: The extracted information is appended to an intermediate Google Sheet by the first Google Sheets node, and an email is sent via Approve Through Email to the user for review and approval. Generate Documents on Approval: A second Intermediate data received Google Sheets Trigger node monitors the approval status in the sheet. Once "Approved", an If node allows the workflow to proceed. Create and Populate Documents: A new client-specific folder is created in Google Drive using Create folder to output. A blank lien template is copied (Make Copy of Template), and its custom variables are populated with the extracted data using Change Custom Variables. Finalize and Store Output: The populated document is converted to PDF (Generate PDF), and both the new PDF (Add PDF To Drive) and the original source document (Move file in Google Drive) are saved to the client's new folder. Update Records and Notify: The Update Creation Google Sheets node marks the document as "Created" in the tracking sheet and updates the document link. Finally, Notify complete sends a notification email about the completion. How to Customize the Workflow Adjust Input Form Fields:** Modify the column names in your initial Google Sheet and update the expressions in the Google Sheets Trigger and Apify Playwright script to find property info nodes to match your form. Change County Website/Scraper:** If you need to fetch data from a different county or property database, you will need to modify the Apify Playwright script to find property info node to call a different Apify actor or configure a new HTTP Request node to interact with your chosen data source. Customize Document Template:** Update the placeholders in your Google Docs Lien Template to match your specific document needs. Ensure corresponding replaceAll actions are updated in the Change Custom Variables node. Modify AI Prompts:** Refine the prompts within the Call Gemini API for legal desc and Property metadata nodes to improve the accuracy of information extraction based on your document types. Notification Preferences:** Adjust the sendTo email addresses and subject/message content in the Approve Through Email and Notify complete nodes. Benefits of this Automation This automation offers significant advantages for legal professionals: Streamlined Organization:** Ensures all relevant documents—original source files, editable templates, and final PDFs—are systematically organized, tracked, and easily accessible within Google Drive. Time-Saving and Efficiency:** Documents are quickly generated and ready for client sharing, leading to faster turnaround times and improved service delivery. Scalability:** Provides a scalable solution for handling a higher volume of document processing tasks without a proportional increase in manual effort. Learn more about Chill Labs and our services on our website: Chill Labs
by PollupAI
This n8n workflow streamlines the onboarding process for new customers by automating personalized email communication, calendar scheduling, and contact assignment in HubSpot. It is perfect for businesses looking to ensure a smooth and personalized onboarding experience for new clients. 🧑💼 Who is this for? Customer success teams who need to onboard new clients efficiently. Sales teams who want to ensure smooth transitions from prospect to customer. Small businesses that want to automate customer onboarding without complex systems. 🧩 What problem is this workflow solving? This workflow reduces the manual effort involved in onboarding new customers by: Automatically sending personalized welcome emails. Scheduling a welcome meeting using a calendar tool. Assigning the customer to a Customer Success Manager (CSM) in HubSpot. ⚙️ What this workflow does Trigger via Webhook or HubSpot: The workflow can be triggered either by a webhook (direct API call) or a HubSpot trigger (e.g., when a new contact is created). HubSpot Connection: Retrieves the list of HubSpot owners (users with contact access). Identifies the owner of the new contact. Calendar Management: Utilizes a Calendar Agent to schedule a welcome meeting with the new customer. The Calendar Agent can create, update, or delete events as needed. Personalized Email Creation: Uses an AI-powered Email Writer (OpenAI) to generate a personalized welcome email. Transforms the email text into HTML for a polished format. Email Sending via Gmail: Sends the personalized email to the customer using Gmail. Sets the new contact’s owner in HubSpot for further communication tracking. 🛠️ Setup Webhook Setup in n8n: Create a new workflow and add a Webhook node. Set the Webhook URL path (e.g., /webhook-customer-onboarding). Make sure the workflow is active. Webhook Setup in HubSpot: Go to HubSpot Developer Account. Navigate to Settings > Integrations > Webhooks. Create a new webhook and set the URL as the n8n Webhook URL. Choose POST as the request method. Test the webhook to ensure it triggers the workflow in n8n. Calendar Agent Configuration: The Calendar Agent can be configured to create, update, or delete events. Connect it to your calendar tool (Google Calendar, Outlook, etc.). Customize the calendar event details (title, description, time). Email Writer Setup: Customize the AI prompt in the Email Writer node to match your brand’s voice. Adjust the email text format for your specific needs. Gmail Integration: Connect your Gmail account in n8n. Set the recipient email to the new customer’s email address. ✏️ How to customize this workflow to your needs Modify the AI-Powered Email**: Adjust the email prompt for the AI model to create a different welcome message. Change the email format or add custom variables (e.g., customer name, service details). Customize Calendar Settings**: Set default time slots for welcome meetings. Specify which calendar to use for scheduling. Add Additional Steps**: Extend the workflow to automatically assign the customer to a specific HubSpot list. Add a follow-up email or survey after the welcome meeting. This workflow is perfect for businesses seeking an efficient and personalized onboarding process, ensuring new customers feel welcomed and supported from day one.
by Sam Nesler
Syncs assignments and completion states to and fro between Canvas LMS and a Notion database. Automatically triggers every 2 hours during the schoolday by default (meaning 7 times a day), but also supports manual refreshing via webhooks. Setup You'll need a few things to get started: A Canvas API key. You can generate one by going to your Canvas account settings and clicking on the "New Access Token" button. The URL looks like https://canvas.wisc.edu/profile/settings You'll also need to replace URLs in Canvas nodes with your institution's domain, unless you're a student at UW-Madison. Canvas nodes are all the HTTP Request nodes except the one labelled "OpenAI Categorization", which is an OpenAI node and will require a key in a later step. A Notion integration token. You can find this by going to your Notion integrations page and clicking "Create new integration". You can make it a "Internal Integration". A Notion database to sync to. I made a template for use with the workflow, but you can use any database that has the following fields: Status (status): Status with at least the options "Not Started" and "Completed" - assignments start out "Not Started", and are marked "Completed" when they are submitted on Canvas. Estimate (select): Select with at least the options "XS", "S", "M", "L", "XL" - this is where the estimated time to complete the assignment will be stored. Even if you don't use AI, they'll start out as "M" Priority (select): Select with at least the options "Could Do", "Should Do", "Must Do" - assignments start out "Should Do" ID (text): this is where the ID of the assignment will be stored. We use this to sync without having a database on the server Due Date (date): this is where the due date of the assignment will be stored Class (text): this is where the name of the class will be stored Link (URL): this is where the link to the assignment will be stored The ID of the Notion database you want to sync to. You can find this by clicking "Share" in the top right of your database and copying the link. The ID is the part of the link that comes after https://www.notion.so/ and before ?v=. So for https://www.notion.so/tsuniiverse/1976e99d91128076b034e7379464560f?v=1976e99d911281e7bd4b000c2cbec692&pvs=4, the ID would be 1976e99d91128076b034e7379464560f. An OpenAI key for assignment length estimation or disable the node. Manual Refreshing Embed the production URL from the Webhook Trigger inside a "toggle list" or "toggle heading" inside Notion, then expand the heading to refresh, like so: