by Shadrack
Streamline your recruitment process with AI-powered resume analysis that goes beyond keyword matching. Overview This workflow revolutionizes hiring by using specialized AI agents to intelligently analyze resumes for different positions. Unlike traditional ATS systems that rely on keyword matching, this solution provides deep, contextual analysis of candidate qualifications and automatically manages the entire screening process from application to response. How it works Application Submission - Candidates complete a form with personal details, select their desired position (ICT, Customer Care, Accounting, or HR), and upload their resume Secure Storage - Resumes are automatically saved to Google Drive for permanent record-keeping Intelligent Routing - A switch node classifies applications by position and routes them to specialized AI agents AI Analysis - Position-specific agents (each with customized prompts and guardrails) extract text from PDFs and rate candidates on a 1-10 scale with detailed commentary CRM Integration - All data (timestamp, candidate info, position, score, AI comments, resume link) flows into Google Sheets for easy tracking Automated Response - A secondary workflow sends interview invitations to high-scoring candidates (7-10) and professional rejection emails to others (below 7) Key Benefits ✅ Contextual Analysis - AI understands skills and experience, not just keywords ✅ Faster Hiring - Close applications once you have enough qualified candidates ✅ No Expertise Required - HR teams don't need technical knowledge in every field ✅ Fully Automated - From submission to interview invitation without manual intervention ✅ Customizable - Adjust AI prompts and scoring criteria for each position ✅ Transparent - All AI reasoning is logged for review Set up steps Time to set up: ~30-45 minutes Prerequisites: n8n instance (cloud or self-hosted) Google Drive account Google Sheets account AI provider credentials (OpenAI, Anthropic, or compatible API) Email service (Gmail, SMTP, or other n8n-supported service) Quick Setup: Import the workflow into your n8n instance Connect your Google Drive and Google Sheets accounts Configure your AI provider credentials in the agent nodes Customize AI prompts for each position in the respective agent nodes (detailed instructions in sticky notes) Set up your email service credentials Customize email templates for invitations and rejections Test with sample resumes for each position Deploy your application form and share the link Detailed configuration instructions are included in sticky notes within the workflow. Use Cases Startups scaling their team quickly HR departments handling high application volumes Agencies managing recruitment for multiple clients Companies hiring for specialized technical roles Customization Options Adjust scoring thresholds for each position Modify AI evaluation criteria via prompts Add additional positions with dedicated agents Integrate with your existing HRIS or ATS Add SMS notifications for candidates Note: This workflow includes two separate flows - the main screening workflow and an automated response workflow. Both are included in the download.
by Cheng Siong Chin
Introduction Automates patient health monitoring by analyzing submitted health data via AI, determining alert necessity, and notifying family and doctors when critical conditions detected. How It Works Webhook receives health data, AI agent analyzes vitals using OpenRouter with structured parsing. Conditional logic checks alert necessity. If yes: prepares data, emails family, checks doctor criteria, sends doctor email, merges results. If no: skips alerts. Combines outcomes and responds to webhook. Workflow Template Webhook → Extract Data → AI Agent → [OpenRouter + Memory + Parser] → Check Alert → [Prepare + Email Family + Check Doctor + Email Doctor] OR [No Alert] → Merge → Combine → Respond Workflow Steps Reception & Extraction: Webhook receives vitals/symptoms, parses JSON payload AI Analysis: OpenRouter analyzes vitals against ranges, accesses history via Memory Tool, formats assessment via Output Parser Routing & Notification: Conditional logic checks severity. Alert path emails family/doctor if critical. No alert logs status Consolidation: Merges outcomes, sends webhook response Setup Instructions Configure webhook endpoint with auth token. Add OpenRouter API key and select model. Set up AI Agent with Memory Tool and Output Parser. Connect Gmail/SMTP for notifications with recipient addresses. Configure alert thresholds (temp >38.5°C, BP >140/90, HR <60/>100). Set doctor notification criteria. Prerequisites n8n instance, OpenRouter API key, AI model access, Patient database, Gmail/SMTP credentials, Family contacts, Doctor contacts, Webhook authentication Use Cases Chronic Disease: Diabetic submits glucose readings. AI detects >250mg/dL, alerts family and endocrinologist. Elderly Care: Senior's vitals monitored via wearable. AI identifies irregular rhythm, emails caregiver and cardiologist. Customization Adjust thresholds by demographics. Add vital types (O2, glucose trends). Customize AI prompts for conditions. Integrate SMS via Twilio. Add escalation logic. Include medication tracking. Connect EHR systems. Implement frequency limits. Add dashboard reporting. Benefits Rapid Response: Detects emergencies in seconds. Intelligent Filtering: Prevents false alarms. Family Peace: Automated notifications keep loved ones informed. Clinical Efficiency:
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Stay ahead of market changes with this Automated Price Intelligence System! This workflow monitors e-commerce product prices 3x daily using advanced web scraping and AI analysis, tracking price changes, comparing against strategic thresholds, and sending intelligent alerts for competitive pricing opportunities. Perfect for e-commerce teams, retailers, and pricing strategists maintaining market competitiveness. What This Template Does Triggers 3x daily (9 AM, 3 PM, 9 PM) for continuous price monitoring. Configures global settings for products, thresholds, and alert recipients. Parses product targets and maps price thresholds by category (premium/default/budget). Uses Decodo scraper to extract real-time pricing data from e-commerce sites. Analyzes product pages with AI to extract structured price information. Logs all price data to Google Sheets for historical tracking and trend analysis. Evaluates prices against thresholds with AI-powered strategic recommendations. Classifies alerts as CRITICAL or STANDARD based on price drop severity. Sends targeted email alerts with strategic action recommendations. Key Benefits Continuous price monitoring across multiple e-commerce platforms AI-powered strategic recommendations for pricing decisions Historical price tracking for trend analysis and forecasting Multi-tier alert system for critical vs. standard price changes Automated competitive intelligence without manual monitoring Configurable thresholds for different product categories Features Triple-daily scheduling for comprehensive market coverage Multi-product monitoring with individual threshold configuration AI-powered price extraction and data structuring Real-time web scraping with Decodo integration Strategic alert classification (CRITICAL/STANDARD) Automated email notifications with actionable insights Google Sheets integration for data centralization Batch processing for efficient multi-product handling Quality assurance with auto-fixing output parsing Requirements Decodo API credentials for web scraping OpenAI API credentials for AI analysis Google Sheets OAuth2 credentials with edit access Gmail OAuth2 credentials for email alerts Environment variables for configuration settings Product URLs with internal tracking IDs Target Audience E-commerce and retail pricing teams Competitive intelligence analysts Pricing strategy and revenue optimization teams E-commerce marketing and sales teams Retail operations and category managers Digital agency e-commerce specialists Step-by-Step Setup Instructions Connect Decodo API credentials for reliable web scraping Set up OpenAI credentials for AI price analysis and strategy recommendations Configure Google Sheets for price history tracking and logging Add Gmail credentials for critical and standard alert notifications Define your product URLs with internal IDs and threshold types Set price thresholds for premium, default, and budget categories Configure alert recipients for different notification levels Test with sample product URLs to verify data extraction and alerting Activate for automated triple-daily price intelligence monitoring Pro Tip: Use coupon code "YARON" to get 23K requests for testing the workflow using the Decodo scraping platform This workflow ensures you never miss a pricing opportunity with automated monitoring, intelligent analysis, and strategic alerting!
by Jordan
This n8n template automatically processes inbound form leads, researches their company via their website, and uses AI to write and send a highly personalized "speed to lead" email instantly. It also tracks subsequent appointment bookings to keep your CRM up to date. Use cases are focused on high velocity sales: Perfect for agency owners, founders, and sales teams who need to respond to new leads in seconds while maintaining a high quality, personal touch. Good to know At the time of writing, this template is optimized for GPT-4o to ensure the highest quality personalization. This template was originally designed using a Lovable form, but it works with any form builder (Typeform, Webflow, WordPress) that can send data via a webhook. How it works The workflow starts when a new lead submits your contact form via the Webhook node. It immediately checks your Airtable CRM to ensure the lead is new and not a duplicate. The system scrapes the provided website URL and uses custom code to clean the HTML data. AI analyzes the cleaned website data to generate a concise summary of the business. A second AI step uses this summary and the lead's goals to draft and send a personalized email via Gmail. The lead data and website summary are logged into your Airtable base. A parallel flow monitors your inbox for booking confirmations and updates the CRM automatically. How to use Create the necessary Base and Table in Airtable to store your leads. Connect your website form to the Webhook URL in the workflow. Update the OpenAI node prompts with your specific email signature and phone number. Run a test submission to ensure the website scraper is capturing data correctly. Requirements Airtable account OpenAI account (GPT-4o model recommended) Gmail account Customising this workflow You can easily adapt this for different communication channels. Try swapping the Gmail node for a Slack notification to alert your team of high value leads, or connect it to HubSpot instead of Airtable for enterprise CRM management.
by ScoutNow
Instantly Track Reddit Discussions and Generate Insights Manually monitoring Reddit for relevant discussions can be overwhelming. This automation does all the heavy lifting by automatically searching for keywords across selected subreddits or the entire Reddit, analyzing each post with GPT-5-Mini, and saving structured insights in your Notion database. Optionally, receive daily email notifications summarizing new discussions without lifting a finger. Key Features Daily Automated Reddit Monitoring** Checks Reddit once a day for new posts matching your keywords in selected subreddits or across all of Reddit. AI-Powered Analysis with Custom Prompts** Uses GPT-5-Mini to: Assign a relevance score to each post based on your defined priorities. Generate a suggested comment tailored to your audience or engagement strategy. Allow custom prompts so you can tweak AI behavior for lead generation, competitor monitoring, or brand mentions. Notion Integration** Saves all posts, relevance scores, and AI-generated comments directly into your Notion database for easy tracking and collaboration. Optional Email Alerts** Sends a clean, HTML-formatted email summarizing new discussions and insights. No API Keys Required for Reddit** Works with the official Reddit API without requiring users to create credentials. Customizable Inputs** Users define keywords, subreddits, relevance criteria, custom prompts, and notification preferences with minimal setup. Setup Instructions 1. Configure Keywords and Subreddits Open the Define Keywords and Subreddits node. Enter your target keywords and choose specific subreddits, or enable search across all Reddit. Map the inputs to the workflow’s search nodes. 2. Connect OpenAI (GPT-5-Mini) Create an OpenAI account and obtain your API key. Open the AI Analysis Node in the workflow. Paste your OpenAI API key into the credentials field. Customize your prompts to define how GPT-5-Mini: Scores relevance based on your specific criteria. Generates suggested comments tailored to your objectives (e.g., lead generation, brand engagement, competitor analysis). Test the node to ensure relevance scoring and suggested comments are generated correctly. 3. Connect Notion Duplicate the provided Notion template into your workspace. Obtain your Notion integration secret from Notion settings. Paste the secret into the Add Each Post to Notion Database node. Test the connection to ensure posts are saved correctly. 4. Optional Email Notifications Setup Google API credentials in the Google Cloud Console. Authenticate Gmail in the workflow. Map the recipient email and customize the notification format. Test sending an email to confirm delivery. 5. Configure Workflow Trigger Import the workflow into n8n. Ensure the daily trigger is enabled. Optionally adjust schedule or notification settings. How It Works Daily Trigger: Starts the workflow once per day. Reddit Search: Queries new posts based on user-defined keywords. AI Analysis: GPT-5-Mini evaluates relevance score and generates suggested comments based on your custom prompts. Save & Notify: Posts are saved to Notion and optionally summarized via email. Example Use Cases Generate leads by identifying potential prospects discussing your industry or product. Monitor brand mentions to stay on top of conversations around your company. Track competitor activity across relevant subreddits. Keep a daily digest of trending posts for market research or content ideas. Automate engagement by posting AI-generated suggested comments. Requirements | Tool | Purpose | | -------------- | ------------------------------------------------- | | Notion Account | Store posts, relevance scores, and comments | | Gmail Account | Optional daily email notifications | | n8n | Run, schedule, and manage the workflow | | OpenAI API Key | Access GPT-5-Mini for relevance scoring and AI-generated comments |
by Daiki Takayama
Transform your Google Form into an intelligent customer support system that automatically analyzes, prioritizes, and responds to every inquiry with AI-powered personalization. Who's it for This workflow is perfect for: Customer support teams** handling high volumes of form submissions Small businesses** looking to provide instant, professional responses 24/7 E-commerce stores** managing product inquiries, complaints, and refunds Service providers** triaging appointment requests and support tickets Event organizers** responding to registration questions How it works When someone submits your Google Form, this workflow: Triggers instantly** from new Google Sheets row Analyzes with AI** to determine urgency (low/medium/high), category (technical/sales/support/billing), and sentiment (positive/neutral/negative) Routes intelligently** based on priority level to appropriate response template Generates personalized reply** using OpenAI, tailored to urgency and sentiment Sends auto-reply** via Gmail with professional formatting Alerts your team** on Slack for high-priority cases requiring human follow-up Logs everything** to tracking sheet for analytics and audit trail Set up steps Time to set up: approximately 15-20 minutes Prerequisites Google Form connected to Google Sheets OpenAI API key (get one at platform.openai.com) Gmail account for sending emails Slack workspace (optional, for team alerts) Configuration steps Connect Google Sheets Trigger to your form response sheet Update column names in "Map Form Column Names" node (default: Name, Email Address, Inquiry) Add OpenAI credentials - uses GPT-4o-mini for cost efficiency (approximately $0.001 per inquiry) Connect Gmail via OAuth2 for sending auto-replies Create tracking sheet with columns: timestamp, name, email, urgency, category, sentiment, summary, keywords, subject, inquiry (Optional) Connect Slack for high-priority notifications Test with sample data before activating Requirements Google Form (free) Google Sheets (free) OpenAI API account (approximately $0.001-0.002 per inquiry with GPT-4o-mini) Gmail account (free) Slack workspace (optional, free tier available) n8n Cloud or self-hosted instance How to customize Adjust triage criteria: Edit the "Analyze with AI Triage" prompt to define what qualifies as high/medium/low urgency for your business. Add custom categories beyond the default five (technical/sales/support/billing/general). Modify response templates: Customize the three response generation nodes to match your brand voice. Add company-specific information, policies, or FAQ links. Adjust tone from formal to casual based on your audience. Add advanced features: Insert CRM integration to check customer history before responding Add condition nodes to route specific categories to different team members Include file attachments (brochures, manuals) based on inquiry type Implement follow-up sequences with Wait nodes Connect multiple notification channels (Discord, email, SMS) Multi-language support: The AI automatically detects and responds in the inquiry's language. Customize prompts with language-specific instructions if needed. Use cases Product support:** Auto-respond to "how to use" questions with relevant documentation Complaint management:** Immediately acknowledge issues and alert team for urgent follow-up Lead qualification:** Instantly engage sales inquiries while routing to appropriate sales rep Appointment requests:** Confirm booking inquiries and provide next steps Feedback collection:** Thank customers and categorize feedback for analysis
by Khairul Muhtadin
This workflow automatically checks a list of desired domain names for availability every 30 minutes. Using the Porkbun API and Google Sheets, it instantly sends detailed notifications via Gmail and Discord the moment a domain becomes available, so you can secure it before anyone else. Why Use This Workflow? Time Savings: Eliminates hours of manual domain checking each week. Set your list once and let the automation monitor your targets 24/7. Competitive Edge: Gain a critical speed advantage in acquiring high-value or expiring domains the second they become available. Scalability: Effortlessly monitor hundreds of domains simultaneously without any extra effort or performance degradation. Ideal For Domain Investors:** Automatically track and snipe valuable expiring domains for their portfolio without constant manual checks. Marketing Agencies & SEO Specialists:** Secure brandable domains for new clients or build out private blog networks by catching domains as soon as they drop. Startups & Entrepreneurs:** Monitor desired brand names without the daily hassle, ensuring they get the perfect domain the moment it's available. How It Works Trigger: A schedule trigger initiates the workflow every 30 minutes. Data Collection: It retrieves a list of domains to monitor from a designated Google Sheet, specifically targeting rows marked as "no" for availability. Processing: The workflow iterates through each domain one by one. Intelligence Layer: It makes an API call to Porkbun to check the current availability of the domain. An IF node then determines if the domain is available (avail == yes). Output & Delivery: If a domain is available, the workflow sends a rich HTML email via Gmail and a formatted message to a Discord channel, complete with pricing details and a direct registration link. Storage & Logging: The Google Sheet is automatically updated to mark the domain as "available," preventing redundant notifications on future runs. Setup Guide Prerequisites | Requirement | Type | Purpose | |-------------|------|---------| | n8n instance | Essential | Workflow execution platform | | Porkbun Account | Essential | API Access for domain checks | | Google Cloud Platform | Essential | Storing domain list (Sheets) & sending alerts (Gmail) | | Discord Server | Optional | Real-time channel notifications | Installation Steps Import the JSON file to your n8n instance. Create a Google Sheet with two columns: Domain (e.g., example.com) and isAvailable (e.g., no). Configure credentials: Porkbun: Log in to Porkbun, go to API Access, create a new key, and copy the API Key and Secret Key into the HTTP Request nodes. A "Validate API KEY" node is included for testing your credentials. Google Sheets/Gmail: Authenticate your Google account for the Google Sheets and Gmail nodes. Discord: Create a Discord Bot and add the credentials to the Discord node. Update environment-specific values: Get Domains from Sheet: Enter your Google Sheet ID and select the correct sheet name. Send Email Alert: Set your recipient's email address in the "To" field. Send Discord Notification: Select your desired Server and Channel ID. Test execution: Add a domain you know is available to your Google Sheet (with isAvailable set to "no"). Run the workflow manually to verify that all connections work and you receive notifications. Technical Details Core Nodes | Node | Purpose | Key Configuration | |------|---------|-------------------| | Schedule Trigger | Initiates the workflow on a recurring basis. | Set the desired interval (default: 30 minutes). | | Google Sheets | Reads the domain list and updates its status. | Sheet ID, Sheet Name, and column mapping. | | SplitInBatches | Processes each domain from the list individually. | Batch size is set to 1 to check domains sequentially. | | HTTP Request | Queries the Porkbun API for domain availability. | Porkbun API endpoint and credentials. | | IF | Routes the workflow based on the API response. | Checks if response.avail equals "yes". | | Gmail | Sends a detailed email alert for available domains. | Recipient email, subject, and HTML body. | | Discord | Sends a concise notification to a Discord channel. | Server ID, Channel ID, and message content. | | Wait | Prevents API rate-limiting. | Pauses for 10 seconds between checking domains. | Workflow Logic The workflow is triggered by a schedule, fetching a list of domains from a Google Sheet. It uses the SplitInBatches node to loop through each domain sequentially. For every domain, it calls the Porkbun API. An IF node checks the response; if available, it triggers notifications and updates the Google Sheet. A Wait node is crucial for respecting API rate limits, ensuring the workflow runs smoothly even with large domain lists. Customization Options Basic Adjustments: Check Frequency**: Modify the "Schedule Trigger" node to run more or less frequently. Notification Channels**: Remove the Gmail or Discord nodes, or add new ones like Slack or Telegram. Email Content**: Customize the HTML in the Gmail node to match your branding. Advanced Enhancements: Auto-Registration**: Extend the workflow to use Porkbun's domain registration API endpoint to automatically purchase the domain when it becomes available (use with caution). Advanced Filtering**: Add logic to only send notifications for domains with specific TLDs (.com, .io, etc.) or that are not marked as "premium." Tiered Notifications**: Set up different notification channels based on the perceived value of the domain, sending high-priority alerts via SMS for critical domains. Troubleshooting Common Issues: | Problem | Cause | Solution | |---------|-------|----------| | Workflow fails at HTTP Request node | Invalid Porkbun API credentials. | Use the separate "Validate API KEY" node to test your keys directly. Regenerate them if needed. | | No domains are processed | Google Sheets node configuration error or the sheet is empty. | Verify the Sheet ID is correct and that the isAvailable column contains "no" for the domains you want to check. | | Authentication errors | Google or Discord credentials have expired or lack permissions. | Re-authenticate the respective nodes in the n8n credentials panel. | Created by: Khaisa Studio Category: Monitoring Tags: Porkbun, Domain, Automation, Google Sheets, Notifications Need custom workflows? Contact us Connect with the creator: Portfolio • Workflows • LinkedIn • Medium • Threads
by Mohamed Abubakkar
Overall Description This workflow fully automates the process of creating and publishing professional LinkedIn posts based on real-world experience in AI automation and workflow engineering, enriched with latest industry insights from the internet, and paired with a relevant high-quality image The workflow ensures: Content relevance to AI automation, backend/frontend systems, IT Professional, human-friendly post tone Image validation before posting Post validation before publishing Automatic success/failure email alerts Worflow Logic Trigger workflow (daily, weekly, custom date) Define posting context (based on your expertise) Fetch latest AI/automation trends (you can change based on your interested trends) Filter insights relevant to your field Generate linkedin ready post using AI Agent (OpenAI API) Generate image from Open AI Image model related to the post Validate image and text quality Merge approved text and image Convert content into LinkedIn format Publish post on LinkedIn Send email notification on success or failure Setup Instructions Credentials OpenAI API Key – For text generation. LinkedIn API Token – For publishing posts. Image Generation API – DALL-E, Pollinations, or MidJourney. Gmail / SMTP – For notifications. Optional APIs: SERP API for trending topics.
by Ramsey Njire
Overview Stop digging through execution logs to find out why a workflow failed. This template provides a "set-it-and-forget-it" monitoring system that uses AI to automatically debug your n8n workflows. Instead of just getting a simple error message, you'll receive a detailed email notification with a root cause analysis and a step-by-step suggested fix from a Gemini-powered AI agent. This saves you valuable time, helps you resolve issues faster, and ensures your critical automations are always running smoothly. Prerequisites Before you begin the setup, please make sure you have the following accounts and information ready: An n8n Instance:** This workflow must be hosted on an active n8n instance. n8n API Key:* You'll need to generate an API key from your n8n instance (Settings > API*). OpenRouter Account:** An account with OpenRouter.ai to access various AI models like Gemini. Google (Gmail) Account:** To send and receive the email notifications. n8n Environment Variable:** Your n8n instance must have the N8N_EDITOR_BASE_URL environment variable configured. This is essential for generating correct links to your workflow executions. Use Cases This template is perfect for anyone who wants to proactively monitor their automations. Here are a few examples: Monitoring Critical Webhooks:** Get instant, intelligent alerts if a workflow handling data from a payment gateway (like Stripe) or a form submission fails due to unexpected data, minimizing service disruption. Managing Complex Automations:** For long, multi-step workflows, the AI can immediately pinpoint the exact failing node and suggest fixes for complex data mapping errors or API issues that are otherwise difficult to trace. Developers & Agencies:** If you manage multiple n8n workflows for clients or different projects, this provides a centralized monitoring system that helps you stay on top of all automation issues without having to manually check each one. How it works This workflow acts as an automated debugging assistant for your n8n instance 🤖. When any of your other workflows fail during an automatic (production) execution, this template will: Capture the error and use the n8n API to fetch the complete structure of the failed workflow for full context. Send the error details and workflow structure to a powerful AI agent (via OpenRouter) for a detailed root cause analysis. Format the AI's diagnosis and suggested fix into a clean HTML email and send it to you via Gmail, including a direct link to the failed execution. Important Note: As per the n8n documentation, the Error Trigger node only runs for errors that happen in production executions (e.g., from a webhook call or a schedule). It will not run when you test a workflow by clicking "Execute workflow" manually. Set up steps (Estimated setup time: 10 minutes) There are two main parts to the setup: configuring this workflow, and then connecting it to the workflows you want to monitor. Part A: Configure the AI Debugger Workflow n8n API Credentials: Create an API key in your n8n instance under Settings > API and add it as a "Header Auth" credential for the Get Workflow JSON node. OpenRouter Credentials: Add your OpenRouter API key to the OpenRouter Chat Model node. Gmail Credentials: Authenticate your Gmail account in the Send Debugging Email node. Recipient Email: Update the "To" field in the final Send Debugging Email node with your own email address. Environment Variable: Ensure your N8N_EDITOR_BASE_URL environment variable is correctly set for your n8n instance. This is required for building the API and execution links. Part B: Connect to Your Other Workflows For every workflow you want this debugger to monitor, you must link it in that workflow's settings. Go to the workflow you want to monitor (your "target" workflow). Click the three dots (...) in the top right corner and select Settings. 3. In the Error Workflow dropdown, select this "AI-Powered Workflow Debugger & Notifier" workflow. Click Save. Now, whenever that target workflow fails during a production run, this debugger workflow will automatically be triggered to analyze the error and notify you. Note that the error workflow only gets triggered on automatic runs, not manual ones: This is a feature of how the error trigger works in n8n. To ensure this, make sure the target workflow is active. It won't trigger the error workflow if it's turned off.
by Oneclick AI Squad
Enhance financial oversight with this automated n8n workflow. Triggered every 5 minutes, it fetches real-time bank transactions via an API, enriches and transforms the data, and applies smart logic to detect critical, high, and medium priority alerts based on error conditions, amounts, or risk scores. It sends multi-channel notifications via email and Slack, logs all data to Google Sheets, and generates summary statistics for comprehensive tracking. 💰🚨 Key Features Real-time monitoring every 5 minutes for instant alerts. Smart prioritization (Critical, High, Medium) based on risk and errors. Multi-channel notifications via email and Slack. Detailed logging and summary reports in Google Sheets. How It Works Schedule Trigger: Runs every 5 minutes. Fetch Transactions: HTTP request retrieves real-time transaction data. API Error?: If condition for error logic is met, sends error alert. Enrich & Transform Data: Advanced risk calculation enhances data. Critical Alert?: If condition (50% or risk > 8) is met, raises alert. High Priority?: If condition (5% or risk > 7) is met, raises alert. Medium Priority?: If condition is met, raises alert. Log Priority to Sheet: Google Sheets appends critical, high, or medium priority data. Send Critical Email: HTML email to execute sheets append. Send High Priority Email: Email to finance team. Send High Priority Slack: Slack notification to finance team. Send Medium Priority Email: Email to finance team. Merge All Alerts: Combines all alerts for comprehensive tracking. Generate Summary Stats: Code block for analytics. Log Summary to Sheet: Summary statistics storage. Setup Instructions Import the workflow into n8n and configure the bank API credentials in "Fetch Transactions." Set up Google Sheets OAuth2 and replace the sheet ID for logging nodes. Configure Gmail API Key and Slack Bot Token for alerts. Test the workflow with sample transaction data exceeding risk or amount thresholds. Adjust priority conditions (e.g., 50%, 5%, risk > 8) based on your risk policy. Prerequisites Bank API access with real-time transaction data (e.g., https://api.bank.com) Google Sheets OAuth2 credentials Gmail API Key for email alerts Slack Bot Token (with chat:write permissions) Structured transaction data format Google Sheet Structure: Create a sheet with columns: Transaction ID Amount Date Risk Score Priority (Critical/High/Medium) Alert Sent Summary Stats Updated At Modification Options Adjust the "Schedule Trigger" interval (e.g., every 10 minutes). Modify "Critical Alert?" and "High Priority?" conditions for custom thresholds. Customize email and Slack templates with branded messaging. Integrate with fraud detection tools for enhanced risk analysis. Enhance "Generate Summary Stats" with additional metrics (e.g., average risk). Discover more workflows – Get in touch with us
by ben daamer
Who it's for Legal, Procurement, and Compliance teams at mid-size companies. ESN and agencies selling AI-powered contract review as a service. How it works Upload a contract PDF through a rich form (with metadata: contract type, counterparty, department, submitter). The workflow first checks Supabase for duplicates to avoid re-analyzing the same contract. Then it runs a two-pass AI analysis: Pass 1 classifies the contract (type, parties, dates, jurisdiction), Pass 2 performs deep risk analysis (clause-by-clause risk levels, missing clauses, obligations, negotiation points, compliance flags). Results are scored 0–100 with risk levels (CRITICAL / HIGH / MEDIUM / LOW). A structured report is generated as both Slack Block Kit (rich formatting) and HTML email. High-risk contracts trigger a Slack alert with full blocks and an email report to the submitter. Low-risk contracts get a summary Slack notification. All analyses are stored in Supabase with full metadata for audit. An Error Trigger sends admin notifications to a dedicated Slack channel. How to set up Run the SQL from the "Supabase Schema" sticky note in Supabase SQL Editor. Set environment variables: SUPABASE_URL, SUPABASE_SERVICE_KEY. Add credentials in n8n: OpenAI (Header Auth), Supabase, Slack, Gmail (optional). Configure the Config node: RISK_THRESHOLD, SLACK_CHANNEL, ADMIN_SLACK_CHANNEL, AI_MODEL, ALERT_EMAIL, ENABLE_EMAIL, CONTRACT_LANG. Activate the workflow and share the Form URL with your team.
by NODA shuichi
Description: An intelligent alarm system that wakes you up early only when necessary. 🛡️🤖 This workflow monitors your local weather and train status every morning at 5:00 AM. It uses Google Gemini to analyze the situation. If there is heavy rain or a train delay, it sends an emergency alert immediately (and can trigger smart home devices). If everything is normal, it waits until your usual wake-up time to send a calm briefing. Key Features: AI Context Awareness: Uses Gemini to intelligently judge "Emergency" situations based on weather descriptions and news headlines. Dynamic Notification: Emergency Mode: Immediate Email alert + Optional SwitchBot trigger (e.g., turn on lights). Normal Mode: Delays notification until your scheduled wake-up time (90 mins later). Targeted Monitoring: Searches specific train lines via Google News RSS. How it works: Trigger: Runs daily at 5:00 AM. Fetch: Gets weather from OpenWeatherMap and train news from Google News RSS. Judge: Gemini analyzes the data. Action: Routes the notification based on the status (Emergency/Normal). Setup Requirements: Credentials: OpenWeatherMap API, Google Gemini API, Gmail. Config: Open the "1. Configuration" node to set your Location, Train Line, and Email.