by SpaGreen Creative
WhatsApp Number Verify & Confirmation System with Rapiwa API and Google Sheets Who is this for? This n8n workflow makes it easy to verify WhatsApp numbers submitted through a form. When someone fills out the form, the automation kicks in—capturing the data via a webhook, checking the WhatsApp number using the Rapiwa API, and sending a confirmation message if the number is valid. All submissions, whether verified or not, are logged into a Google Sheet with a clear status. It’s a great solution for businesses, marketers, or developers who need a reliable way to verify leads, manage event signups, or onboard customers using WhatsApp. How it works? This n8n automation listens for form submissions via a webhook, validates the provided WhatsApp number using the Rapiwa API, sends a confirmation message if the number is verified, and then appends the submission data to a Google Sheet, marking each entry as verified or unverified. Features Webhook Trigger**: Captures form submissions via HTTP POST Data Cleaning**: Formats and sanitizes the WhatsApp number Rapiwa API Integration**: Checks if the number is registered on WhatsApp Conditional Messaging**: Sends confirmation messages only to verified WhatsApp users Google Sheets Integration**: Appends all submissions with a validity status Auto Timestamping**: Adds the submission date in YYYY-MM-DD format Throttling Support**: Built-in delay to avoid hitting API or sheet rate limits Separation of Verified/Unverified**: Distinct handling for both types of entries Nodes Used in the Workflow Webhook** Format Webhook Response Data** (Code) Loop Over Items** (Split In Batches) Cleane Number** (Code) check valid whatsapp number** (HTTP Request) If** (Conditional) Send Message Using Rapiwa** verified append row in sheet** (Google Sheets) unverified append row in sheet** (Google Sheets) Wait1** How to set up? Webhook Add a Webhook node to the canvas. Set HTTP Method to POST. Copy the Webhook URL path (/a9b6a936-e5f2-4xxxxxxxxxe0a970d5). In your frontend form or app, make a POST request to: The request body should include: { "business_name": "ABC Corp", "location": "New York", "whatsapp": "+1 234-567-8901", "email": "user@example.com", "name": "John Doe" } Format Webhook Response Data Add a Code node after the Webhook node. Use this JavaScript code: const result = $input.all().map(item => { const body = item.json.body || {}; const submitted_date = new Date().toISOString().split('T')[0]; return { business_name: body.business_name, location: body.location, whatsapp: body.whatsapp, email: body.email, name: body.name, submitted_date: submitted_date }; }); return result; Loop Over Items Insert a SplitInBatches node after the data formatting. Set the Batch Size to a reasonable number (e.g. 1 or 10). This is useful for processing multiple submissions at once, especially if your webhook receives arrays of entries. Note: If you expect only one submission at a time, it still helps future-proof your workflow. Cleane Number Add a Code node named Cleane Number. Paste the following JavaScript: const items = $input.all(); const updatedItems = items.map((item) => { const waNo = item?.json["whatsapp"]; const waNoStr = typeof waNo === 'string' ? waNo : (waNo !== undefined && waNo !== null ? String(waNo) : ""); const cleanedNumber = waNoStr.replace(/\D/g, ""); item.json["whatsapp"] = cleanedNumber; return item; }); return updatedItems; Check WhatsApp Number using Rapiwa Add an HTTP Request node. Set: Method: POST URL: https://app.rapiwa.com/api/verify-whatsapp Add authentication: Type: HTTP Bearer Credentials: Select or create Rapiwa token In Body Parameters, add: number: ={{ $json.whatsapp }} This API call checks if the WhatsApp number exists and is valid. Expected Output: { "success": true, "data": { "number": "+88017XXXXXXXX", "exists": true, "jid": "88017XXXXXXXXXXXXX", "message": "✅ Number is on WhatsApp" } } Conditional If Check Add an If node after the Rapiwa validation. Configure the condition: Left Value: ={{ $json.data.exists }} Operation: true If true → valid number → go to messaging and append as "verified". If false → go to unverified sheet directly. Note: This step branches the flow based on the WhatsApp verification result. Send WhatsApp Message (Rapiwa) Add an HTTP Request node under the TRUE branch of the If node. Set: Method: POST URL: https://app.rapiwa.com/api/send-message Authentication: Type: HTTP Bearer Use same Rapiwa token Body Parameters: number: ={{ $json.data.phone }} message_type: text message: Hi {{ $('Cleane Number').item.json.name }}, Thanks! Your form has been submitted successfully. This sends a confirmation message via WhatsApp to the verified number. Google Sheets – Verified Data Add a Google Sheets node under the TRUE branch (after the message is sent). Set: Operation: Append Document ID: Choose your connected Google Sheet Sheet Name: Set to your active sheet (e.g., Sheet1) Column Mapping: Business Name: ={{ $('Cleane Number').item.json.business_name }} Location: ={{ $('Cleane Number').item.json.location }} WhatsApp Number: ={{ $('Cleane Number').item.json.whatsapp }} Email : ={{ $('Cleane Number').item.json.email }} Name: ={{ $('Cleane Number').item.json.name }} Date: ={{ $('Cleane Number').item.json.submitted_date }} validity: verified Use OAuth2 Google Sheets credentials for access. Note: Make sure the sheet has matching column headers. Google Sheets – Unverified Data Add a Google Sheets node under the FALSE branch of the If node. Use the same settings as the verified node, but set: validity: unverified This stores entries with unverified WhatsApp numbers in the same Google Sheet. Wait Node Add a Wait node after both Google Sheets nodes. Set Wait Time: Value: 2 seconds This delay prevents API throttling and adds buffer time before processing the next item in the batch. Google Sheet Column Reference A Google Sheet formatted like this ➤ Sample Sheet | Business Name | Location | WhatsApp Number | Email | Name | validity | Date | |---------------------|--------------------|------------------|----------------------|------------------|------------|------------| | SpaGreen Creative | Dhaka, Bangladesh | 8801322827799| contact@spagreen.net | Abdul Mannan | unverified | 2025-09-14 | | SpaGreen Creative | Bagladesh | 8801322827799| contact@spagreen.net| Abdul Mannan | verified | 2025-09-14 | > Note: The Email column includes a trailing space. Ensure your column headers match exactly to prevent data misalignment. How to customize the workflow Modify confirmation message with your brand tone Add input validation for missing or malformed fields Route unverified submissions to a separate spreadsheet or alert channel Add Slack or email notifications on new verified entries Notes & Warnings Ensure your Google Sheets credential has access to the target sheet Rapiwa requires an active subscription for API access Monitor Rapiwa API limits and adjust wait time as needed Keep your webhook URL protected to avoid misuse Support & Community WhatsApp Support: Chat Now Discord: Join SpaGreen Community Facebook Group: SpaGreen Support Website: spagreen.net Developer Portfolio: Codecanyon SpaGreen
by khaled
What Problem Does It Solve? Business owners, managers, and accountants waste valuable time manually entering daily expenses, supplier payments, and employee advances into Odoo. Getting quick balance reports usually requires logging into the ERP, navigating multiple menus, and generating complex reports. Managing post-dated checks often relies on manual tracking, leading to missed due dates. This workflow solves these by: -- Allowing users to record financial transactions simply by sending a natural language message (e.g., via Telegram or Botpress). -- Automatically fetching real-time account balances and supplier statements, returning them instantly in the chat. -- Setting up automated calendar reminders for post-dated check due dates. -- Handling the entire double-entry accounting process in the background without human intervention. How to Configure It Chat Platform Setup** -- Add the webhook URL from this workflow to your Telegram Bot, Botpress, or preferred chat interface. Odoo Setup** -- Connect your Odoo credentials in n8n. -- Open the "Build [X] Entry" code nodes and replace the placeholder journal_id and currency_id with your actual Odoo system IDs. AI Setup** -- Add your OpenAI API key (or swap the node for Google Gemini/Anthropic). -- Open the "AI Financial Agent" node and update the # ACCOUNT MAPPING section with your specific Odoo Chart of Accounts codes. Calendar Setup (Optional)** -- Connect your Google Calendar credentials if you want the workflow to automatically schedule reminders for check due dates. How It Works Webhook catches the new text message from your chat platform. An AI Agent analyzes the Arabic natural language and extracts the intent, amount, date, check details, and specific account categories. Routing: -- For Expenses, Payments, or Advances → The workflow searches Odoo for the correct IDs, builds a balanced double-entry journal record, creates it, and posts it. -- For Post-Dated Checks → Extracts the due date and creates a Google Calendar event before posting the entry to Odoo. -- For Balance Inquiries → Fetches the relevant ledger lines, calculates total debits/credits, and formats a clean Arabic text summary. A success confirmation or the requested financial report is instantly sent back to the user in the chat. Customization Ideas Expand the AI prompt and routing switch to handle Customer Invoices or internal Petty Cash transfers. Add an approval step (e.g., sending a Slack/Email button) before the workflow officially "posts" large transactions in Odoo. Change the AI prompt to support multiple languages or different regional dialects. Log a backup of all financial chat requests into Google Sheets or a Notion database for auditing. For more info Contact Me
by Cheng Siong Chin
How It Works This workflow automates student progress monitoring and academic intervention orchestration through intelligent AI-driven analysis. Designed for educational institutions, learning management systems, and academic advisors, it solves the critical challenge of identifying at-risk students while coordinating timely interventions across faculty and support services. The system receives student data via webhook, fetches historical learning records, and merges these sources for comprehensive progress analysis. It employs a dual-agent AI framework for student progress validation and academic orchestration, detecting performance gaps, engagement issues, and intervention opportunities. The workflow intelligently routes findings based on validation status, triggering orchestration actions for students requiring support while logging compliant progress for successful learners. By executing multi-channel interventions through HTTP APIs and email notifications, it ensures educators and students receive timely guidance while maintaining complete audit trails for academic accountability and accreditation compliance. Setup Steps Configure Student Data Webhook trigger endpoint Connect Workflow Configuration node with academic performance parameters Set up Fetch Student Learning History node with LMS API credentials Configure Merge Student Data node for data consolidation Connect Student Progress Validation Agent with Claude/OpenAI API credentials Set up AI processing nodes Configure Route by Validation Status node with performance thresholds Connect Academic Orchestration Agent with AI API credentials for intervention planning Set up orchestration processing Prerequisites Claude/OpenAI API credentials for AI agents, learning management system API access Use Cases Universities identifying students requiring academic support, online learning platforms detecting engagement drops Customization Adjust validation thresholds for institutional academic standards Benefits Reduces student identification lag by 75%, eliminates manual progress tracking
by Roshan Ramani
Who's it for This workflow is ideal for: Content creators who want to replicate successful LinkedIn strategies Social media managers monitoring competitor content performance Marketing teams analyzing trending topics in their industry Personal brands looking to create data-driven content Agencies managing multiple LinkedIn accounts What it does This comprehensive workflow automates the entire LinkedIn content lifecycle: it scrapes viral posts from target accounts, analyzes engagement patterns, identifies trending topics, generates original AI-powered content based on those trends, creates accompanying images, and automatically publishes to your LinkedIn profile or company page. How it works Phase 1: Data Collection (Runs every 12 hours) Scheduler triggers the workflow twice daily Fetches LinkedIn profile URLs from Google Sheets Processes profiles in batches of 3 to respect API limits Uses Apify API to scrape recent posts from each profile Adds 3-second delays between requests to avoid rate limiting Filters for high-engagement posts (20+ likes, comments, or reposts) Saves viral posts to Google Sheets with full metadata Phase 2: Content Generation (Triggered by new data) Monitors Google Sheets for new viral posts every minute Filters posts published within the last 3 days that haven't been analyzed Aggregates trending content into a single dataset Analyzes patterns using Google Gemini AI to identify: Common themes and topics Engagement triggers and hooks Successful content structures Trending hashtags and formats Generates original LinkedIn post with proper formatting Creates AI image prompt optimized for minimal text Generates professional image using Google Imagen Publishes complete post to your LinkedIn account Marks analyzed posts as complete to prevent duplication Setup steps 1. Configure Google Sheets Create a new Google Sheet with two tabs: Tab 1: "usernames & links" - Add LinkedIn profile URLs you want to monitor Tab 2: "scrape data" - Leave empty (auto-populated by workflow) Connect your Google Sheets credentials in both nodes Replace all instances of YOUR_GOOGLE_SHEET_ID with your actual sheet ID Replace SHEET_GID values with your actual sheet GIDs 2. Set up Apify API Sign up for Apify account and get API token Replace YOUR_APIFY_API_TOKEN in "Scrape LinkedIn Posts API" node Note: Apify has free tier with limited requests 3. Configure Google Gemini credentials Obtain Google PaLM API credentials Add credentials to both "Google Gemini Chat Model" and "Generate an image" nodes 4. Set up LinkedIn publishing Connect your LinkedIn credentials in "Publish to LinkedIn" node If posting as organization, replace YOUR_LINKEDIN_ORGANIZATION_ID with your company page ID If posting as individual, change "postAs" parameter to "person" 5. Configure scheduling Default schedule: every 12 hours Adjust "LinkedIn Content Automation Scheduler" trigger if needed Consider your API rate limits when changing frequency 6. Test the workflow Manually trigger Phase 1 to scrape posts Verify data appears in Google Sheets "scrape data" tab Wait for Phase 2 trigger or manually activate it Check that content is generated and published correctly Verify posts are marked as analyzed in Google Sheets Requirements Google Sheets API access (free) Google Sheets Trigger OAuth2 (free) Apify API token (free tier available, $49/month for more) Google PaLM/Gemini API key (pay-per-use pricing) LinkedIn OAuth credentials (free) How to customize Adjust scraping targets: Add more LinkedIn profile URLs to your Google Sheets Change batch size in "Process Profiles in Batches" (default: 3) Modify post limit per profile in Apify API call (default: 1 post) Modify engagement filters: Edit "Filter High-Engagement Posts" node thresholds Default: 20+ likes OR 20+ comments OR 20+ reposts Adjust based on your niche's typical engagement rates Add additional criteria like views or impressions Customize content analysis window: Change "Filter Recent Posts (3 Days)" to analyze different timeframes Options: 24 hours for fast-moving trends, 7 days for broader patterns Balance between recency and data volume Refine AI content generation: Edit system prompt in "LinkedIn Content Strategy AI" node Adjust content length, tone, or style preferences Add industry-specific guidelines Include brand voice requirements Modify hashtag strategy Customize image generation: Edit image prompt structure in AI prompt Change visual style, colors, or composition Adjust for brand guidelines Modify dimensions or aspect ratios Change posting schedule: Adjust "LinkedIn Content Automation Scheduler" frequency Consider optimal posting times for your audience Balance between content quality and posting frequency Coordinate with other marketing activities Enhance data collection: Increase posts per profile in Apify settings Add more profile URLs to monitor Implement competitor tracking Track additional metrics like impressions or click-through rates Add notifications: Connect Slack/Email nodes after successful posts Set up alerts for high-performing content Create reports of analyzed trends Monitor API usage and errors
by Oneclick AI Squad
Automate your post-event networking with this intelligent n8n workflow. Triggered instantly after an event, it collects attendee and interaction data, enriches profiles with LinkedIn insights, and uses GPT-4 to analyze engagement and generate tailored follow-up messages. High-value leads are prioritized, messages are sent via email, LinkedIn, or Slack, and all activity is logged in your CRM and database. Save hours of manual follow-up while boosting relationship-building and ROI. 🤝✨ Advanced Features Webhook automation** – Starts instantly on event completion Multi-Source Enrichment** – Combines event data, interactions, and LinkedIn profiles AI-Powered Insights** – GPT-4 analyzes behavior and suggests personalized talking points Smart Priority Filtering** – Routes leads into High, Medium, and Low priority paths Personalized Content Generation** – AI crafts custom emails and LinkedIn messages Multi-Channel Outreach** – Sends via Email, LinkedIn DM, and Slack CRM Integration** – Automatically updates HubSpot with contact notes and engagement PostgreSQL Logging** – Stores full interaction history and analytics ROI Dashboard** – Tracks response rates, meetings booked, and pipeline impact What It Does Collects attendee data from your event platform Enriches with LinkedIn profiles & real-time interaction logs Scores networking potential using engagement algorithms Uses AI to analyze conversations, roles, and mutual interests Generates hyper-personalized follow-up emails and LinkedIn messages Sends messages through preferred channels (email, LinkedIn, Slack) Updates HubSpot CRM with follow-up status and next steps Logs all actions and tracks analytics for performance reporting Workflow Process The Webhook Trigger initiates the workflow via POST request with event and attendee data. Get Attendees** fetches participant list from the event platform. Get Interactions** pulls Q&A, chat, poll, and networking activity logs. Enrich LinkedIn Data** retrieves professional profiles, job titles, and company details via LinkedIn API. Merge & Enrich Data** combines all sources into a unified lead profile. AI Analyze Profile** uses GPT-4 to evaluate interaction depth, role relevance, and conversation context. Filter High Priority** routes top-tier leads (e.g., decision-makers with strong engagement). Filter Medium Priority** handles warm prospects for lighter follow-up. AI Agent1** generates personalized email content using chat model and memory. Generate Email** creates a professional, context-aware follow-up email. Send Email** delivers the message to the lead’s inbox. AI Agent2** crafts a concise, friendly LinkedIn connection message. Generate LinkedIn Msg** produces a tailored outreach note. Send LinkedIn** posts the message via LinkedIn API. Slack Notification** alerts your team in real-time about high-priority outreach. Update CRM (HubSpot)** adds contact, tags, and follow-up tasks automatically. Save to Database (Insert)** logs full lead journey and message content in PostgreSQL. Generate Analytics** compiles engagement metrics and success rates. Send Response** confirms completion back to the event system. Setup Instructions Import the workflow JSON into n8n Configure credentials: Event Platform API (for attendees & interactions) LinkedIn API (OAuth2) OpenAI (GPT-4) SMTP (for email) or Email Service (SendGrid, etc.) HubSpot API Key PostgreSQL Database Slack Webhook URL Trigger with a webhook POST containing event ID and settings Watch personalized outreach happen automatically! Prerequisites Event platform with webhook + attendee/interaction API LinkedIn Developer App with API access OpenAI API key with GPT-4 access HubSpot account with API enabled PostgreSQL database (table for leads & logs) Slack workspace (optional, for team alerts) Example Webhook Payload { "eventId": "evt_spring2025", "eventName": "Annual Growth Summit", "triggerFollowUp": true, "priorityThreshold": { "high": 75, "medium": 50 } } Modification Options Adjust scoring logic in AI Analyze Profile (e.g., weight Q&A participation higher) Add custom email templates in Generate Email with your brand voice Include meeting booking links (Calendly) in high-priority messages Route VIP leads to Send SMS via Twilio Export analytics to Google Sheets or BI tools (Looker, Tableau) Add approval step before sending LinkedIn messages Ready to 10x your event ROI? Get in touch with us for custom n8n automation!
by Rahul Joshi
Description Turn incoming Gmail messages into structured Zendesk tickets, enriched by Azure OpenAI, and log key details to Google Sheets for tracking. Ideal for IT Support teams needing fast, consistent intake and documentation. ⚡ What This Template Does Fetches new emails via Gmail Trigger. ✉️ Normalizes Gmail data and formats it for downstream steps. Enriches and structures content with Azure OpenAI Chat Model and Output Parsers. Creates Zendesk tickets from the processed data. 🎫 Appends or updates logs in Google Sheets for auditing and reporting. 📊 Key Benefits Saves time by automating ticket creation and logging. ⏱️ Improves ticket quality with AI-driven normalization and structure. Ensures consistent records in Google Sheets for easy reporting. Reduces manual errors in IT Support intake. ✅ Features Gmail-triggered intake flow for new messages. AI enrichment using Azure OpenAI Chat Model with parsing and memory tooling. Zendesk ticket creation (create: ticket) with structured fields. Google Sheets logging (appendOrUpdate: sheet). Modular design with Execute Workflow nodes for reuse and scaling. Requirements n8n instance (Cloud or self-hosted). Gmail credentials configured in n8n for the Gmail Trigger. Zendesk credentials with permission to create tickets. Google Sheets credentials with access to the target spreadsheet (append/update enabled). Azure OpenAI credentials configured for the Azure OpenAI Chat Model and associated parsing. Target Audience IT Support and Helpdesk teams handling email-based requests. 🛠️ Operations teams standardizing inbound email workflows. Agencies and MSPs offering managed support intake. Internal automation teams centralizing ticket capture and logging. Step-by-Step Setup Instructions Connect Gmail credentials in n8n and select the inbox/label for the Gmail Trigger. Add Zendesk credentials and confirm ticket creation permissions. Configure Google Sheets credentials and select the target sheet for logs. Add Azure OpenAI credentials to the Azure OpenAI Chat Model node and verify parsing steps. Import the workflow, assign credentials to each node, update any placeholders, and run a test. Rename the final email/logging nodes descriptively (e.g., “Log to Support Sheet”) and schedule if needed.
by Abdulrahman Alhalabi
NGO TPM Request Management System Benefits For Beneficiaries: 24/7 Accessibility** - Submit requests anytime via familiar Telegram interface Language Flexibility** - Communicate in Arabic through text or voice messages Instant Acknowledgment** - Receive immediate confirmation that requests are logged No Technical Barriers** - Works on basic smartphones without special apps For TPM Teams: Centralized Tracking** - All requests automatically logged with timestamps and user details Smart Prioritization** - AI categorizes issues by urgency and type for efficient response Action Guidance** - Specific recommended actions generated for each request type Performance Analytics** - Track response patterns and common issues over time For NGO Operations: Cost Reduction** - Automated intake reduces manual processing overhead Data Quality** - Standardized categorization ensures consistent reporting Audit Trail** - Complete record of all beneficiary interactions for compliance Scalability** - Handle high volumes without proportional staff increases How it Works Multi-Input Reception - Accepts both text messages and voice recordings via Telegram Voice Transcription - Uses OpenAI Whisper to convert Arabic voice messages to text AI Categorization - GPT-4 analyzes requests and categorizes issues (aid distribution, logistics, etc.) Action Planning - AI generates specific recommended actions for TPM team in Arabic Data Logging - Records all requests, categories, and actions in Google Sheets with user details Confirmation Feedback - Sends acknowledgment message back to users via Telegram Set up Steps Setup Time: ~20 minutes Create Telegram Bot - Get bot token from @BotFather and configure webhook Configure APIs - Set up OpenAI (transcription + chat) and Google Sheets credentials Customize AI Prompts - Adjust system messages for your NGO's specific operations Set Up Spreadsheet - Link Google Sheets for request tracking and reporting Test Workflows - Verify both text and voice message processing paths Detailed Arabic language configuration and TPM-specific categorization examples are included as sticky notes within the workflow. What You'll Need: Telegram Bot Token (free from @BotFather) OpenAI API key (Whisper + GPT-4) Google Sheets API credentials Google Spreadsheet for logging requests Sample Arabic text/voice messages for testing Key Features: Dual input support (text + voice messages) Arabic language processing and responses Structured data extraction (category + recommended action) Complete audit trail with user information Real-time confirmation messaging TPM team-specific workflow optimization
by Cheng Siong Chin
How It Works Scheduled runs collect data from oil markets, global shipping movements, news sources, and official reports. The system performs statistical checks to detect anomalies and volatility shifts. An AI-driven geopolitical model evaluates emerging risks and assigns a crisis score. Based on severity thresholds, results are routed to the appropriate alert channels for rapid response. Setup Steps Data Sources: Connect the oil price API, OPEC report feeds, shipping databases, and news sources. AI Model: Configure the OpenRouter ChatGPT model for geopolitical and risk analysis. Alerts: Define severity rules and route alerts to Email, Slack, or Dashboard APIs. Storage: Configure a database for historical records, audit logging, and trend tracking. Prerequisites Oil market API credentials; news feed access; OPEC data source; OpenRouter API key; Slack/email/dashboard integrations Use Cases Supply chain risk monitoring; energy market crisis detection; geopolitical threat assessment; trader decision support; operational risk management Customization Adjust risk thresholds; add market data sources; modify alert routing rules Benefits Reduces crisis detection lag 90%; consolidates fragmented data; enables proactive response
by n8n Team
This n8n workflow, which runs every Monday at 5:00 AM, initiates a comprehensive process to monitor and analyze network security by scrutinizing IP addresses and their associated ports. It begins by fetching a list of watched IP addresses and expected ports through an HTTP request. Each IP address is then processed in a sequential loop. For every IP, the workflow sends a GET request to Shodan, a renowned search engine for internet-connected devices, to gather detailed information about the IP. It then extracts the data field from Shodan's response, converting it into an array. This array contains information on all ports Shodan has data for regarding the IP. A filter node compares the ports returned from Shodan with the expected list obtained initially. If a port doesn't match the expected list, it is retained for further processing; otherwise, it's filtered out. For each such unexpected port, the workflow assembles data including the IP, hostnames from Shodan, the unexpected port number, service description, and detailed data from Shodan like HTTP status code, date, time, and headers. This collected data is then formatted into an HTML table, which is subsequently converted into Markdown format. Finally, the workflow generates an alert in TheHive, a popular security incident response platform. This alert contains details like the title indicating unexpected ports for the specific IP, a description comprising the Markdown table with Shodan data, medium severity, current date and time, tags, Traffic Light Protocol (TLP) set to Amber, a new status, type as 'Unexpected open port', the source as n8n, a unique source reference combining the IP with the current Unix time, and enabling follow and JSON parameters options. This comprehensive workflow thus aids in the proactive monitoring and management of network security.
by Rahul Joshi
Description Automate your weekly social media analytics with this end-to-end AI reporting workflow. 📊🤖 This system collects real-time Twitter (X) and Facebook metrics, merges and validates data, formats it with JavaScript, generates an AI-powered HTML report via GPT-4o, saves structured insights in Notion, and shares visual summaries via Slack and Gmail. Perfect for marketing teams tracking engagement trends and performance growth. 🚀💬 What This Template Does 1️⃣ Starts manually or on-demand to fetch the latest analytics data. 🕹️ 2️⃣ Retrieves follower, engagement, and post metrics from both X (Twitter) and Facebook APIs. 🐦📘 3️⃣ Merges and validates responses to ensure clean, complete datasets. 🔍 4️⃣ Runs custom JavaScript to normalize and format metrics into a unified JSON structure. 🧩 5️⃣ Uses Azure OpenAI GPT-4o to generate a visually rich HTML performance report with tables, emojis, and insights. 🧠📈 6️⃣ Saves the processed analytics into a Notion “Growth Chart” database for centralized trend tracking. 🗂️ 7️⃣ Sends an email summary report to the marketing team, complete with formatted HTML insights. 📧 8️⃣ Posts a concise Slack update comparing platform performance and engagement deltas. 💬 9️⃣ Logs any validation or API errors automatically into Google Sheets for debugging and traceability. 🧾 Key Benefits ✅ Centralizes all social metrics into a single automated flow. ✅ Delivers AI-generated HTML reports ready for email and dashboard embedding. ✅ Reduces manual tracking with Notion and Slack syncs. ✅ Ensures data reliability with built-in validation and error logging. ✅ Gives instant, visual insights for weekly marketing reviews. Features Multi-platform analytics integration (Twitter X + Facebook Graph API). JavaScript node for dynamic data normalization. Azure OpenAI GPT-4o for HTML report generation. Notion database update for long-term trend storage. Slack and Gmail nodes for instant sharing and communication. Automated error capture to Google Sheets for workflow reliability. Visual, emoji-enhanced reporting with HTML formatting and insights. Requirements Twitter OAuth2 API credentials for access to public metrics. Facebook Graph API access token for page insights. Azure OpenAI API key for GPT-4o report generation. Notion API credentials with write access to “Growth Chart” database. Gmail OAuth2 credentials for report dispatch. Slack Bot Token with chat:write permission for posting analytics summaries. Google Sheets OAuth2 credentials for maintaining the error log. Environment Variables TWITTER_API_KEY FACEBOOK_ACCESS_TOKEN AZURE_OPENAI_API_KEY NOTION_GROWTH_DB_ID GMAIL_REPORT_RECIPIENTS SLACK_REPORT_CHANNEL_ID GOOGLE_SHEET_ERROR_LOG_ID Target Audience 📈 Marketing and growth teams tracking cross-platform performance 💡 Social media managers needing automated reporting 🧠 Data analysts compiling weekly engagement metrics 💬 Digital agencies managing multiple brand accounts 🧾 Operations and analytics teams monitoring performance KPIs Step-by-Step Setup Instructions 1️⃣ Connect all API credentials (Twitter, Facebook, Notion, Gmail, Slack, and Sheets). 2️⃣ Paste your Facebook Page ID and Twitter handle in respective API nodes. 3️⃣ Verify your Azure OpenAI GPT-4o connection and prompt text for HTML report generation. 4️⃣ Update your Notion database structure to match “Growth Chart” property names. 5️⃣ Add your marketing email in the Gmail node and test delivery. 6️⃣ Specify the Slack channel ID where summaries will be posted. 7️⃣ Optionally, connect a Google Sheet tab for error tracking (error_id, message). 8️⃣ Execute the workflow once manually to validate data flow. 9️⃣ Activate or schedule it for weekly or daily analytics automation. ✅
by siyad
This n8n workflow automates the process of monitoring inventory levels for Shopify products, ensuring timely updates and efficient stock management. It is designed to alert users when inventory levels are low or out of stock, integrating with Shopify's webhook system and providing notifications through Discord (can be changed to any messaging platform) with product images and details. Workflow Overview Webhook Node (Shopify Listener): This node is set up to listen for Shopify's inventory level webhook. It triggers the workflow whenever there is an update in the inventory levels. The webhook is configured in Shopify settings, where the n8n URL is specified to receive inventory level updates. Function Node (Inventory Check): This node processes the data received from the Shopify webhook. It extracts the available inventory and the inventory item ID, and determines whether the inventory is low (less than 4 items) or out of stock. Condition Nodes (Inventory Level Check): Two condition nodes follow the function node. One checks if the inventory is low (low_inventory equals true), and the other checks if the inventory is out of stock (out_of_stock equals true). GraphQL Node (Product Details Retrieval): Connected to the condition nodes, this node fetches detailed information about the product using Shopify's GraphQL API. It retrieves the product variant, title, current inventory quantity, and the first product image. HTTP Node (Discord Notification): The final node in the workflow sends a notification to Discord. It includes an embed with the product title, a warning message ("This product is running out of stock!"), the remaining inventory quantity, product variant details, and the product image. The notification ensures that relevant stakeholders are immediately informed about critical inventory levels.
by Yaron Been
This workflow automatically analyzes sales territory performance, comparing revenue, win rates, and activity across regions. Remove the guesswork from territory planning and drive balanced growth. Overview On a weekly schedule, the workflow pulls CRM data for each territory, merges it with demographic and market size info scraped via Bright Data, and feeds everything into OpenAI for performance benchmarking. Outliers—both high and low performers—are highlighted in a Google Data Studio dashboard and summarized in a Slack message. Tools Used n8n** – Orchestrates data collection and analysis CRM API** – Source of sales metrics by territory Bright Data** – Scrapes external market indicators (population, GDP, etc.) OpenAI** – Normalizes and benchmarks territories Google Sheets / Data Studio** – Stores and visualizes results Slack** – Sends the weekly summary How to Install Import the Workflow into n8n. Connect Your CRM API credentials. Configure Bright Data credentials. Set Up OpenAI API key. Authorize Google services & Slack. Customize Territory Definitions in the Set node. Use Cases Sales Leadership**: Rebalance territories based on potential. Revenue Operations**: Identify underserved regions. Financial Planning**: Allocate resources where ROI is highest. Incentive Design**: Reward reps fairly based on potential. Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #territorymanagement #salesanalytics #brightdata #openai #n8nworkflow #nocode #revenueops