by Łukasz
What is This? This automation simulates Scrum Master role on daily meetings. Essentially it is an AI Scrum Master using different sources of data. As intelligent support system for Scrum Masters that leverages data from Asana, Slack, and direct developer responses for comprehensive sprint status analysis and identification of areas requiring intervention. As such it is usable for Scrum Masters (of course) but Scrum Team aswell, Product Owner and possibly Business Owner. Who is it For? This automation is designed for Agile teams to support the Scrum Master role by collecting and analyzing data from various sources to identify potential impediments and support the team in sprint delivery. How Does It Work? The workflow has four main data entry points, that are launched either on-click or on workdays. First is collecting project section information from Asana. The automation retrieves project structure, available sections, and their organization, allowing the AI to understand the team's work context. Second is getting recently modified tasks in the Asana project. The system tracks changes in tasks, their status, assignments, and updates to detect potential delays or issues. Third is obtaining communication in the team's Slack channel. The flow collects data about recent conversations, discussion threads, and team communication to identify warning signals or areas requiring attention. Fourth is directly collecting responses from developers about the current sprint - their progress, impediments, concerns, and support needs. All collected data is passed to an AI model that analyzes it within the Scrum methodology context and identifies: Potential impediments in sprint delivery Areas requiring Scrum Master intervention Recommendations for team support Warning signals regarding Sprint Goal achievement Output is being pushed to Slack channel so it can be potentially used by another iteration of same flow itself via Slack channel history. Requirements You need Asana oAuth credentials You need OpenAI / alternative AI for processing data You need to have Slack app with proper permissions channels:history chat:write groups:history im:history mpim:history users.profile:write users:write Configuration Set up node "Asana Project and Slack Channel". Provide Asana project ID and Slack Channel ID (optional) Set up node "Get Scrum Master Answers". There are daily questions/answers that are being sent to channel. Alternative use You can get rid of the whole "Ask Users Daily ScrumMaster Questions" part if you don't want to do it simirarly as "daily Scrum standups". In such case whole flow is essentially changed to static analyzer of project status based on Slack and Asana. Extensions and Customizations There are many possibilities to extend this automation depending on team needs. For example, you can add integration with additional project management tools, implement different notification schemes based on detected issue criticality, or adjust data collection frequency to match the team's work rhythm. Disclaimers and Notes Whole automation has one important assumption: project is run on single Slack channel and on single Asana board. Of cource this can be extended, but is beyond currently designed scope. Adding new sources for AI to analyze should be fairly easy - just add another branch of data and push it to AI prompt. This automation represents a proof-of-concept and should not replace an actual Scrum Master. The Scrum Master role extends far beyond data collection and analysis - it requires deep understanding of team dynamics, business context, and interpersonal skills. As Scrum.org emphasizes, the Scrum Master doesn't need to be present during Daily Scrum, and their role is to ensure the meeting happens, but developers are responsible for conducting the meeting. Mindlessly executing daily questions without proper context analysis can lead to situations where the Scrum Master becomes a team manager instead of a self-organization facilitator. A real Scrum Master analyzes much more data than what's collected by automation - they observe team dynamics, understand business context, identify deeper root causes of problems, and support the team in developing self-organization skills. AI can be a valuable support tool, but it cannot replace the human intuition, empathy, and experience essential in this role. The automation should be treated as a tool supporting the Teams's work, providing additional insights and helping identify areas requiring attention, but always under the supervision and interpretation of an experienced Scrum practitioner.
by Harshil Agrawal
This is a workflow that sends daily astronomy picture of the day using the NASA node to a channel on Telegram. Cron node: The Cron node triggers the workflow daily at 8 PM. You can update the time in the Cron node to trigger the workflow at your desired time. NASA node: After the Cron node triggers the workflow, the NASA node fetches the Astronomy Picture of the Day from the NASA API. You can also get the binary file of the image. Toggle Download Image to true to get the file. Telegram node: The Telegram node sends the image to a Telegram channel. If you want to share the image on another platform, you can replace the Telegram node with the node of that platform. For example, if you want to post the image on a channel on Slack, replace the Telegram node with the Slack node. You can learn to build this workflow on the documentation page of the NASA node.
by Obsidi8n
How it Works This n8n template makes it possible to send emails directly from your Obsidian notes. It leverages the power of the Obsidian Post Webhook plugin, allowing seamless integration between your notes and the email workflow. What it does: Receives note content and metadata from Obsidian via a Webhook. Parses YAML frontmatter to define email recipients, subject, and more. Automatically processes attachments, encoding them into an email-friendly format. Sends emails via Gmail and confirms the status back to Obsidian. Includes a testing feature to verify everything works before going live. Set-up Steps Webhook Configuration: Set your n8n POST Webhook URL in the Obsidian Obsidian Post Webhook plugin settings. Email Integration: Submit the Gmail credentials in n8n email nodes. Test the Workflow: Run a test from Obsidian to ensure the template functions correctly. Activate and Enjoy: Start sending customized emails with attachments from your notes in no time!
by tanaypant
This workflow automatically queries a Postgres database to find outlier readings for which SMS notifications have not been sent. This is Workflow 2 in the blog tutorial Database activity monitoring and alerting. Prerequisites A Postgres database set up and credentials A Twilio account and credentials Nodes Cron node triggers the workflow every minute, so the database is queried at regular intervals. Postgres nodes extract values from, and update values in the database. Twilio node sends an alert SMS about the outlier reading to a specified phone number. Set node sets the notification value to true.
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically monitors local event platforms (Eventbrite, Meetup, Facebook Events) and aggregates upcoming events that match your criteria. Never miss a networking or sponsorship opportunity again. Overview A scheduled trigger scrapes multiple event sites via Bright Data, filtering by location, date range, and keywords. OpenAI classifies each event (conference, meetup, workshop) and extracts key details such as venue, organizers, and ticket price. Updates are posted to Slack and archived in Airtable for quick lookup. Tools Used n8n** – Core automation engine Bright Data** – Reliable multi-site scraping OpenAI** – NLP-based event categorization Slack** – Delivers daily event digests Airtable** – Stores enriched event records How to Install Import the Workflow: Add the .json file to n8n. Configure Bright Data: Provide your account credentials. Set Up OpenAI: Insert your API key. Connect Slack & Airtable: Authorize both services. Customize Filters: Edit the initial Set node to adjust city, radius, and keywords. Use Cases Community Managers**: Curate a calendar of relevant events. Sales Teams**: Identify trade shows and meetups for prospecting. Event Planners**: Track competing events when choosing dates. Marketers**: Spot speaking or sponsorship opportunities. Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #eventmonitoring #brightdata #openscraping #openai #slackalerts #n8nworkflow #nocode #meetup #eventbrite
by Oneclick AI Squad
This comprehensive n8n workflow automates the entire travel business call management process, from initial customer inquiries to trip bookings and marketing outreach. The system handles incoming calls, validates trip details, processes bookings, captures leads, and manages outbound marketing campaigns to promote trip organizer services. It streamlines the complete sales cycle while maintaining organized data records for business intelligence. Essential Information The system operates across four distinct workflows to handle different aspects of travel call management. All call data is automatically captured and stored in organized spreadsheets for analysis and follow-up. The workflow validates trip details before processing to ensure data accuracy and prevent booking errors. Outbound marketing campaigns are automatically triggered based on lead detection and formatting. System Architecture Call Handling Pipeline**: The Detect Incoming Call node captures all incoming customer calls, followed by the Validate Trip Details node which verifies and processes trip information, and the Deliver Organizer Info node that provides relevant trip organizer details to callers. Booking Management Flow**: The Capture Voice Input node records customer booking requests, the Update Booking Record node processes and stores booking information, and the Send Booking Confirmation node delivers confirmation details to customers. Lead Generation Process**: The Detect New Lead node identifies potential customers from call data, the Format Lead Information node structures the lead data for marketing use, and the Initiate Marketing Outreach node launches targeted marketing campaigns. Data Management System**: The Receive Call Response node collects call interaction data, the Log User Input node records customer information in spreadsheets, and the Relay Response to System node ensures data synchronization across all components. Implementation Guide Import the workflow into n8n and configure phone system integration for call detection and voice capture. Set up spreadsheet connections for booking records, lead management, and call logging. Configure marketing automation tools for outbound campaign management. Test each workflow section independently before enabling the complete system. Monitor call handling accuracy and adjust validation rules as needed. Technical Dependencies Phone system API or telephony service for call detection and voice processing Spreadsheet service (Google Sheets, Excel Online) for data storage and management Marketing automation platform for outbound campaign execution Voice recognition service for capturing and processing customer input CRM integration for lead management and customer tracking Database & Sheet Structure Call Tracking Sheet**: Columns should include Call_ID, Customer_Phone, Call_Time, Call_Duration, Call_Status, Trip_Interest, Organizer_Assigned Booking Records Sheet**: Required columns are Booking_ID, Customer_Name, Customer_Phone, Destination, Travel_Dates, Group_Size, Booking_Status, Confirmation_Sent Lead Management Sheet**: Essential columns include Lead_ID, Customer_Name, Phone_Number, Email, Trip_Preference, Lead_Source, Lead_Status, Marketing_Campaign_Sent Trip Organizer Database**: Contains Organizer_ID, Organizer_Name, Specialization, Contact_Info, Availability_Status, Performance_Rating Marketing Outreach Log**: Tracks Campaign_ID, Lead_ID, Campaign_Type, Send_Date, Response_Status, Follow_up_Required Customization Possibilities Adjust the Validate Trip Details node to include specific travel validation rules or partner requirements. Modify the Format Lead Information node to match your CRM system's data structure and marketing campaign formats. Configure the Initiate Marketing Outreach node to integrate with your preferred marketing platforms and campaign templates. Customize the data logging structure in the Log User Input node to capture additional customer information or booking details. Add additional validation steps or approval workflows between booking capture and confirmation sending.
by Brian Burnett
Basics Provides a mechanism to save all your workflows into a github repository and path (of your choosing). These can then be shared through your entire org and used to track changes (if you make any sad 'oopsies'. Flow Obtains and creates listing of currently configured workflows. Iterates through each workflow looking at the following Github source (if present) Actual workflow code (from N8N) Workflow code is sorted and compared for any changes If changed (or new) the workflows are saved / archived into github. Configuration Most of the configuration is done in the Globals node which houses the repo detail for github nodes. The only other dependency is that it by default looks for a GitHub credential, if you use something other than that precise wording you will need to change the credential used on the respective nodes. We gave it 'Manage' rights, but that was only so that it was able to override a requirement for checks to complete? Most would probably only need 'Write' privileges. Background Well, so we initially started using N8N just as a kubernetes-based service housed with its DB running inside the pod. Worked great for getting to know N8N and we jut kept all our workflows and credentials listed in a readme. Fast forward about a year... We have migrated this into our 'production' toolsets and maintain a bunch of team worflows inside it (not company-wide, but LOTS of team fun). While trying to spin a copy of our production RDS database, the ++actual++ production database was deleted, and in doing so AWS was nice enough to wipe our snapshots too!! Yea! Thankfully it only took us a few hours to get everything back up and running thanks to this, so I'm sharing it for everyone to benefit. We have used it to restore old workflows, changes, and now to test our full DR proceedures! (Ok, I might have taken that a bit far)
by Oneclick AI Squad
This automated n8n workflow scrapes job listings from Upwork using Apify, processes and cleans the data, and generates daily email reports with job summaries. The system uses Google Sheets for data storage and keyword management, providing a comprehensive solution for tracking relevant job opportunities and market trends. What is Apify? Apify is a web scraping and automation platform that provides reliable APIs for extracting data from websites like Upwork. It handles the complexities of web scraping including rate limiting, proxy management, and data extraction while maintaining compliance with website terms of service. Good to Know Apify API calls may incur costs based on usage; check Apify pricing for details Google Sheets access must be properly authorized to avoid data sync issues The workflow includes data cleaning and deduplication to ensure high-quality results Email reports provide structured summaries for easy review and decision-making Keyword management through Google Sheets allows for flexible job targeting How It Works The workflow is organized into three main phases: Phase 1: Job Scraping & Initial Processing This phase handles the core data collection and initial storage: Trigger Manual Run - Manually starts the workflow for on-demand job scraping Fetch Keywords from Google Sheet - Reads the list of job-related keywords from the All Keywords sheet Loop Through Keywords - Iterates over each keyword to trigger Apify scraping Trigger Apify Scraper - Sends HTTP request to start Apify actor for job scraping Wait for Apify Completion - Waits for the Apify actor to finish execution Delay Before Dataset Read - Waits a few seconds to ensure dataset is ready for processing Fetch Scraped Job Dataset - Fetches the latest dataset from Apify Process Raw Job Data - Filters jobs posted in the last 24 hours and formats the data Save Jobs to Daily Sheet - Appends new job data to the daily Google Sheet Update Keyword Job Count - Updates job count in the All Keywords summary sheet Phase 2: Data Cleaning & Deduplication This phase ensures data quality and removes duplicates: Load Today's Daily Jobs - Loads all jobs added in today's sheet for processing Remove Duplicates by Title/Desc - Removes duplicates based on title and description matching Save Clean Job Data - Saves the cleaned, unique entries back to the sheet Clear Old Daily Sheet Data - Deletes old or duplicate entries from the sheet Reload Clean Job Data - Loads clean data again after deletion for final processing Phase 3: Daily Summary & Email Report This phase generates summaries and delivers the final report: Generate Keyword Summary Stats - Counts job totals per keyword for analysis Update Summary Sheet - Updates the summary sheet with keyword statistics Fetch Final Summary Data - Reads the summary sheet for reporting purposes Build Email Body - Formats email with statistics and sheet link Send Daily Report Email - Sends the structured daily summary email to recipients Data Sources The workflow utilizes Google Sheets for data management: AI Keywords Sheet - Contains keyword management data with columns: Keyword (text) - Job search terms Job Count (number) - Number of jobs found for each keyword Status (text) - Active/Inactive status Last Updated (timestamp) - When keyword was last processed Daily Jobs Sheet - Contains scraped job data with columns: Job Title (text) - Title of the job posting Description (text) - Job description content Budget (text) - Job budget or hourly rate Client Rating (number) - Client's rating on Upwork Posted Date (timestamp) - When job was posted Job URL (text) - Direct link to the job posting Keyword (text) - Which keyword found this job Scraped At (timestamp) - When data was collected Summary Sheet - Contains daily statistics with columns: Date (date) - Report date Total Jobs (number) - Total jobs found Keywords Processed (number) - Number of keywords searched Top Keyword (text) - Most productive keyword Average Budget (currency) - Average job budget Report Generated (timestamp) - When summary was created How to Use Import the workflow into n8n Configure Apify API credentials and Google Sheets API access Set up email credentials for daily report delivery Create three Google Sheets with the specified column structures Add relevant job keywords to the AI Keywords sheet Test with sample keywords and adjust as needed Requirements Apify API credentials and actor access Google Sheets API access Email service credentials (Gmail, SMTP, etc.) Upwork job search keywords for targeting Customizing This Workflow Modify the Process Raw Job Data node to filter jobs by additional criteria like budget range, client rating, or job type. Adjust the email report format to include more detailed statistics or add visual aids, such as charts. Customize the data cleaning logic to better handle duplicate detection based on your specific requirements, or add additional data sources beyond Upwork for comprehensive job market analysis.
by WeblineIndia
Facebook Group Auto-Moderation This workflow automatically monitors Facebook Group posts, analyzes them using AI, detects policy violations, logs incidents, notifies moderators and automatically hides high-severity posts to keep the community clean and safe. This workflow listens to new Facebook Group posts in real time, processes each post individually and sends the post content to AI for moderation.If a post violates group rules (spam, scam, hate, adult content or aggressive promotion), the workflow alerts moderators, stores the violation in Airtable and automatically hides the post if the severity is high. You receive: Real-time AI moderation of Facebook Group posts** Automatic hiding of high-risk content** Slack alerts for moderation actions** Airtable logging for audit and tracking** Ideal for Facebook Group admins who want fast, consistent and automated moderation without manual review of every post. Quick Start – Implementation Steps Connect the Facebook Group Webhook to your n8n instance. Add your Facebook Page Access Token as an environment variable. Connect OpenAI credentials for content moderation. Configure Slack for alerts and Airtable for logging. Test using sample Facebook post data. Activate the workflow. What It Does This workflow automates Facebook Group moderation: Receives new group posts via webhook. Splits and processes posts one by one. Normalizes post data (ID, message, user, time). Sends post content to AI for moderation analysis. Determines: Violation or not Category (spam, scam, hate, adult, etc.) Severity (low / medium / high) Logs violations into Airtable. Sends alerts to Slack. Automatically hides posts marked as high severity. Notifies the team whether auto-hide succeeded or failed. Who’s It For This workflow is ideal for: Facebook Group admins & moderators Community management teams Social media operations teams Platforms handling large group volumes Anyone needing automated moderation at scale Requirements to Use This Workflow To run this workflow, you need: n8n instance** (cloud or self-hosted) Facebook Group Webhook subscription** Facebook Page Access Token** OpenAI API key** Slack workspace** with API access Airtable base** + Personal Access Token How It Works Receive Facebook Post – Webhook captures new group posts. Process Posts – Posts are handled one at a time. Normalize Data – Extracts clean post and user details. AI Moderation – AI analyzes the post for rule violations. Violation Check – Determines whether action is needed. Severity Check – Only high-risk posts are auto-hidden. Hide Post – Facebook API hides the post automatically. Log & Notify – Slack alerts + Airtable records are created. Setup Steps Import the workflow JSON into n8n. Configure the Webhook node and subscribe it to your Facebook Group. Add FB\_PAGE\_ACCESS\_TOKEN in n8n environment variables. Connect OpenAI, Slack and Airtable credentials. Verify Airtable field names match the workflow mapping. Test using pinned sample data. Activate the workflow. How To Customize Nodes Customize Moderation Rules Edit the AI Content Moderation node to: Adjust strictness Add or remove categories Change severity logic Customize Slack Alerts You can add: Emojis Mentions (@channel / @here) Direct links to the Facebook post Customize Auto-Hide Logic Change the Severity High? IF node to: Auto-hide medium severity Disable auto-hide completely Add manual approval steps Add-Ons (Optional Enhancements) You can extend this workflow to: Add moderator approval before hiding posts Auto-ban repeat offenders Track user violation history Generate daily moderation summaries Add sentiment analysis Create dashboards using Airtable Interfaces Support multiple Facebook Groups Use Case Examples 1\. Spam Control Automatically hide promotional or scam posts. 2\. Community Safety Detect hate or adult content instantly. 3\. Moderator Efficiency Reduce manual review workload. 4\. Audit & Compliance Maintain a clear violation history in Airtable. 5\. Large Group Management Scale moderation without adding moderators. Troubleshooting Guide | Issue | Possible Cause | Solution | |----------------------|----------------------------|-------------------------------------------------------| | No posts received | Webhook not subscribed | Verify Facebook webhook setup | | AI result missing | OpenAI error | Check API key & rate limits | | Post not hidden | Token permission issue | Verify Page Access Token permissions | | Slack alert not sent | Invalid Slack credentials | Reconnect Slack API | | Airtable error | Field mismatch | Match Airtable column names exactly | Need Help? If you need help extending this workflow with multi-group moderation, advanced AI rules, dashboards or production-scale automation, our n8n automation experts at WeblineIndia can assist with custom workflow design and deployment.
by Lorena
This workflow synchronizes data both ways between Pipedrive and HubSpot. Cron node** schedules the workflow to run every minute. Pipedrive* and *Hubspot nodes** pull in both lists of persons from Pipedrive and contacts from HubSpot. Merge1* and *Merge2 nodes** with the option Remove Key Matches identify the items that uniquely exist in HubSpot and Pipedrive, respectively. Update Pipedrive* and *Update HubSpot nodes** take those unique items and add them in Pipedrive and HubSpot, respectively.
by Lucas Perret
This workflow monitor G2 reviews URLS. When a new review is published, it will: trigger a Slack notification record the review in Google Sheets To install it, you'll need: access to Slack, Google Sheets and ScrapingBee Full guide here: https://lempire.notion.site/Scrape-G2-reviews-with-n8n-3f46e280e8f24a68b3797f98d2fba433?pvs=4
by Jonathan
This workflow uses a Hubspot Trigger to check for new companies. It then checks the companies website exists using the HTTP node. If it doesn't, a message is sent to Slack. To configure this workflow you will need to set the credentials for the Hubspot and Slack Nodes. You will also need to select the Slack channel to use for sending the message.