by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically monitors local event platforms (Eventbrite, Meetup, Facebook Events) and aggregates upcoming events that match your criteria. Never miss a networking or sponsorship opportunity again. Overview A scheduled trigger scrapes multiple event sites via Bright Data, filtering by location, date range, and keywords. OpenAI classifies each event (conference, meetup, workshop) and extracts key details such as venue, organizers, and ticket price. Updates are posted to Slack and archived in Airtable for quick lookup. Tools Used n8n** – Core automation engine Bright Data** – Reliable multi-site scraping OpenAI** – NLP-based event categorization Slack** – Delivers daily event digests Airtable** – Stores enriched event records How to Install Import the Workflow: Add the .json file to n8n. Configure Bright Data: Provide your account credentials. Set Up OpenAI: Insert your API key. Connect Slack & Airtable: Authorize both services. Customize Filters: Edit the initial Set node to adjust city, radius, and keywords. Use Cases Community Managers**: Curate a calendar of relevant events. Sales Teams**: Identify trade shows and meetups for prospecting. Event Planners**: Track competing events when choosing dates. Marketers**: Spot speaking or sponsorship opportunities. Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #eventmonitoring #brightdata #openscraping #openai #slackalerts #n8nworkflow #nocode #meetup #eventbrite
by Oneclick AI Squad
This automated n8n workflow scrapes job listings from Upwork using Apify, processes and cleans the data, and generates daily email reports with job summaries. The system uses Google Sheets for data storage and keyword management, providing a comprehensive solution for tracking relevant job opportunities and market trends. What is Apify? Apify is a web scraping and automation platform that provides reliable APIs for extracting data from websites like Upwork. It handles the complexities of web scraping including rate limiting, proxy management, and data extraction while maintaining compliance with website terms of service. Good to Know Apify API calls may incur costs based on usage; check Apify pricing for details Google Sheets access must be properly authorized to avoid data sync issues The workflow includes data cleaning and deduplication to ensure high-quality results Email reports provide structured summaries for easy review and decision-making Keyword management through Google Sheets allows for flexible job targeting How It Works The workflow is organized into three main phases: Phase 1: Job Scraping & Initial Processing This phase handles the core data collection and initial storage: Trigger Manual Run - Manually starts the workflow for on-demand job scraping Fetch Keywords from Google Sheet - Reads the list of job-related keywords from the All Keywords sheet Loop Through Keywords - Iterates over each keyword to trigger Apify scraping Trigger Apify Scraper - Sends HTTP request to start Apify actor for job scraping Wait for Apify Completion - Waits for the Apify actor to finish execution Delay Before Dataset Read - Waits a few seconds to ensure dataset is ready for processing Fetch Scraped Job Dataset - Fetches the latest dataset from Apify Process Raw Job Data - Filters jobs posted in the last 24 hours and formats the data Save Jobs to Daily Sheet - Appends new job data to the daily Google Sheet Update Keyword Job Count - Updates job count in the All Keywords summary sheet Phase 2: Data Cleaning & Deduplication This phase ensures data quality and removes duplicates: Load Today's Daily Jobs - Loads all jobs added in today's sheet for processing Remove Duplicates by Title/Desc - Removes duplicates based on title and description matching Save Clean Job Data - Saves the cleaned, unique entries back to the sheet Clear Old Daily Sheet Data - Deletes old or duplicate entries from the sheet Reload Clean Job Data - Loads clean data again after deletion for final processing Phase 3: Daily Summary & Email Report This phase generates summaries and delivers the final report: Generate Keyword Summary Stats - Counts job totals per keyword for analysis Update Summary Sheet - Updates the summary sheet with keyword statistics Fetch Final Summary Data - Reads the summary sheet for reporting purposes Build Email Body - Formats email with statistics and sheet link Send Daily Report Email - Sends the structured daily summary email to recipients Data Sources The workflow utilizes Google Sheets for data management: AI Keywords Sheet - Contains keyword management data with columns: Keyword (text) - Job search terms Job Count (number) - Number of jobs found for each keyword Status (text) - Active/Inactive status Last Updated (timestamp) - When keyword was last processed Daily Jobs Sheet - Contains scraped job data with columns: Job Title (text) - Title of the job posting Description (text) - Job description content Budget (text) - Job budget or hourly rate Client Rating (number) - Client's rating on Upwork Posted Date (timestamp) - When job was posted Job URL (text) - Direct link to the job posting Keyword (text) - Which keyword found this job Scraped At (timestamp) - When data was collected Summary Sheet - Contains daily statistics with columns: Date (date) - Report date Total Jobs (number) - Total jobs found Keywords Processed (number) - Number of keywords searched Top Keyword (text) - Most productive keyword Average Budget (currency) - Average job budget Report Generated (timestamp) - When summary was created How to Use Import the workflow into n8n Configure Apify API credentials and Google Sheets API access Set up email credentials for daily report delivery Create three Google Sheets with the specified column structures Add relevant job keywords to the AI Keywords sheet Test with sample keywords and adjust as needed Requirements Apify API credentials and actor access Google Sheets API access Email service credentials (Gmail, SMTP, etc.) Upwork job search keywords for targeting Customizing This Workflow Modify the Process Raw Job Data node to filter jobs by additional criteria like budget range, client rating, or job type. Adjust the email report format to include more detailed statistics or add visual aids, such as charts. Customize the data cleaning logic to better handle duplicate detection based on your specific requirements, or add additional data sources beyond Upwork for comprehensive job market analysis.
by Jonathan
This workflow uses a Hubspot Trigger to check for new companies. It then checks the companies website exists using the HTTP node. If it doesn't, a message is sent to Slack. To configure this workflow you will need to set the credentials for the Hubspot and Slack Nodes. You will also need to select the Slack channel to use for sending the message.
by Don Jayamaha Jr
A next-generation AI-powered DeFi health monitor that tracks wallet positions across Aave V3 using GPT-4o and LangChain. It delivers human-readable reports via Telegram and Gmail, triggered on schedule or manually. Built for professionals monitoring multiple DeFi wallets. 🧩 System Components | Component | Role | | --------------------------------- | ------------------------------------------------------------- | | ✅ Scheduler | Triggers the workflow periodically | | ✅ Google Sheets Wallet Loader | Loads all monitored wallet addresses | | ✅ Set Variables | Injects dynamic wallet + date | | ✅ AAVE Portfolio AI Agent | GPT-4o + LangChain AI that generates human-readable summaries | | ✅ Moralis API Nodes (3) | Collect Aave V3 supply/borrow/collateral data | | ✅ OpenAI Chat Model (gpt-4o-mini) | Interprets on-chain data and explains it | | ✅ Telegram Delivery | Sends summary to Telegram chat | | ✅ Gmail Email Sender | Sends full HTML report to email | | ✅ HTML Formatter | Beautifies AI output into email structure | ⚙️ How It Works Scheduled or manual trigger Pulls wallet addresses from Google Sheets For each wallet: Pulls Aave data from Moralis GPT-4o AI generates report Sends summary to Telegram Sends full HTML report via Gmail 🛠 Installation Steps 1. Import the Workflow Upload AAVE_Portfolio_Professional_AI_Agent.json to your n8n instance. 2. Connect These Credentials | Service | Required Credential Type | | -------- | ---------------------------- | | Moralis | HTTP Header Auth (X-API-Key) | | OpenAI | GPT-4o via OpenAI API Key | | Telegram | Telegram Bot API Token | | Gmail | Gmail OAuth2 Credential | 3. Create Google Sheet Column name must be: wallet_address Add wallet addresses you want monitored 📬 Output Format Telegram Message Example 📊 Aave DeFi Health Report Wallet: 0xABC...123 Date: 2025-05-21 ▪️ Pool: Aave Ethereum USDC • Supply: $10,040 • Borrowed: $5,500 • Health Factor: 3.43 • Liquidation Risk: No Email Report Full HTML + plain text versions Auto-generated date + styled per wallet Includes notes and threshold warnings 🧠 Smart Features GPT-4o generates clear human summaries Monitors multiple wallets in one run Flags liquidation risk dynamically Logs daily performance snapshots 💡 Customization Ideas Add Telegram slash command /aave <wallet> Expand to monitor Compound, Lido, or Uniswap Export to Notion, Slack, or Data Studio 🧾 Licensing & Attribution © 2025 Treasurium Capital Limited Company Architecture, prompts, and report formatting are intellectual property protected. No unauthorized rebranding, redistribution, or resale permitted. 🔗 For support or licensing inquiries: LinkedIn – Don Jayamaha 🚀 Track all your Aave DeFi positions using AI—delivered via Telegram + Gmail. Perfect for funds, traders, and DeFi power users. 🎥 Watch the Live Demo:
by Abdullah
What it does Automatically respond to Google Form entries submitted via Google Sheets. This workflow notifies your Slack team, sends a personalized Gmail response to the user, and adds the user to Google Contacts — all triggered instantly upon new row addition in your connected Sheet. Who's it for Perfect for lead capture forms, client inquiries, or feedback submissions. Trigger: When a new row is added to a connected Google Sheet (usually linked to a Google Form). Slack Notification: Sends a Slack message to your selected channel with the form data. Gmail Message: Sends an automatic email reply to the submitter (using their email from the form). Add Google Contact: Automatically creates a new contact in Google Contacts using the form data. This setup is ideal for automating client communication and internal team alerts without manual input.
by Nima Salimi
Overview Automate your daily contact imports from NocoDB to Brevo.The workflow updates the record status in NocoDB at each step. For every email campaign, it’s essential to keep your Brevo contact list updated so you can send personalized and targeted emails. This flow automates that process. ✅ Tasks ⏰ Runs automatically every day 🗂 Fetches only new/unimported records from NocoDB 🔍 Checks for missing required fields 🚫 Filters out disposable/temporary emails 📬 Creates contacts in Brevo 📝 Updates NocoDB status after each step 🛠 How to Use 1️⃣ Set your schedule The Schedule Trigger node runs the flow daily adjust to your preferred time. 2️⃣ Prepare your table in NocoDB Your NocoDB table should contain at least: id first_name last_name email status (default: 0-not-imported) 3️⃣ Configure your credentials Connect your NocoDB API Token in the NocoDB nodes. Connect your Brevo API Key in the Brevo node. 4️⃣ Map your fields In the Brevo: Create Contact node, make sure first name, last name, and email match your NocoDB column names. 📌 Notes 🛡 Make sure your NocoDB project/table IDs match the ones in this template. 🚀 This workflow processes contacts one-by-one to avoid heavy API calls and rate limit issues with Brevo. ✅ status values: 0-not-imported → new record 1-empty-fields → missing required fields 2-disposal-email → disposable email detected 3-contact-created → successfully created in Brevo
by n8n Team
This n8n workflow automates the monitoring and notification of Palo Alto Networks security advisories. It is triggered manually from within the n8n UI or scheduled to run daily at midnight using the Schedule Trigger. The workflow begins by fetching the latest security advisories from Palo Alto Networks' RSS feed. Each advisory is then processed, and relevant information is extracted and categorized, including the advisory type, subject, and severity. The workflow checks the publication date of each advisory to ensure that it was posted within the last 24 hours, filtering out older advisories. The workflow then splits into two paths based on the advisory type: GlobalProtect and Traps. In the GlobalProtect path, advisories related to GlobalProtect are identified and used to create Jira issues. The Jira issues include a summary with the advisory title and a description that provides details about the advisory, its severity, link, and publication date. In the Traps path, advisories related to Traps are recognized, and dummy data (which should be replaced with logic to retrieve valid user emails) is generated for sample purposes. These email addresses are then used to send email notifications using the Gmail node. Each email's subject includes the type of advisory, while the body contains the advisory title and a link for more information. Potential issues when setting up this workflow for the first time might involve configuring the Schedule Trigger to match the desired time zone. Additionally, ensuring that the Jira and Gmail nodes are configured correctly with the required credentials and email addresses is crucial. The placeholder for generating dummy data for email recipients should be replaced with logic to retrieve valid user emails. Proper error handling and testing with real and sample advisories can help identify and resolve any potential issues during setup.
by Yaron Been
Description This workflow automatically monitors online forums for specific keywords, topics, or competitor mentions. It helps you stay on top of relevant discussions without manually checking multiple forums throughout the day. Overview This workflow automatically monitors selected forums for new posts, keywords, or competitor activity. It uses Bright Data to scrape forum content and can notify you or save results to a spreadsheet or other service. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping forum content without getting blocked. (Optional) Google Sheets/Discord/Telegram:** For notifications or data storage. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Notifications (Optional): Configure notification/storage nodes as needed. Customize: Specify the forums and keywords to monitor. Use Cases Brand Monitoring:** Stay updated on mentions of your brand or product. Competitor Tracking:** Monitor competitor activity in your industry forums. Community Managers:** Get alerts for new threads or topics. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #forums #brightdata #webscraping #monitoring #forummonitoring #brandmentions #competitortracking #keywordtracking #n8nworkflow #workflow #nocode #communitymanagement #onlineforums #discussionboards #brandawareness #marketresearch #socialmediamonitoring #contentmonitoring #reputationmanagement #digitalmarketing #businessintelligence #onlinediscussions #competitoranalysis
by Oneclick AI Squad
This automated n8n workflow detects and manages fraudulent booking transactions through comprehensive AI-powered analysis and multi-layered security checks. The system processes incoming travel booking data, performs IP geolocation verification, enriches transaction details with AI insights, calculates dynamic risk scores, and executes automated responses based on threat levels. All transactions are logged and appropriate notifications are sent to relevant stakeholders. Good to Know The workflow combines multiple detection methods, including IP geolocation, AI analysis, and risk scoring algorithms Google Gemini Chat Model provides advanced natural language processing for transaction analysis Risk levels are dynamically calculated and categorized as CRITICAL, HIGH, or standard risk Automated blocking and flagging system protects against fraudulent transactions in real-time All transaction data is logged to Google Sheets for audit trails and pattern analysis The system respects API rate limits and includes proper error handling mechanisms How It Works 1. Initial Data Ingestion & Extraction Monitors and captures incoming booking transaction data from various sources Extracts key booking details including user information, payment data, booking location, and transaction metadata Performs initial data validation and formatting for downstream processing 2. IP Geolocation and AI Analysis IP Geolocation Check**: Validates booking IP addresses by checking geolocation details and comparing against expected user locations AI Agent Integration**: Utilizes Google Gemini Chat Model to analyze booking patterns, user behavior, and transaction anomalies Enhanced Data Processing**: Enriches transaction data with geographical context and AI-driven risk indicators 3. Risk Calculation and Decision Logic Enhanced Risk Calculator**: Combines AI-generated risk scores with geolocation-based factors, payment method analysis, and historical patterns Critical Risk Check**: Flags transactions with risk levels marked as CRITICAL for immediate action High Risk Check**: Identifies HIGH risk transactions requiring additional verification steps Dynamic Scoring**: Adjusts risk calculations based on real-time threat intelligence and pattern recognition 4. Action & Notification Block User Account**: Automatically blocks user accounts for CRITICAL risk transactions to prevent immediate fraud Flag for Review**: Marks HIGH risk transactions for manual review by fraud prevention teams Send Notifications**: Dispatches real-time alerts via email and messaging systems to security teams Automated Responses**: Sends appropriate messages to users based on transaction status and risk level 5. Logging & Response Log to Google Sheets**: Records all transaction details, risk scores, and actions taken for comprehensive audit trails Flag for Review**: Maintains detailed logs of flagged transactions for pattern analysis and machine learning improvements Response Tracking**: Monitors and logs all automated responses and manual interventions How to Use Import the workflow into your n8n instance Configure Google Gemini Chat Model API credentials for AI analysis Set up IP geolocation service API access for location verification Configure Google Sheets integration for transaction logging Establish Gmail/email credentials for notification delivery Define risk thresholds and scoring parameters based on your fraud tolerance levels Test the workflow with sample booking data to verify all components function correctly Monitor initial deployments closely to fine-tune risk scoring algorithms Establish manual review processes for flagged transactions Set up regular monitoring and maintenance schedules for optimal performance Requirements Google Gemini Chat Model API access IP Geolocation service API credentials Google Sheets API integration Gmail API or SMTP email service for notifications n8n instance with appropriate node modules installed Customizing This Workflow Risk Scoring Parameters**: Adjust risk calculation algorithms and thresholds based on your specific fraud patterns and business requirements AI Model Configuration**: Fine-tune Google Gemini prompts and analysis parameters for improved accuracy in your use case Notification Channels**: Add or modify notification methods including Slack, SMS, or webhook integrations Data Sources**: Extend input methods to accommodate additional booking platforms or payment processors Logging Destinations**: Configure alternative or additional logging systems such as databases or external SIEM platforms Geographic Rules**: Customize geolocation validation rules based on your service areas and customer base Automated Actions**: Modify or expand automated response actions based on your fraud prevention policies Review Workflows**: Integrate with existing fraud review systems or ticketing platforms for seamless manual review processes
by Airtop
About this AI Agent This workflow is designed to automate web interactions by simulating a human user, using a combination of the Agent node and AI tools powered by Airtop. How does this workflow works? Form Submission Trigger: The workflow starts with a form submission trigger node named "On form submission". This node collects user instructions for the web AI agent, including a prompt and an optional Airtop profile name for sites requiring authentication. AI Agent: The core of the workflow is the "AI Agent" node, which uses a smart web agent to manage a remote web browser. It is designed to fulfill user requests by interacting with the browser through various tools. Browser Session Management Start Browser: The "Start browser" node initiates a new browser session and window. It is essential for obtaining the sessionId and windowId required for subsequent operations. Session and Window Management: The workflow includes nodes for creating and managing browser sessions and windows, such as "Session" and "Window". Web Interaction Tools: Load URL: This node loads a specified URL into the browser window. Query: The "Query" node allows the agent to ask questions and extract information from the current web page. Click: This node simulates clicking on elements within the web page. Type: The "Type" node types text into specified elements on the page. Session Termination: The "End session" node is used to terminate the browser session once the tasks are completed. Output Handling Structured Output Parser: This node processes the agent's results into a structured format. Output: The final results are set and prepared for output. Slack Integration: Although currently disabled, there is a "Slack" node intended to send messages to a Slack channel, potentially for notifications or live view URLs. Seting up your agent Airtop API Credentials: Users must have valid Airtop API credentials to interact with the web browser tools. This includes nodes like "Click", "Query", "Load URL", "End session", "Type", "Session", and "Window". Slack API Credentials (Optional): If users want to enable Slack notifications, they need to configure Slack OAuth2 credentials. The Slack node is currently disabled but can be used to send messages to a Slack channel. Anthropic API Credentials: The "Claude 3.5 Haiku" node requires Anthropic API credentials. Users need to have access to this API to utilize the language model features.
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically tracks individual sales rep performance—calls, emails, meetings, quota attainment—and surfaces coaching insights. Free your managers from manual report building and focus on improvement. Overview On a daily schedule, the workflow queries CRM and telephony/email logs, aggregating activity metrics per rep. OpenAI analyzes patterns, flags underperformance or standout achievements, and suggests tailored coaching tips. Results are delivered as a nicely formatted Slack message and stored in Airtable. Tools Used n8n** – Automation backbone CRM + Telephony APIs** – Provide activity data OpenAI** – Generates insights and coaching tips Slack** – Sends manager digest Airtable** – Maintains historical performance records How to Install Import the Workflow into n8n. Connect Data Sources: Add CRM, VoIP, and email API keys. Set Up OpenAI: Enter your API key. Authorize Slack & Airtable. Customize Metrics: Modify the aggregation node to focus on your KPIs. Use Cases Sales Coaching**: Provide reps with daily feedback. Performance Management**: Quickly identify top and low performers. Incentive Programs**: Track achievements for rewards. Revenue Operations**: Unify activity data into one source. Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #salesperformance #openai #salescoaching #n8nworkflow #nocode #revenueops
by David Ashby
What it is- I wanted to create a simple, easy-to-use, MCP server for your Discord bot(s). How to set up- Literally all you do is select your bot auth (or crease a new Discord Bot auth if you havn't entered your key in n8n before) and that's IT! How to use it- You can now ask your bot to do things via any MCP client, including from within N8N workflows! Note: If you need an example, you can check out my simple quickstart Discord MCP Server that uses 4o to send messages to channels on your server and users who are members of the server the bot is in. Need help? Want access to more workflows and even live Q&A sessions with a top verified n8n creator.. All 100% free? Join the community