by lin@davoy.tech
This workflow template, "n8n Error Report to LINE," is designed to streamline error handling by sending real-time notifications to your LINE account whenever an error occurs in any of your n8n workflows. By integrating with the LINE Messaging API , this template ensures you stay informed about workflow failures, allowing you to take immediate action and minimize downtime. Whether you're a developer managing multiple workflows or a business owner relying on automation, this template provides a simple yet powerful way to monitor and resolve errors efficiently. Who Is This Template For? Developers: Who manage complex n8n workflows and need real-time error notifications. DevOps Teams: Looking to enhance monitoring and incident response for automated systems. Business Owners: Who rely on n8n workflows for critical operations and want to ensure reliability. Automation Enthusiasts: Seeking tools to simplify error tracking and improve workflow performance. What Problem Does This Workflow Solve? When automating processes with n8n, errors can occur due to various reasons such as misconfigurations, API changes, or unexpected inputs. Without proper error handling, these issues may go unnoticed, leading to delays or disruptions. This workflow solves that problem by: 1) Automatically detecting errors in your n8n workflows. 2) Sending instant notifications to your LINE account with details about the failed workflow, including its name and execution URL. Allowing you to quickly identify and resolve issues, ensuring smooth operation of your automated systems. What This Workflow Does 1) Error Trigger: The workflow is triggered automatically whenever an error occurs in any n8n workflow configured to use this error-handling flow. 2) Send Notification via LINE: Using the LINE Push API , the workflow sends a message to your LINE account with key details about the error, such as the workflow name and execution URL. You can also customize the notification message to include additional information or format it to suit your preferences. Setup Guide Pre-Requisites Access to the LINE Developers Console with a registered bot and access to the Push API. https://developers.line.biz/console/ [API Reference]( https://developers.line.biz/en/reference/messaging-api/#send-narrowcast-message) Basic knowledge of n8n workflows and JSON formatting. An active n8n instance where you can configure error workflows. Step-by-Step Setup Configure the Error Trigger: Set this workflow as the default error workflow in your n8n instance. https://docs.n8n.io/flow-logic/error-handling/ Set Up LINE Push API: Replace <UID HERE> in the HTTP Request node with your LINE user ID to ensure notifications are sent to the correct account.
by Davide
How It Works This workflow creates an AI-powered Telegram chatbot with session management, allowing users to: Start new conversations** (/new). Check current sessions** (/current). Resume past sessions** (/resume). Get summaries** (/summary). Ask questions** (/question). Key Components: Session Management**: Uses Google Sheets to track active/expired sessions (storing SESSION IDs and STATE). /new creates a session; /resume reactivates past ones. AI Processing**: OpenAI GPT-4 generates responses with contextual memory (via Simple Memory node). Summarization: Condenses past conversations when requested. Data Logging**: All interactions (prompts/responses) are saved to Google Sheets for audit and retrieval. User Interaction**: Telegram commands trigger specific actions (e.g., /question [query] fetches answers from session history). Main Advantages 1. Multi-session Handling Each user can create, manage, and switch between multiple sessions independently, perfect for organizing different conversations without confusion. 2. Persistent Memory Conversations are stored in Google Sheets, ensuring that chat history and session states are preserved even if the server or n8n instance restarts. 3. Commands for Full Control With intuitive commands like /new, /current, /resume, /summary, and /question, users can manage sessions easily without needing a web interface. 4. Smart Summarization and Q&A Thanks to OpenAI models, the workflow can summarize entire conversations or answer specific questions about past discussions, saving time and improving the chatbot’s usability. 5. Easy Setup and Scalability By using Google Sheets instead of a database, the workflow is easy to clone, modify, and deploy — ideal for quick prototyping or lightweight production uses. 6. Modular and Extensible Each logical block (new session, get current session, resume, summarize, ask question) is modular, making it easy to extend the workflow with additional features like analytics, personalized greetings, or integrations with CRM systems. Setup Steps Prerequisites: Telegram Bot Token**: Create via BotFather. Google Sheets**: Duplicate this template. Two sheets: Session (active/inactive sessions) and Database (chat logs). OpenAI API Key**: For GPT-4 responses. Configuration: Telegram Integration: Add your bot token to the Telegram Trigger and Telegram Send nodes. Google Sheets Setup: Authenticate the Google Sheets nodes with OAuth. Ensure sheet names (Session, Database) and column mappings match the template. OpenAI & Memory: Add your API key to the OpenAI Chat Model nodes. Adjust contextWindowLength in the Simple Memory node for conversation history length. Testing: Use Telegram commands to test: /new: Starts a session. /question [query]: Tests AI responses. /summary: Checks summarization. Deployment: Activate the workflow; the bot will respond to Telegram messages in real-time. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Mark Shcherbakov
Video Guide I prepared a detailed guide that showed the whole process of integrating the Binance API and storing data in Airtable to manage funding statements associated with tokens in a wallet. Youtube Link Who is this for? This workflow is ideal for developers, financial analysts, and cryptocurrency enthusiasts who want to automate the process of managing funding statements and token prices. It’s particularly useful for those who need a systematic approach to track and report funding fees associated with tokens in their wallets. What problem does this workflow solve? Managing funding statements and token prices across multiple platforms can be cumbersome and error-prone. This workflow automates the process, allowing users to seamlessly fetch funding fees from Binance and record them alongside token prices in Airtable, minimizing manual data entry and potential discrepancies. What this workflow does This workflow integrates the Binance API with an Airtable database, facilitating the storage and management of funding statements linked to tokens in a wallet. The agent can: Fetch funding fees and current positions from Binance. Aggregate data to create structured funding statements. Insert records into Airtable, ensuring proper linkage between funding data and tokens. API Authentication: The workflow establishes authentication with the Binance API using a Crypto Node to handle API keys and signatures, ensuring secure and verified requests. Data Collection: It retrieves necessary data, including funding fees and current positions with properly formatted API requests to ensure seamless communication with Binance. Airtable Integration: The workflow inserts aggregated funding statements and token data into the corresponding Airtable records, managing token existence checks to avoid duplicate entries. Setup Set Up Airtable Database: Create an Airtable base with tables for Funding Statements and Tokens. Generate Binance API Key: Log in and create an API key with appropriate permissions. Set Up Authentication in N8N: Utilize a Crypto Node for Binance API authentication. Configure API Request to Binance: Set request method and headers for communication with the Binance API. Fetch Funding Fees and Current Positions: Retrieve funding data and current positions efficiently. Aggregate and Create Statements: Aggregate data to create detailed funding statements. Insert Data into Airtable: Input the structured data into Airtable and manage token records. Using Get Price Node: Implement a Get Price Node to maintain current token price tracking without additional setup.
by Gleb D
This n8n workflow automates the enrichment of a company list by discovering and extracting each company’s official LinkedIn URL using Bright Data’s search capabilities and Google Gemini AI for HTML parsing and result interpretation. Who is this template for? This workflow is ideal for sales, business development, and data research professionals who need to collect official LinkedIn company profiles for multiple organizations, starting from a list of company names in Google Sheets. It’s especially useful for teams who want to automate sourcing LinkedIn URLs, enrich their prospect database, or validate company data at scale. How it works Manual Trigger: The workflow is started manually (useful for controlled batch runs and testing). Read Company Names: Company names are loaded from a specified Google Sheets table. Loop Over Each Company: Each company is processed one-by-one: A custom Google Search URL is generated for each name. A Bright Data Web Unlocker request is sent to fetch Google search results for “site:linkedin.com [company name]”. Parse LinkedIn Profile URL Using AI: Google Gemini (or your specified LLM) analyzes the fetched search page and extracts the most likely official LinkedIn company profile. Result Handling: If a profile is found, it’s stored in the results. If not, an empty result is created, but you can add custom logic (notifications, retries, etc.). Batch Data Enrichment: All found company URLs are bundled into a single request for further enrichment from a Bright Data dataset. Export: The workflow appends the final, structured data for each company to another sheet in your Google Sheets file. Setup instructions 1. Replace API Keys: Insert your Bright Data API key in these nodes: Bright Data Web Request - Google Search for Company LinkedIn URL HTTP Request - Post API call to Bright Data Snapshot Progress HTTP Request - Getting data from Bright Data 2. Connect Google Sheets: Set up your Google Sheets credentials and specify the sheet for reading input and writing output. 3. Customize Output Structure: Adjust the Python code node (see sticky note in the template) if you want to include additional or fewer fields in your output. 4. Adjust for Scale or Error Handling: You can modify the logic for “not found” results (e.g., to notify a Slack channel or retry failed companies). 5. Run the Workflow: Start manually, monitor the run, and check your Google Sheet for results. Customization guidance Change Input/Output Sheets: Update the sheet names or columns if your source/target spreadsheet has a different structure. Use Another AI Model: Replace the Google Gemini node with another LLM node if preferred. Integrate Alerts: Add Slack or email nodes to notify your team when a LinkedIn profile is not found or when the process is complete.
by Jimleuk
This n8n template lets you summarize team member activity on Slack for the past week and generates a report. For remote teams, chat is a crucial communication tool to ensure work gets done but with so many conversations happening at once and in multiple threads, ideas, information and decisions usually live in the moment and get lost just as quickly - and all together forgotten by the weekend! Using this template, this doesn't have to be the case. Have AI crawl through last week's activity, summarize all threads and generate a casual and snappy report to bring the team back into focus for the current week. A project manager's dream! How it works A scheduled trigger is set to run every Monday at 6am to gather all team channel messages within the last week. Each message thread are grouped by user and data mined for replies. Combined, an AI analyses the raw messages to pull out interesting observations and highlights. The summarized threads of the user are then combined together and passed to another AI agent to generate a higher level overview of their week. These are referred to as the individual reports. Next, all individual reports are summarized together into a team weekly report. This allows understanding of group and similar activities. Finally, the team weekly report is posted back to the channel. The timing is important as it should be the first message of the week and ready for the team to glance over coffee. How to use Ideally works best per project and where most of the comms happens on a single channel. Avoid combining channels and instead duplicate this workflow for more channels. You may need to filter for specific team members if you want specific team updates. Customise the report to suit your organisation, team or the channel. You may prefer to be more formal if clients or external stakeholders are also present. Requirements Slack for chat platform Gemini for LLM (or switch for other models) Customising this workflow If the slack channel is busy enough already, consider posting the final report to email. Pull in project metrics to include in your report. As extra context, it may be interesting to tie the messages to production performance. Use an AI Agent to query for knowledgebase or tickets relevant to the messages. This may be useful for attaching links or references to add context. Channel not so busy or way too busy for 1 week? Play with the scheduled trigger and set an interval which works for your team.
by Niranjan G
🛡️ Automated AWS Key Compromise Remediation Description This n8n workflow provides a secure, enterprise-grade response system for AWS IAM access key compromises with built-in form submission and human approval mechanisms. When an AWS access key is suspected to be compromised, this workflow enables rapid containment through a secure web form interface with basic authentication, human approval via Slack, and automated damage prevention through immediate key deactivation, credential invalidation, and comprehensive security reporting. How This Workflow is Useful Secure Form-Based Response Authenticated Form Submission**: Secure web form with basic authentication for capturing compromise details Human Approval Workflow**: Slack-based approval system for sensitive security operations Rapid Key Deactivation**: Instantly disables compromised access keys after approval Credential Invalidation**: Creates and applies security policies to invalidate existing temporary credentials Policy Analysis**: Automatically scans and analyzes both inline and attached IAM policies for the affected user AI-Powered Reporting**: Generates detailed security reports with intelligent analysis and team notifications Business Value Reduces Mean Time to Response (MTTR)**: Automates manual security procedures that typically take hours Minimizes Security Exposure**: Immediate containment prevents potential data breaches and unauthorized resource access Ensures Compliance**: Provides audit trails and documentation required for security compliance frameworks Cost Prevention**: Prevents potential financial damage from compromised credentials being used maliciously Rapid Response Capability**: Streamlines security response procedures when incidents are detected Technical Benefits AWS Best Practices**: Implements official AWS security recommendations for key compromise response Scalable Architecture**: Handles multiple access keys and complex IAM policy structures Error Handling**: Robust error handling ensures workflow continues even if individual steps fail Audit Trail**: Complete logging of all actions taken during the incident response Integration Ready**: Easily integrates with existing security tools and notification systems Use Cases 1. Incident Response Automation Automated response to security alerts from AWS CloudTrail Integration with SIEM systems for immediate key compromise response 24/7 security monitoring and automated containment 2. Compliance and Audit Meeting regulatory requirements for incident response documentation Providing audit trails for security compliance frameworks (SOC 2, ISO 27001, PCI DSS) Demonstrating due diligence in security incident handling 3. Multi-Account Management Centralized security response across multiple AWS accounts Consistent incident response procedures across different environments Standardized security automation for enterprise AWS deployments 4. Security Training and Testing Security team training on AWS incident response procedures Tabletop exercises and security drills Testing and validation of security response capabilities Key Features Core Functionality ✅ Secure Form Interface: Web form with basic authentication for secure data submission ✅ Human Approval Gate: Slack-based approval workflow for sensitive operations ✅ Authenticated Data Processing: Secure handling of form submissions with validation ✅ Immediate Key Deactivation: Instant disabling of compromised credentials after approval ✅ Security Policy Generation: Automatic creation and attachment of credential invalidation policies ✅ Policy Analysis: Deep analysis of user permissions and attached policies ✅ AI Security Analysis: Intelligent security report generation with risk assessment ✅ Team Notifications: Real-time Slack notifications to security teams ✅ Comprehensive Logging: Complete audit trail of all response actions Technical Specifications Secure Form Interface**: Web form with basic authentication for secure data capture Human Approval System**: Slack-based approval workflow for sensitive operations AWS API Integration**: Direct integration with AWS IAM APIs Authentication Layer**: Basic auth protection for form submissions Error Handling**: Robust error handling with continuation on non-critical failures Scalable Processing**: Handles multiple policies and complex IAM structures Security Best Practices**: No hardcoded credentials, uses AWS credential management Modular Design**: Easy to customize and extend for specific organizational needs Prerequisites Required Credentials AWS Credentials** with IAM permissions for: ListAccessKeys, UpdateAccessKey ListUserPolicies, ListAttachedUserPolicies CreatePolicy, AttachUserPolicy GetPolicy, GetPolicyVersion, GetUserPolicy Required Integrations Slack Workspace** for approval workflow and team notifications Basic Authentication Setup** for secure form access Optional Integrations AI Language Model** (Claude/OpenAI) for intelligent security analysis and report generation Installation and Setup Import the workflow into your n8n instance Configure AWS credentials in n8n credential manager Set up basic authentication for the secure form interface Configure Slack integration for approval notifications and team alerts Set up AI model (optional) for enhanced security analysis and reporting Configure approval workflow in Slack for human oversight Test in development environment before production use Workflow Inputs Secure Form Submission This workflow uses a secure web form with basic authentication to capture compromise details: Username**: The AWS IAM username of the compromised account Access Key ID**: The specific access key ID that has been compromised Authentication & Approval Process Form Authentication: Basic authentication protects the submission form Data Processing: Secure handling and validation of submitted credentials Human Approval: Slack notification sent to security team for approval Automated Execution: Upon approval, the workflow executes the security response This multi-layered approach ensures that sensitive security operations require both authentication and human oversight before execution. 🚀 Automate with Slack Integration Want to fully automate and simplify this workflow? Connect it with Slack for seamless team collaboration and instant response capabilities! Interactive Slack Automation Combine this AWS Key Compromise Response workflow with our Interactive Slack Approval & Data Submission System to create a fully automated incident response pipeline: Instant Slack Alerts**: Receive immediate notifications when key compromises are detected One-Click Response**: Trigger the AWS response workflow directly from Slack with interactive buttons Team Collaboration**: Enable security teams to respond collectively through Slack channels Approval Workflows**: Add human approval gates before executing critical security actions Real-time Updates**: Get live status updates and completion notifications in Slack How the Complete Solution Works Detection: External security monitoring tools (CloudTrail, SIEM, etc.) detect potential key compromise Secure Form Access: Security team accesses the authenticated web form to submit compromise details Form Submission: Credentials are securely submitted through the basic auth-protected form Human Approval: Slack notification sent to security team for review and approval Approved Execution: Upon approval, the AWS security response executes automatically Real-time Updates: Progress and completion notifications sent back to Slack Security Analysis: AI-powered analysis and comprehensive reporting delivered to the team Get Started with Full Automation To enable automatic notifications and complete the automation pipeline, use the Interactive Slack Approval & Data Submission System with Webhooks workflow: https://n8n.io/workflows/5049-interactive-slack-approval-and-data-submission-system-with-webhooks/ This integration transforms manual security responses into streamlined, team-collaborative automation that reduces response time from hours to minutes. Security Considerations Form Authentication**: Basic authentication protects the submission interface Human Approval Gate**: Slack-based approval prevents unauthorized execution AWS Credential Management**: Uses AWS credential management best practices No Sensitive Data Storage**: No sensitive data stored in workflow configuration Least-Privilege Access**: Implements least-privilege access principles Complete Audit Trails**: Provides complete audit trails for compliance Secure Data Processing**: Encrypted handling of form submissions and approvals Immediate Damage Prevention**: Designed for rapid containment after approval ⚠️ Important Disclaimer Use with Caution: Disabling access keys without proper understanding can significantly impact your personal or business operations. This workflow immediately deactivates AWS access keys, which may disrupt running applications, automated processes, or services that depend on these credentials. AWS Best Practices Recommendation: Use IAM Roles instead of Access Keys** whenever possible for enhanced security IAM roles provide temporary credentials and eliminate the need for long-term access keys Follow the principle of least privilege when assigning permissions Regularly rotate and audit your AWS credentials Implement proper monitoring and alerting for credential usage Before Using This Workflow: Ensure you understand which services and applications use the target access key Have a rollback plan in case of accidental disruption Test in a non-production environment first Coordinate with your team before executing in production For comprehensive AWS security best practices, refer to the AWS Security Best Practices Guide. For more workflows and automation solutions, visit: https://n8n.io/creators/niranjan/
by Kanaka Kishore Kandregula
Boost Sales with Automated Magento 2 Product and Coupon Notifications This n8n workflow automatically posts new Magento products & coupons to Telegram while preventing duplicates. Key benefits: ✅ Increase conversions with time-sensitive alerts (creates urgency) ✅ Reduce missed opportunities with 24/7 monitoring ✅ Improve customer engagement through rich media posts ✅ Save hours per week by automating manual posting Why This Works: Triggers impulse buys with real-time notifications Eliminates human error in duplicate posting Scales effortlessly as your catalog grows Provides analytics through database tracking Perfect for e-commerce stores wanting to: Announce new arrivals instantly Promote limited-time offers effectively Maintain consistent social presence Track performance through MySQL This workflow automatically: ✅ Detects new products AND coupons in Magento ✅ Prevents duplicate postings with MySQL tracking ✅ Posts rich formatted alerts to Telegram ✅ Runs on a customizable schedule ✨ Key Features For Products: Product name, price, and image Direct store link Media gallery support For Coupons: Coupon code and status Usage limits (times used/available) Active/inactive status indicator Core System: 🔒 MySQL duplicate prevention⏰ 1-hour schedule (customizable)📱 Telegram notifications with Markdown 🛠️ Configuration Guide Database Setup CREATE TABLE IF NOT EXISTS posted_items (item_id INT PRIMARY KEY, item_type ENUM('product', 'coupon') NOT NULL, item_value VARCHAR(255), posted BOOLEAN DEFAULT FALSE, created_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP, updated_at TIMESTAMP DEFAULT CURRENT_TIMESTAMP ON UPDATE CURRENT_TIMESTAMP); Required Credentials Magento API (HTTP Header Auth) MySQL Database Telegram Bot Sticky Notes `❗ IMPORTANT SETUP NOTES ❗ For products: Ensure 'url_key' exists in custom_attributes For coupons: Magento REST API must expose coupon rules MySQL user needs INSERT/SELECT privileges Telegram bot must be added to your channel first 🔄 SCHEDULING: - Default: Checks every 1 hours at :00 - Adjust in Schedule Trigger node ` ⚙️ Technical Details Workflow Logic: Checks for new products/coupons via Magento API Verifies against MySQL database Only posts if record doesn't exist Updates database after successful post Error Handling: Automatic skip if product/coupon exists Empty result handling Connection timeout protection 🌟 Why This Template? Complete Solution**: Handles both products AND coupons Battle-Tested**: Prevents all duplicates reliably Ready-to-Use**: Just add your credentials Fully Customizable**: Easy to modify for different needs Perfect for e-commerce stores using Magento 2 who want automated, duplicate-free social notifications!
by Vadym Nahornyi
> ⚠️ Multi-language WhatsApp Error Notifier Get instant WhatsApp alerts when any workflow fails — perfect for mobile-first monitoring and fast incident response. ✅ No coding required ✅ Works with any workflow via Error Workflow ✅ Step-by-step setup instructions included in: 🇬🇧 English 🇪🇸 Español 🇩🇪 Deutsch 🇫🇷 Français 🇷🇺 Русский 📦 What This Template Does This template sends real-time WhatsApp notifications when a workflow fails. It uses the WhatsApp Business Cloud API to deliver a preformatted error message directly to your phone. The message includes: Workflow name Error message Last executed node Example message: Error on WorkFlow: {{ $json.workflow.name }} Message: {{ $json.execution.error.message }} lastNodeExecuted: {{ $json.execution.lastNodeExecuted }} ⚙️ Prerequisites Before using this template, make sure you have: A verified Facebook Business account Access to WhatsApp Business Cloud API A sender phone number (registered in Meta) An access token (used as credentials in n8n) A pre-approved message template (or be within the 24h session window) More info from Meta Docs → 🚀 How to Use Open the template and insert your WhatsApp credentials Enter your target phone number (e.g. your own) in international format Customize the message body if needed Save the workflow but do not activate it In any other workflow → open Settings → set this as your Error Workflow 🌐 Multi-language Setup Guide Included This template includes full setup instructions with screenshots and message formatting help in: 🇬🇧 English 🇪🇸 Español 🇩🇪 Deutsch 🇫🇷 Français 🇷🇺 Русский Choose your language inside the embedded sticky note in the workflow.
by Incrementors
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 📦 Multi-Platform Price Finder: Scraping Prices with Bright Data & Telegram An intelligent n8n automation that fetches real-time product prices from marketplaces like Amazon, Wayfair, Lowe's, and more using Bright Data's dataset, and sends promotional messages via Telegram using AI—perfect for price tracking, deal alerts, and affiliate monetization. 📋 Overview This automation tracks product prices across top e-commerce platforms using Bright Data and sends out alerts via Telegram based on the best available deals. The workflow is designed for affiliate marketers, resellers, and deal-hunting platforms who want real-time competitive pricing. ✨ Key Features 🔎 Multi-Platform Scraping: Supports Amazon, Wayfair, Lowe's, and more ⚡ Bright Data Integration: Access to structured product snapshots 📢 AI-Powered Alerts: Generates Telegram-ready promo messages using AI 🧠 Lowest Price Logic: Filters and compares products across sources 📈 Data Merge & Processing: Combines multiple sources into a single stream 🔄 Keyword-Driven Search: Searches using dynamic keywords from form input 📦 Scalable Design: Built for multiple platform processing simultaneously 🧼 Clean Output: Strips unnecessary formatting before publishing 🎯 What This Workflow Does Input Search Keywords**: User-defined keyword(s) from a form trigger Platform Sources**: Wayfair, Lowe's, Amazon, etc. Bright Data API Key**: Needed for authenticated scraping Processing Steps User Input via n8n form trigger (keyword-based) Bright Data API Trigger for each marketplace Status Polling: Wait until scraping snapshot is ready Data Retrieval: Fetches JSON results from Bright Data snapshot Data Cleaning & Normalization: Price, title, and URL are extracted Merging Products from all platforms Find Lowest Price Product using custom JS logic AI Prompt Generation via Claude/Anthropic Telegram Formatting and alert message creation Output 🛍️ Product Title 💰 Final Price 🔗 Product URL ✉️ Promotional Message (for Telegram/notifications) 🚀 Setup Instructions Step 1: Import Workflow Open n8n > Workflows > + Add Workflow Import the provided JSON file Step 2: Configure Bright Data Add credentials under Credentials → Bright Data API Set the appropriate dataset_id for each platform Ensure dataset includes title, price, and url fields Step 3: Enable Keyword Trigger Use the built-in Form Trigger node Input: Single keyword field (SearchHere) Step 4: Telegram or AI Integration Modify prompt node for your language or tone Add Telegram webhook or integration where needed 📖 Usage Guide Adding Keywords Trigger the form with a product keyword like iPhone 15 Wait for workflow to fetch best deals and generate Telegram message Understanding AI-Powered Output AI creates a short, engaging message like: > "🔥 Deal Alert: Get the iPhone 15 for just ₹74,999! Limited stock—Check it out: [link]" Debugging Output Output node shows cleaned JSON with title, price, url, and message If no valid results, debug message is returned with sample structure info 🔧 Customization Options Add More Marketplaces Clone any HTTP Request node (e.g., for Wayfair) Update dataset_id and required output fields Modify Price Logic Update the Code1 node to change comparison (e.g., highest price instead of lowest) Change Message Format Edit the AI Agent prompt to customize tone/language Add emoji, CTAs, or markdown formatting as needed 🧪 Test & Activation Add a few sample keywords via form trigger Run manually or set as a webhook for external app input Check final AI-generated message in output node 🚨 Troubleshooting | Issue | Solution | |-------|----------| | No Data Returned | Ensure keyword matches real products | | Status Not 'Ready' | Bright Data delay; add Wait nodes | | Invalid API Key | Check Bright Data credentials | | AI Errors | Adjust prompt or validate input fields | 📊 Use Cases 💰 Affiliate Campaigns: Show best deals across platforms 🛒 Deal Pages: Post live offers with product links 🧠 Competitor Analysis: Track cross-platform pricing 🔔 Alert Bots: Send real-time alerts to Telegram or Slack ✅ Quick Setup Checklist [x] Bright Data API credentials configured [x] n8n form trigger enabled [x] Claude or AI model connected [x] All HTTP requests working [x] AI message formatting verified 🌐 Example Output { "title": "Apple iPhone 15 Pro Max", "price": 1199, "url": "https://amazon.com/iphone-15", "message": "🔥 Grab the Apple iPhone 15 Pro Max for just $1199! Limited deal—Check it out: https://amazon.com/iphone-15" } 📬For any questions or support, please contact: 📧 <info@incrementors.com> or fill out this form: https://www.incrementors.com/contact-us/
by Hybroht
Source Discovery - Automatically Search More Up-to-Date Information Sources 🎬 Overview Version : 1.0 This workflow utilizes various nodes to discover and analyze potential sources of information from platforms like Google, Reddit, GitHub, Bluesky, and others. It is designed to streamline the process of finding relevant sources based on specified search themes. ✨ Features Automated source discovery from multiple platforms. Filtering of existing and undesired sources. Error handling for API requests. User-friendly configuration options. 👤 Who is this for? This workflow is ideal for researchers, content marketers, journalists, and anyone looking to efficiently gather and analyze information from various online sources. 💡 What problem does this solve? This workflow addresses the challenge of manually searching for relevant information sources, saving time and effort while ensuring that users have access to the most pertinent content. Ideal use-cases include: Resource Compilation for Academic and Educational Purposes Journalism and Research Content Marketing Competitor Analysis 🔍 What this workflow does The workflow gathers data from selected platforms through search terms. It filters out known and undesired sources, analyzes the content, and provides insights into potential sources relevant to the user's needs. 🔄 Workflow Steps 1. Search Queries Fetch sources using SerpAPI search, DuckDuckGo, and Bluesky. Utilizes GitHub repositories to find relevant links. Leverages RSS feeds from subreddits to identify potential sources. 2. Filtering Step Removes existing and undesired sources from the results. 3. Source Selection Analyzes the content of the identified sources for relevance. 📌 Expected Input / Configuration The workflow is primarily configured via the Configure Workflow Args (Manual) node or the Global Variables custom node. Search themes: Keywords or phrases relevant to the desired content. Lists of known sources and undesired sources for filtering. 📦 Expected Output A curated list of potential sources relevant to the specified search themes, along with insights into their content. 📌 Example ⚙️ n8n Setup Used n8n version:** 1.105.3 n8n-nodes-serpapi:** 0.1.6 n8n-nodes-globals:** 1.1.0 n8n-nodes-bluesky-enhanced**: 1.6.0 n8n-nodes-duckduckgo-search**: 30.0.4 LLM Model:** mistral-small-latest (API) Platform:** Podman 4.3.1 on Linux Date:** 2025-08-06 ⚡ Requirements to Use / Setup Self-hosted or cloud n8n instance. Install the following custom nodes: SerpAPI, Bluesky, and DuckDuckGo Search. n8n-nodes-serpapi n8n-nodes-duckduckgo-search n8n-nodes-bluesky-enhanced Install the Global Variables Node for enhanced configuration: n8n-nodes-globals (or use Edit Field (Set) node instead) Provide valid credentials to nodes for your preferred LLM model, SerpAPI, and Bluesky. Credentials for GitHub recommended. ⚠️ Notes, Assumptions \& Warnings Ensure compliance with the terms of service of any platforms accessed or discovered in this workflow, particularly concerning data usage and attribution. Monitor API usage to avoid hitting rate limits. The workflow may encounter errors such as 403 responses; in such cases, it will continue by ignoring the affected substep. Duplicate removal is applied, but occasional overlaps might still appear depending on the sources. This workflow assumes familiarity with n8n, APIs, and search engines. Using AI agents (Mistral or substitute LLMs) requires access to their API services and keys. This is not a Curator of News. It is designed to find websites that are relevant and useful to your searches. If you are looking for a relevant news selector, please check this workflow. ℹ️ About Us This workflow was developed by the Hybroht team. Our goal is to create tools that harness the possibilities of technology and more. We aim to continuously improve and expand functionalities based on community feedback and evolving use cases. For questions, reach out via contact@hybroht.com. ⚖️ Warranty & Legal Notice This free workflow is provided "as-is" without any warranties of any kind, either express or implied, including but not limited to the implied warranties of merchantability, fitness for a particular purpose, or non-infringement. By using this workflow, you acknowledge that you do so at your own risk. We shall not be held responsible for any damages, losses, or liabilities arising from the use or inability to use this workflow, including but not limited to any direct, indirect, incidental, or consequential damages. It is your responsibility to ensure that your use of this workflow complies with all applicable laws and regulations.
by Max
N8N Automated Twitter Reply Bot Workflow For latest version, check: dziura.online/automation Latest documentation can be find here You must have Apify community node installed before pasting the JSON to your workflow. Overview This n8n workflow creates an intelligent Twitter/X reply bot that automatically scrapes tweets based on keywords or communities, analyzes them using AI, generates contextually appropriate replies, and posts them while avoiding duplicates. The bot operates on a schedule with intelligent timing and retry mechanisms. Key Features Automated tweet scraping** from Twitter/X using Apify actors AI-powered reply generation** using LLM (Large Language Model) Duplicate prevention** via MongoDB storage Smart scheduling** with timezone awareness and natural posting patterns Retry mechanism** with failure tracking Telegram notifications** for status updates Manual trigger** option via Telegram command Required Credentials & Setup 1\. Telegram Bot Create a bot via @BotFather on Telegram Get your Telegram chat ID to receive status messages Credential needed**: Telegram account (Bot token) 2\. MongoDB Database Set up a MongoDB database to store replied tweets and prevent duplicates Create a collection (default name: collection\_name) Credential needed**: MongoDB account (Connection string) Tutorial**: MongoDB Connection Guide 3\. Apify Account Sign up at Apify.com Primary actors used**: Search Actor: api-ninja/x-twitter-advanced-search - For keyword-based tweet scraping (ID: 0oVSlMlAX47R2EyoP) Community Actor: api-ninja/x-twitter-community-search-post-scraper - For community-based tweet scraping (ID: upbwCMnBATzmzcaNu) Credential needed**: Apify account (API token) 4\. OpenRouter (LLM Provider) Sign up at OpenRouter.ai Used for AI-powered tweet analysis and reply generation Model used**: x-ai/grok-3 (configurable) Credential needed**: OpenRouter account (API key) 5\. Twitter/X API Set up developer account at developer.x.com Note**: Free tier limited to ~17 posts per day Credential needed**: X account (OAuth2 credentials) Workflow Components Trigger Nodes 1\. Schedule Trigger Purpose**: Runs automatically every 20 minutes Smart timing**: Only active between 7 AM - 11:59 PM (configurable timezone) Randomization**: Built-in probability control (~28% execution chance) to mimic natural posting patterns 2\. Manual Trigger Purpose**: Manual execution for testing 3\. Telegram Trigger Purpose**: Manual execution via /reply command in Telegram Usage**: Send /reply to your bot to trigger the workflow manually Data Processing Flow 1\. MongoDB Query (Find documents) Purpose**: Retrieves previously replied tweet IDs to avoid duplicates Collection**: collection\_name (configure to match your setup) Projection**: Only fetches tweet\_id field for efficiency 2\. Data Aggregation (Aggregate1) Purpose**: Consolidates tweet IDs into a single array for filtering 3\. Keyword/Community Selection (Keyword/Community List) Purpose**: Defines search terms and communities Configuration**: Edit the JSON to include your keywords and Twitter community IDs Format:{ "keyword\_community\_list": \[ "SaaS", "Entrepreneur", "1488663855127535616" // Community ID (19-digit number) \], "failure": 0 } 4\. Random Selection (Randomized community, keyword) Purpose**: Randomly selects one item from the list to ensure variety 5\. Routing Logic (If4) Purpose**: Determines whether to use Community search or Keyword search Logic**: Uses regex to detect 19-digit community IDs vs keywords Tweet Scraping (Apify Actors) Community Search Actor Actor**: api-ninja/x-twitter-community-search-post-scraper Purpose**: Scrapes tweets from specific Twitter communities Configuration:{ "communityIds": \["COMMUNITY\_ID"\], "numberOfTweets": 40 } Search Actor Actor**: api-ninja/x-twitter-advanced-search Purpose**: Scrapes tweets based on keywords Configuration:{ "contentLanguage": "en", "engagementMinLikes": 10, "engagementMinReplies": 5, "numberOfTweets": 20, "query": "KEYWORD", "timeWithinTime": "2d", "tweetTypes": \["original"\], "usersBlueVerifiedOnly": true } Filtering System (Community filter) The workflow applies multiple filters to ensure high-quality replies: Text length**: >60 characters (substantial content) Follower count**: >100 followers (audience reach) Engagement**: >10 likes, >3 replies (proven engagement) Language**: English only Views**: >100 views (visibility) Duplicate check**: Not previously replied to Recency**: Within 2 days (configurable in actor settings) AI-Powered Reply Generation LLM Chain (Basic LLM Chain) Purpose**: Analyzes filtered tweets and generates contextually appropriate replies Model**: Grok-3 via OpenRouter (configurable) Features**: Engagement potential scoring User authority analysis Timing optimization Multiple reply styles (witty, informative, supportive, etc.) <100 character limit for optimal engagement Output Parser (Structured Output Parser) Purpose**: Ensures consistent JSON output format Schema:{ "selected\_tweet\_id": "tweet\_id\_here", "screen\_name": "author\_screen\_name", "reply": "generated\_reply\_here" } Posting & Notification System Twitter Posting (Create Tweet) Purpose**: Posts the generated reply as a Twitter response Error handling**: Catches API limitations and rate limits Status Notifications Success**: Notifies via Telegram with tweet link and reply text Failure**: Notifies about API limitations or errors Format**: HTML-formatted messages with clickable links Database Storage (Insert documents) Purpose**: Saves successful replies to prevent future duplicates Fields stored**: tweet\_id, screen\_name, reply, tweet\_url, timestamp Retry Mechanism The workflow includes intelligent retry logic: Failure Counter (If5, Increment Failure Counter1) Logic**: If no suitable tweets found, increment failure counter Retry limit**: Maximum 3 retries with different random keywords Wait time**: 3-second delay between retries Final Failure Notification Trigger**: After 4 failed attempts Action**: Sends Telegram notification about unsuccessful search Recovery**: Manual retry available via /reply command Configuration Guide Essential Settings to Modify MongoDB Collection Name: Update collection\_name in MongoDB nodes Telegram Chat ID: Replace 11111111111 with your actual chat ID Keywords/Communities: Edit the list in Keyword/Community List node Timezone: Update timezone in Code node (currently set to Europe/Kyiv) Actor Selection: Enable only one actor (Community OR Search) based on your needs Filter Customization Adjust filters in Community filter node based on your requirements: Minimum engagement thresholds Text length requirements Time windows Language preferences LLM Customization Modify the AI prompt in Basic LLM Chain to: Change reply style and tone Adjust engagement criteria Modify scoring algorithms Set different character limits Usage Tips Start small: Begin with a few high-quality keywords/communities Monitor performance: Use Telegram notifications to track success rates Adjust filters: Fine-tune based on the quality of generated replies Respect limits: Twitter's free tier allows ~17 posts/day Test manually: Use /reply command for testing before scheduling Troubleshooting Common Issues No tweets found: Adjust filter criteria or check keywords API rate limits: Reduce posting frequency or upgrade Twitter API plan MongoDB connection: Verify connection string and collection name Apify quota: Monitor Apify usage limits LLM failures: Check OpenRouter credits and model availability Best Practices Monitor your bot's replies for quality and appropriateness Regularly update keywords to stay relevant Keep an eye on engagement metrics Adjust timing based on your audience's activity patterns Maintain a balanced posting frequency to avoid appearing spammy Documentation Links Full Documentation**: Google Doc Guide Latest Version**: dziura.online/automation MongoDB Setup Tutorial**: YouTube Guide This workflow provides a comprehensive solution for automated, intelligent Twitter engagement while maintaining quality and avoiding spam-like behavior.
by Nik B.
Automatically fetches daily sales, shifts, and receipts from Loyverse. Calculates gross profit, net operating profit, other key metrics, saves them to a Google Sheet and sends out a daily report via email. Who’s it for This template is for any business owner, manager, or analyst using Loyverse POS who needs more advanced financial reporting. If you're a restaurant, bar, or retail owner who wants to automatically track daily net profit, compare sales to historical averages, and build a custom financial dashboard in Google Sheets, this workflow is for you. How it works / What it does This workflow runs automatically on a daily schedule. It fetches all sales data and receipts from your Loyverse account for the previous business day, defined by your custom shift times (even past midnight). A powerful Code node then processes all the data to calculate the metrics that Loyverse either doesn't provide at all, or only spreads out across several separate reports instead of in one consolidated place. Already set up are metrics like... -Total Revenue, Gross Profit, and Net Operating Profit Cash handling differences (over/under) Average spend per receipt (ATV) 30-day rolling Net Operating Profit (NOP) Performance vs. your historical weekday average Finally, it appends the single, calculated row of daily metrics to a Google Sheet and sends an easily customizable summary report to your email. How to set up This workflow includes detailed Sticky Notes to guide you through the setup process. Because every business has a unique POS configuration (different POS devices, categories, and payment types), you'll need to set up a few things manually before executing the workflow. I've tried to make this as easy as possible to follow, and the entire setup should only take about 15 minutes. Preparations & Credential setup Subscribe to "Integrations" Add-on in Loyverse ($9 / month) to gain API access. Create an Access token in Loyverse Create Credentials: In your n8n instance, create credentials for Loyverse (use "Generic" > "Bearer Auth"), Google Sheets (OAuth2), and your Email (SMTP or other). Make a copy of a prep-configured Google Spreadsheet (Link in the second sticky note inside the workflow). Fill MASTER CONFIG: Open the MASTER CONFIG node. Follow the comments inside to add your Google Sheet ID, Sheet Names, business hours, timezone, and Loyverse IDs (for POS devices, payment types, and categories). Configure Google Sheet Nodes Configure Read Historical Data: Open this node. Follow the instructions in the nearby Sticky Note to paste the expressions for your Document ID and Sheet Name. Configure Save Product List: Open this node. Paste in the expressions for Document ID and Sheet Name. The column mapper will load; map your sheet columns (e.g., item_name) to the data on the left (e.g., {{ $json.item_name }}). Configure Save Latest Sales Data: Open this node. Paste in the expressions for Document ID and Sheet Name. Save and run the workflow. After that, the column mapper will load. This is the most important step: map your sheet's column names (e.g., "Total Revenue") to the calculated metrics from the Calculate All Metrics node (e.g., {{ $json.totalGrossRevenue }}). Activate the workflow. 🫡 Requirements Loyverse Integrations Subscription Loyverse Access Token Credentials for Loyverse (Bearer Auth) Credentials for Google Sheets (OAuth2) Credentials for Email/SMTP sender How to customize the workflow This template is designed to be highly flexible. Central Configuration: Almost all customization (POS devices, categories, payment types, sheet names) is done in the MASTER CONFIG node. You don't need to dig through other nodes. Add/Remove Metrics: The Calculate All Metrics node has additional metrics already set up, just add the relevant collumns to the SalesData sheet or even add your own calculations to the node. Any new metric you add (e.g., metrics.myNewMetric = 123) will be available to map in the Save Latest Sales Data node. Email Body: You can easily edit the Send email node to change the text or add new metrics from the Calculate All Metrics node.