by Εukasz
What Is This? This workflow is a comprehensive solution for automating website audits and optimizations, leveraging advanced technologies to boost SEO effectiveness and overall site performance. Who Is It For? Designed for SEO specialists, digital marketers, webmasters, and content teams, this workflow empowers anyone responsible for website performance to automate and scale their audit processes. Agencies managing multiple client sites, in-house SEO teams aiming to save time on routine checks, and developers seeking to integrate data-driven insights into their deployment pipelines will all find this solution invaluable. By combining your siteβs sitemap with Google Search Console and Google Analytics data, then applying AI-powered analysis, the workflow continuously uncovers actionable recommendations to boost search visibility, improve user engagement, and accelerate page performance. Whether you manage a single blog or oversee a sprawling e-commerce platform, this automated pipeline delivers precise, prioritized SEO improvements without manual data wrangling. How Does It Work? This end-to-end site analysis automation consists of five main stages: 1. URL Discovery Processes the sitemap.xml using HTTP Request and XML nodes to extract all site URLs. 2. Search Console Performance Analysis Uses the Google Search Console API to fetch detailed metrics for each page, including search position, clicks, impressions, and CTR. 3. Analytics Data Collection Connects to the Google Analytics API to automatically retrieve traffic metrics such as pageviews, average session duration, bounce rate, and conversions. 4. AI Data Processing Employs OpenAI models to perform in-depth analysis of the collected data. The artificial intelligence engine merges insights from all sources, identifies patterns, and produces detailed optimization recommendations. AI analyses website itsefl aswell. Consider testing different models. I do recommend at least trying out o4-mini. 5. Recommendation Generation Creates tailored suggestions for each page, in form of HTML table, that is being sent to your email. How To Set It Up? Accounts: An active n8n account or instance, API keys for Google Search Console and Google Analytics, an OpenAI access token. Enabled Google APIs: You will neeed at least following scopes: Google Search Console API Google Analytics Aadmin API Google Analytics Data API Scheduling: The workflow can run manually for ad hoc audits or be scheduled (daily, weekly) for continuous site monitoring. Testing: There are two nodes that are optional: "Sort for testing purposes" and "Limit for testing purposes" Together they randomly select items from sitemap and limit them to few so you don't need to run hundreds of sitemap.xml items at once, but you can run just a random batch first. Globals: There is node called "Globals- CHANGE ME!". You need to set up proper variables in there, which are: sitemap_url - self exlpainatory search_console_selector - for example "sc-domain:sailingbyte.com" but can be URL aswell- depends on how did you set up your search console analysis_start_date and analysis_end_date - date range for analytics, by default last 30 days analytics_selector_id - ID of Google Analytics setup, it is a large integer, you can find it in analytics url preceeded with letter "p", ex (your number is where there are X's): https://analytics.google.com/analytics/web/#/pXXXXXXXXX/reports/intelligenthome report_receiver - email which will receive report What's More? That's actually it. I hope that this automation will help your website improvement will be much easier! Thank you, perfect! Glad I could help. Visit my profile for other automations for businesses. And if you are looking for dedicated software development, do not hesitate to reach out!
by Raz Hadas
Description Transform your investment strategy with a fully automated, AI-driven trading bot. This workflow bridges the gap between AI-powered market insights and real-world trading by executing buy and sell orders directly through the Alpaca paper trading API. Designed to work in tandem with the Automated Stock Sentiment Analysis workflow, this solution takes the top-performing stocks based on daily news sentiment and automatically rebalances your portfolio. It's perfect for algorithmic traders, data-driven investors, and n8n enthusiasts who want to see their AI analysis translate into tangible actions, all while maintaining a comprehensive log of every transaction in Google Sheets. Key Features & Benefits Automated Trading Execution:** Automatically places buy and sell orders on the Alpaca paper trading platform without manual intervention. Sentiment-Driven Decisions:** Leverages the output from the sentiment analysis workflow to make informed decisions, selling positions with waning sentiment and buying into those with high positive sentiment. Dynamic Portfolio Rebalancing:** Intelligently calculates which positions to close and how to allocate the resulting funds into new, high-potential stocks. Paper Trading Ready:** Safely test and refine your trading strategies in a risk-free environment using Alpaca's paper trading API. Daily Performance Tracking:** Automatically logs your account equity and daily percentage change to a Google Sheet, giving you a clear view of your portfolio's performance. Detailed Trade Logging:** Every buy and sell order is meticulously recorded in a Google Sheet for easy review and historical analysis. Scheduled and Autonomous:** The entire process runs on a daily schedule, making it a "set and forget" solution for systematic trading. How It Works This workflow executes a sophisticated, automated trading strategy in a few key stages: Daily Kick-off & Snapshot: The workflow triggers on a daily schedule, first fetching your current Alpaca account balance and logging it to a Google Sheet to track daily performance. Strategy Formulation: It then reads the daily sentiment scores produced by the accompanying "Stock Sentiment Analysis" workflow. A Code node filters these results to identify the top four stocks with the highest positive sentiment. The Decision Engine: The core of the workflow is a custom Code node that acts as the trading brain. It: Retrieves your currently open positions from Alpaca. Compares your holdings against the day's top four sentiment stocks. Generates a "sell list" of positions you hold that are no longer in the top four. Generates a "buy list" of top-sentiment stocks that you don't yet own. Calculates the total cash value from the "sell list" and determines the exact notional value to invest in each stock on the "buy list." Trade Execution: The workflow first iterates through the "sell list" and executes a DELETE request to Alpaca for each, closing the positions. A Wait node pauses the workflow for two minutes to ensure the sell orders are filled and the account balance is updated. It then iterates through the "buy list," executing POST requests to Alpaca to purchase the new assets with the calculated funds. Record Keeping: All executed orders (both buys and sells) are merged and logged in a dedicated Google Sheet, giving you a permanent and detailed transaction history. Nodes Used Schedule Trigger HttpRequest (Alpaca API) Google Sheets Code (JavaScript) SplitOut Wait Merge This workflow is the perfect next step for anyone looking to take their AI analysis to the next level. Take the emotion out of your trading and let this bot systematically execute your data-driven strategy.
by Jonathan | NEX
Are you drowning in a sea of security notifications? Do your analysts spend more time sifting through low-level logs than investigating real threats? This workflow transforms n8n into an autonomous SOC (Security Operations Center) Analyst, tackling alert fatigue head-on. Leveraging the NixGuard Security RAG connector, this workflow automates the entire alert triage process. It ingests raw security events (from sources like Wazuh, your SIEM, or EDR), uses AI to analyze and assign a priority, and then intelligently routes the alert to the correct Slack channel. How It Works: Ingest & Filter: The workflow runs on a schedule, fetching all recent security alerts. It first performs a basic filtering to isolate events that meet a minimum severity threshold (e.g., level 7+). AI Analysis & Prioritization: The aggregated high-severity alerts are then sent to the AI with a specific prompt, asking it to analyze the situation and return a structured JSON object containing a single, overall priority (Critical, High, Info) and a concise summary. Intelligent Routing: A Switch node reads the AI-assigned priority and routes the notification to the appropriate destination. Critical alerts go to your #security-incident-response channel, high-priority alerts to #security-investigations, and informational ones to #security-logs. Key Features & Benefits: Eliminate Alert Fatigue:** Drastically reduce the noise by having AI pre-process and categorize alerts before they hit your team. Automate SOC Tier 1 Triage:** Free up your human analysts from repetitive triage tasks so they can focus on high-value investigation and threat hunting. Faster Incident Response:** Route critical alerts to the right people in real-time, cutting down on crucial response time. Consistent Prioritization:** Use AI to ensure a consistent, unbiased approach to alert prioritization, 24/7. Smart Routing Logic:** Go beyond simple keyword matching. The Switch node ensures alerts are delivered to the team best equipped to handle them based on AI-assessed severity. Who is this for? SOC Analysts & Security Engineers** looking to automate alert triage and incident response workflows. SecOps and DevOps Teams** who want to build a more efficient, automated security operations pipeline. IT Managers and Directors** aiming to improve their team's efficiency and reduce the risk of missing critical alerts. Anyone using Wazuh, a SIEM, or other security tools that generate a high volume of alerts. Stop manually triaging alerts. Install this workflow to build your own AI-powered security automation platform and let your team focus on what matters most. Don't have the main workflow yet? Get it HERE! π Learn more about NixGuard: thenex.world π Get started with a free security subscription: thenex.world/security/subscribe Tags / Keywords: AI, Security, SOC, Automation, Triage, Alerting, Cybersecurity, Wazuh, SIEM, Slack, Incident Response, Alert Fatigue, SecOps, Generative AI, LLM, NixGuard, Routing
by Evoort Solutions
β¨ AI Text Summarizer with Google Sheets Logging Summarize large blocks of text into concise outputs using the Text Summarizer AI API and automatically log results in Google Sheets. This automation is ideal for content creators, marketers, researchers, and teams who need efficient summarization and record-keepingβwithout writing a single line of code. π Features | Feature | Description | |--------------------------|-----------------------------------------------------------------------------| | Web Form Trigger | Collects title, content, mode (Paragraph/Bullet), and length preferences. | | API Integration | Sends content to Text Summarizer AI API via RapidAPI. | | Conditional Logic | Routes success and error cases appropriately. | | Google Sheets Logging| Stores summaries or error messages into Google Sheets. | | Error Handling | Captures failed summaries and ensures no data is lost. | π§ Problem Solved Manually summarizing long-form content is time-consuming and repetitive. Storing this output in structured logs (like Google Sheets) adds an extra layer of manual effort. This workflow solves that by: π Automating AI-driven summarization using Text Summarizer AI API. π Storing results (or fallback errors) directly into a Google Sheet. β Ensuring no request goes unloggedβeven on API failure. π― Use Cases π Blog Writers: Quickly convert articles into summaries for social captions. π Students: Break down textbook or lecture content. π§ Knowledge Management: Turn raw meeting notes into concise summaries. π’ SEO Teams: Use bullet-point output for schema markup or meta descriptions. π§© Nodes in the Flow | Node Name | Purpose | |----------------------|-------------------------------------------------------------------------| | On form submission | Captures user input via form (title, content, mode, length). | | Mapping | Formats input to match Text Summarizer AI API specs. | | HTTP Request | Sends POST request to the summarization API on RapidAPI. | | If | Validates whether a summary was returned. | | Wait | Adds short delay before writing to spreadsheet (success). | | Google Sheets | Appends summary data to Google Sheet. | | Wait1 | Adds delay for error handling path. | | Google Sheets1 | Logs failure with an βError occurredβ message. | β Benefits π§ AI-Powered: Uses Text Summarizer AI API for fast and contextual summaries. π Organized Logs: Google Sheets integration ensures easy tracking and auditing. π Reliable: Captures all submissions, including failed ones. π§© Customizable: Easily adapt inputs or connect to other tools like Notion, Slack, or Airtable. π οΈ Requirements β n8n account (Cloud or Self-Hosted) β Access to Text Summarizer AI API β Google account (for Sheets integration) π API key from RapidAPI (used in the HTTP Request node) π Workflow Overview User submits text and preferences (mode, length) via a form. Workflow triggers and transforms the input. Formatted data is sent to the Text Summarizer AI API. If a valid summary is returned: Log it into Google Sheets. If the API fails: Log the error message instead. π₯ Import Instructions Open n8n and import the workflow JSON. Replace the x-rapidapi-key in the HTTP node with your personal RapidAPI key. Configure your Google Sheets credentials. Deploy and test. β π Suggested Extensions π’ Notion database logging π£ Slack/Discord notification for each summary π΅ CSV download or Airtable sync π· Tags ai summarization text-processing rapidapi google-sheets automation markdown n8n Text Summarizer AI API Create your free n8n account and set up the workflow in just a few minutes using the link below: π Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!
by Simeon
Google Calendar MCP β Context-Aware Calendar Operations This n8n template implements an MCP (Model Context Protocol)-compliant module for managing Google Calendar events in a context-aware, conflict-free manner. π§ What It Does This MCP enables structured interaction with Google Calendar based on context and intent, ensuring reliable, reusable operations with awareness of existing data and state. β Core Capabilities Context-aware event creation** Prevents overlapping by validating time availability before creating new events. Gap validation** Checks if a time range is busy or free, enabling smarter scheduling decisions. Conditional updates** Only updates events after confirming their existence and current state. Safe deletion** Removes events using MCP principles of validation and traceability. π How to Use To use this MCP in your context-aware systems: Deploy the template in your n8n instance. Locate the Server node in the workflow β it exposes a Server-Sent Events (SSE) URL. Copy that SSE URL. Use that URL as the entry point for your MCP client or orchestrator. This URL acts as the communication bridge, allowing you to interact with the MCP-compliant Google Calendar logic using standard MCP semantics.
by Julien DEL RIO
This template is inspired by Save your workflows into a GitHub repository by hikerspath and Back Up Your n8n Workflows To Github by jon-n8n. Basic Retrieve all workflows from an n8n instance and save it on a gitlab project. If the workflow exist, il will only save the changes. Flow What the workflow does : Sets custom parameters Gets workflows Iterates through each workflow one by one Get the file from Gitlab if exists Compare the files as objects (not as strings) Return a status on the workflow Create, Edit or ignore the file depending on the status Return a list of status for each workflow Configuration Select a credential in each Gitlab nodes. Edit the data in node "Globals" : repo.owner : slug of the user or team owning the repo repo.name : slug of the repository repo.branch : branch to commit on repo.path : from root of the repository. Should end with / Comments Error on gitlab nodes will not stop the run but will list the current workflow as error in the results Some fields are ignored to determined if there is changes : updatedAt : should be ignored if only ignores fields are changed globals : it's running information, no need to follow the changes
by Evoort Solutions
AI-Powered Image Background Removal Workflow with Google Sheets Integration Flow Description: This workflow utilizes AI-powered image background removal integrated with Google Sheets to create a fully automated and streamlined process for handling and managing image files. The flow is triggered when a user uploads an image through a form. The image is sent to the API Background Remover AI, where it undergoes automatic background removal. Upon successful processing, the new image is uploaded to a temporary file storage service using the Temp File Upload. Afterward, the relevant data, including the image link and status, is logged in a Google Sheets document for easy access and tracking. In case the process fails, the system automatically logs a failure status in the same Google Sheet, along with the reason (if applicable). This allows users to have a transparent, organized, and real-time view of both successful and failed background removal attempts. Used APIs: Background Remover AI: An AI-powered service that removes backgrounds from images. This service offers a fast, accurate, and scalable solution for background removal in images. Temp File Upload: This API facilitates the upload of processed images to a temporary file storage service, making it easy to access and manage files before permanent storage. Use Case: This workflow is highly beneficial for businesses and developers who need to process multiple images automatically. It helps automate tedious tasks such as background removal, making it an efficient tool for industries like: E-commerce**: Automatically removing backgrounds from product images for clean, professional-looking listings across online platforms such as Amazon, eBay, or Shopify. Content Creation**: Content creators can quickly remove backgrounds from images for blogs, social media posts, and marketing campaigns, saving significant time in photo editing. Real Estate**: Real estate businesses can use this workflow to enhance property images by removing unwanted backgrounds, making them look more polished and appealing for listings. Advertising & Marketing**: This workflow simplifies image preparation for digital ads, banners, and promotional content by automatically cleaning up images for a more professional look. Benefits: Time-Saving: By automating the background removal process via the **Background Remover AI API, you eliminate the need for manual image editing, saving time and resources. AI-Powered Accuracy**: The AI-powered background removal service ensures precise and high-quality results consistently. Seamless Integration with Google Sheets: All successful and failed image processing attempts are automatically logged into a **Google Sheets document, ensuring you have a transparent, real-time record of each operation. Error Tracking**: In case of failure, detailed error logs are created in Google Sheets, allowing easy tracking and troubleshooting. Efficient Cloud Storage: The **Temp File Upload API stores processed images securely in the cloud, offering a temporary solution before permanent storage. Google Sheets Table Example: The data from the workflow will be automatically added to a Google Sheets document, creating an organized table with information about the processed images. The table will have the following columns: | Image Name | Link | Status | Expire At | |-----------------|----------|------------|---------------------| | image1.jpg | Link | Success | 2025-07-25T12:00:00Z | | image2.jpg | Link | Success | 2025-07-25T12:00:00Z | | image3.jpg | Not found | Failed | 2025-07-24T12:00:00Z | | image4.jpg | Link | Success | 2025-07-25T12:00:00Z | Columns Explained: Image Name**: The name of the image file uploaded by the user. Link**: A direct link to the processed image stored in temporary file storage. Status: Indicates whether the background removal was **Successful or Failed. Expire At**: The expiration date and time when the temporary file link will no longer be accessible. This table provides real-time tracking of each image processing event, offering full visibility of the workflow results. It is ideal for businesses or developers who need to keep a record of their image-processing operations. Additional Features: Automatic Error Logging**: If the background removal fails for any reason, a failure entry is recorded in Google Sheets with a timestamp and an error message. Custom Expiry Time**: The system automatically sets an expiry time for the processed image, allowing temporary access before it expires and is removed from storage. Scalable Process**: The workflow can easily handle multiple form submissions and process images in bulk, making it scalable for various use cases. Create your free n8n account and set up the workflow in just a few minutes using the link below: π Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!
by Patrick Campbell
Who's this for Small business owners, finance teams, accountants, and bookkeepers who use Xero for invoicing and want to improve cash flow by automating payment reminders. If you're spending time manually following up on unpaid invoices or struggling with late payments, this workflow eliminates the manual effort and ensures consistent, timely communication with customers while maintaining a complete audit trail. What it does This workflow automatically monitors all invoices in your Xero account and sends friendly payment reminders to customers when invoices are approaching their due date. It runs daily at noon, checks every invoice, calculates how many days until payment is due, sends personalized email reminders for invoices due within 7 days, and logs each reminder activity back into Xero's invoice history. The automation ensures no invoice slips through the cracks, reduces the administrative burden of accounts receivable management, and maintains professional customer relationships through polite, timely remindersβall while keeping your Xero records up to date with reminder tracking. How it works The workflow executes automatically every day at 12 PM and follows this process: Triggers the daily check using the Schedule Trigger node Fetches all invoices from your Xero account using the Xero API integration Filters out invoices that are already marked as "PAID" to avoid sending unnecessary reminders Calculates the number of days remaining until each unpaid invoice is due using a JavaScript code node Identifies invoices that are due within the next 7 days (customizable threshold) Sends personalized email reminders to customers via Microsoft Outlook, including invoice number, due date, and amount Logs the reminder activity back into Xero's invoice history with the date sent and days until due Creates a complete audit trail in Xero showing when reminders were sent for each invoice The workflow only sends reminders for invoices meeting the criteria, so customers aren't bombarded with unnecessary emails. The Xero history logging ensures your team can see at a glance which customers have been reminded and when, preventing duplicate reminders and providing accountability. Requirements Xero account with API access enabled (available to all Xero users at no additional cost) Microsoft Outlook or Office 365 account for sending email reminders Valid email addresses configured for all customers in your Xero contact records n8n instance (self-hosted or cloud) with credentials configured for: Xero OAuth2 connection (used twice: once for fetching invoices, once for logging history) Microsoft Outlook OAuth2 connection Setup instructions 1. Enable Xero API access Ensure your Xero account has API access enabled. This is available by default for all Xero accounts. You'll need administrator access to create the API connection. 2. Configure n8n credentials In your n8n instance, set up OAuth2 credentials for: Xero:** Follow n8n's Xero credential documentation to authorize access to your Xero organization. Make sure the credentials have permission to both read invoices and write to invoice history. Microsoft Outlook:** Set up OAuth2 connection to allow n8n to send emails on your behalf 3. Assign credentials to nodes Open the workflow and assign your configured credentials to these nodes: "Fetch All Xero Invoices" β Select your Xero credential "Send Email Reminder to Customer" β Select your Microsoft Outlook credential "Log Reminder in Xero History" β Select your Xero credential (same as above) 4. Customize the email template Edit the "Send Email Reminder to Customer" node to personalize the message: Update the sender name and signature Add your company branding or logo Include payment instructions or online payment links Adjust the tone to match your customer communication style Add any legal disclaimers or terms if required Customize the subject line if needed 5. Adjust the reminder threshold (optional) By default, reminders are sent for invoices due within 7 days. To change this: Open the "Calculate Days Until Due" code node Find the line: isDueSoon: diffDays <= 7 && diffDays >= 0 Change 7 to your preferred number of days (e.g., 14 for two weeks notice) 6. Test the workflow Before enabling the daily schedule: Use the manual trigger to test with your actual Xero data Verify that invoices are fetched correctly Check that the date calculations are accurate Send a test email to yourself to review the message format Confirm the reminder is logged in Xero's invoice history Verify only qualifying invoices trigger reminders 7. Activate the workflow Once testing is complete, activate the workflow. It will run automatically every day at noon (or your customized schedule time).
by Rahi
Workflow 1: Domain and Email Health π©Ί This part of the workflow is triggered every 5 hours by the Schedule Trigger1 node. Its purpose is to pull health metrics for both email domains and individual email addresses. How it Works: ++Schedule Trigger:++ The Schedule Trigger1 node initiates the workflow every 5 hours. ++API Requests:++ Two separate HTTP Request nodes, HTTP Request5 and HTTP Request6, make API calls to Smartlead. ++HTTP Request5 calls++ the endpoint for domain-wise health metrics. ++HTTP Request6 calls++ the endpoint for email-wise health metrics. Both requests use the same api_key and a date range from 2025-07-04 to the current day. ++Data Splitting:++ The Split Out5 and Split Out6 nodes take the JSON response from the API calls and split the data into individual items. This is necessary so each row of data can be processed and added to Google Sheets separately. ++Google Sheets Integration:++ Finally, the Append or update row in sheet5 and Append or update row in sheet6 nodes update two different Google Sheets: ++Append or update row in sheet5 adds++ or updates rows in the DomainHealth sheet, matching on the domain column. ++Append or update row in sheet6 adds++ or updates rows in the EmailHealth sheet, matching on the from_email column. Workflow 2: Global and Campaign-Specific Analytics π This second part of the workflow is triggered every 2 hours by the Schedule Trigger node. Its goal is to get a day-by-day overview of email engagement and campaign-specific performance. How it Works: Schedule Trigger: The Schedule Trigger node starts this workflow every 2 hours. ++API Requests:++ Two HTTP Request nodes, HTTP Request and HTTP Request1, call different Smartlead API endpoints. ++HTTP Request++ retrieves day-wise overall stats for email engagement. ++HTTP Request1 ++retrieves overall stats for each campaign. ++Data Splitting:++ The Split Out and Split Out1 nodes separate the JSON responses into individual data items for processing. ++Google Sheets Integration:++ The Append or update row in sheet and Append or update row in sheet1 nodes then write the data to Google Sheets. ++Append or update row in sheet++ updates the Sheet1 sheet with day-wise metrics, using the date as a matching column. ++Append or update row in sheet1++ updates the CampaignWise sheet with campaign performance metrics, using the campaign id to match rows.
by Oneclick AI Squad
AI-Driven Tax Compliance & Deadline Management System Description Automate tax deadline monitoring with AI-powered insights. This workflow checks your tax calendar daily at 8 AM, uses GPT-4 to analyze upcoming deadlines across multiple jurisdictions, detects overdue and critical items, and sends intelligent alerts via email and Slack only when immediate action is required. Perfect for finance teams and accounting firms who need proactive compliance management without manual tracking. ποΈπ€π Good to Know AI-Powered**: GPT-4 provides risk assessment and strategic recommendations Multi-Jurisdiction**: Handles Federal, State, and Local tax requirements automatically Smart Alerts**: Only notifies executives when deadlines are overdue or critical (β€3 days) Priority Classification**: Categorizes deadlines as Overdue, Critical, High, or Medium priority Dual Notifications**: Critical alerts to leadership + daily summaries to team channel Complete Audit Trail**: Logs all checks and deadlines to Google Sheets for compliance records How It Works Daily Trigger - Runs at 8:00 AM every morning Fetch Data - Pulls tax calendar and company configuration from Google Sheets Analyze Deadlines - Calculates days remaining, filters by jurisdiction/entity type, categorizes by priority AI Analysis - GPT-4 provides strategic insights and risk assessment on upcoming deadlines Smart Routing - Only sends alerts if overdue or critical deadlines exist Critical Alerts - HTML email to executives + Slack alert for urgent items Team Updates - Slack summary to finance channel with all upcoming deadlines Logging - Records compliance check results to Google Sheets for audit trail Requirements Google Sheets Structure Sheet 1: TaxCalendar DeadlineID | DeadlineName | DeadlineDate | Jurisdiction | Category | AssignedTo | IsActive FED-Q1 | Form 1120 Q1 | 2025-04-15 | Federal | Income | John Doe | TRUE Sheet 2: CompanyConfig (single row) Jurisdictions | EntityType | FiscalYearEnd Federal, California | Corporation | 12-31 Sheet 3: ComplianceLog (auto-populated) Date | AlertLevel | TotalUpcoming | CriticalCount | OverdueCount 2025-01-15 | HIGH | 12 | 3 | 1 Credentials Needed Google Sheets - Service Account OAuth2 OpenAI - API Key (GPT-4 access required) SMTP - Email account for sending alerts Slack - Bot Token with chat:write permission Setup Steps Import workflow JSON into n8n Add all 4 credentials Replace these placeholders: YOUR_TAX_CALENDAR_ID - Tax calendar sheet ID YOUR_CONFIG_ID - Company config sheet ID YOUR_LOG_ID - Compliance log sheet ID C12345678 - Slack channel ID tax@company.com - Sender email cfo@company.com - Recipient email Share all sheets with Google service account email Invite Slack bot to channels Test workflow manually Activate the trigger Customizing This Workflow Change Alert Thresholds: Edit "Analyze Deadlines" node: Critical: Change <= 3 to <= 5 for 5-day warning High: Change <= 7 to <= 14 for 2-week notice Medium: Change <= 30 to <= 60 for 2-month lookout Adjust Schedule: Edit "Daily Tax Check" trigger: Change hour/minute for different run time Add multiple trigger times for tax season (8 AM, 2 PM, 6 PM) Add More Recipients: Edit "Send Email" node: To: cfo@company.com, director@company.com CC: accounting@company.com BCC: archive@company.com Customize Email Design: Edit "Format Email" node to change colors, add logo, or modify layout Add SMS Alerts: Insert Twilio node after "Is Critical" for emergency notifications Integrate Task Management: Add HTTP Request node to create tasks in Asana/Jira for critical deadlines Troubleshooting | Issue | Solution | |-------|----------| | No deadlines found | Check date format (YYYY-MM-DD) and IsActive = TRUE | | AI analysis failed | Verify OpenAI API key and account credits | | Email not sending | Test SMTP credentials and check if critical condition met | | Slack not posting | Invite bot to channel and verify channel ID format | | Permission denied | Share Google Sheets with service account email | π Professional Services Need help with implementation or customization? Our team offers: π― Custom workflow development π’ Enterprise deployment support π Team training sessions π§ Ongoing maintenance π Custom reporting & dashboards π Additional API integrations Discover more workflows β Get in touch with us
by Rully Saputra
Automated SEO Watchlist: Continuous Audits Powered by Decodo, Gemini and Google Sheets Automate continuous SEO audits with Decodo and Gemini AI β live data, smart insights, and Google Sheets tracking with team alerts. Whoβs it for This workflow is designed for SEO specialists, marketing teams, agencies, and website owners who want an effortless, automated way to monitor SEO health. Itβs perfect for ongoing audits, content monitoring, and proactive SEO management β without the manual workload. How it works / What it does Every five days, the workflow: Reads a list of URLs from Google Sheets. Uses Decodo to fetch live on-page data β titles, meta descriptions, headings, schema, links, and Core Web Vitals. Passes that data to Gemini AI for an advanced SEO analysis and scoring based on key factors (content, metadata, links, speed, and structure). Parses results via a Structured Output Parser for clean JSON output. Stores findings in Google Sheets and sends a Telegram alert when the audit completes. Why Decodo matters Decodo is the backbone of this workflow. It powers the real-time page inspection, ensuring Gemini AI has complete, accurate data to analyze. Decodo transforms static audits into live, intelligent monitoring β making your SEO insights far more actionable and reliable. How to set up Connect your Decodo API credentials. Add your Google Sheets URL list. Configure your Telegram bot credentials. Enable the workflow β it runs automatically every 5 days. Requirements Decodo API credentials Google Sheets OAuth connection Telegram Bot token n8n instance (Cloud or Self-hosted) How to customize the workflow Change the trigger interval in the Schedule Trigger node. Modify the SEO Analyzer (LLM Chain) weights for different scoring. Extend the Store Result node to integrate with dashboards or databases. Adjust the AI prompt for additional SEO checks (e.g., backlinks, readability, image optimization). β Highlights Automated SEO auditing Real-time data from Decodo Smart analysis powered by Gemini AI Structured reporting in Google Sheets Team notifications via Telegram
by vinci-king-01
Social Media Sentiment Analysis Dashboard with AI and Real-time Monitoring π― Target Audience Social media managers and community managers Marketing teams monitoring brand reputation PR professionals tracking public sentiment Customer service teams identifying trending issues Business analysts measuring social media ROI Brand managers protecting brand reputation Product managers gathering user feedback π Problem Statement Manual social media monitoring is overwhelming and often misses critical sentiment shifts or trending topics. This template solves the challenge of automatically collecting, analyzing, and visualizing social media sentiment data across multiple platforms to provide actionable insights for brand management and customer engagement. π§ How it Works This workflow automatically monitors social media platforms using AI-powered sentiment analysis, processes mentions and conversations, and provides real-time insights through a comprehensive dashboard. Key Components Scheduled Trigger - Runs the workflow at specified intervals to maintain real-time monitoring AI-Powered Sentiment Analysis - Uses advanced NLP to analyze sentiment, emotions, and topics Multi-Platform Integration - Monitors Twitter, Reddit, and other social platforms Real-time Alerting - Sends notifications for critical sentiment changes or viral content Dashboard Integration - Stores all data in Google Sheets for comprehensive analysis and reporting π Google Sheets Column Specifications The template creates the following columns in your Google Sheets: | Column | Data Type | Description | Example | |--------|-----------|-------------|---------| | timestamp | DateTime | When the mention was recorded | "2024-01-15T10:30:00Z" | | platform | String | Social media platform | "Twitter" | | username | String | User who posted the content | "@john_doe" | | content | String | Full text of the post/comment | "Love the new product features!" | | sentiment_score | Number | Sentiment score (-1 to 1) | 0.85 | | sentiment_label | String | Sentiment classification | "Positive" | | emotion | String | Primary emotion detected | "Joy" | | topics | Array | Key topics identified | ["product", "features"] | | engagement | Number | Likes, shares, comments | 1250 | | reach_estimate | Number | Estimated reach | 50000 | | influence_score | Number | User influence metric | 0.75 | | alert_priority | String | Alert priority level | "High" | π οΈ Setup Instructions Estimated setup time: 20-25 minutes Prerequisites n8n instance with community nodes enabled ScrapeGraphAI API account and credentials Google Sheets account with API access Slack workspace for notifications (optional) Social media API access (Twitter, Reddit, etc.) Step-by-Step Configuration 1. Install Community Nodes Install required community nodes npm install n8n-nodes-scrapegraphai npm install n8n-nodes-slack 2. Configure ScrapeGraphAI Credentials Navigate to Credentials in your n8n instance Add new ScrapeGraphAI API credentials Enter your API key from ScrapeGraphAI dashboard Test the connection to ensure it's working 3. Set up Google Sheets Connection Add Google Sheets OAuth2 credentials Grant necessary permissions for spreadsheet access Create a new spreadsheet for sentiment analysis data Configure the sheet name (default: "Sentiment Analysis") 4. Configure Social Media Monitoring Update the websiteUrl parameters in ScrapeGraphAI nodes Add URLs for social media platforms you want to monitor Customize the user prompt to extract specific sentiment data Set up keywords, hashtags, and brand mentions to track 5. Set up Notification Channels Configure Slack webhook or API credentials Set up email service credentials for alerts Define sentiment thresholds for different alert levels Test notification delivery 6. Configure Schedule Trigger Set monitoring frequency (every 15 minutes, hourly, etc.) Choose appropriate time zones for your business hours Consider social media platform rate limits 7. Test and Validate Run the workflow manually to verify all connections Check Google Sheets for proper data formatting Test sentiment analysis with sample content π Workflow Customization Options Modify Monitoring Targets Add or remove social media platforms Change keywords, hashtags, or brand mentions Adjust monitoring frequency based on platform activity Extend Sentiment Analysis Add more sophisticated emotion detection Implement topic clustering and trend analysis Include influencer identification and scoring Customize Alert System Set different thresholds for different sentiment levels Create tiered alert systems (info, warning, critical) Add sentiment trend analysis and predictions Output Customization Add data visualization and reporting features Implement sentiment trend charts and graphs Create executive dashboards with key metrics Add competitor sentiment comparison π Use Cases Brand Reputation Management**: Monitor and respond to brand mentions Crisis Management**: Detect and respond to negative sentiment quickly Customer Feedback Analysis**: Understand customer satisfaction and pain points Product Launch Monitoring**: Track sentiment around new product releases Competitor Analysis**: Monitor competitor sentiment and engagement Influencer Identification**: Find and engage with influential users π¨ Important Notes Respect social media platforms' terms of service and rate limits Implement appropriate delays between requests to avoid rate limiting Regularly review and update your monitoring keywords and parameters Monitor API usage to manage costs effectively Keep your credentials secure and rotate them regularly Consider privacy implications and data protection regulations π§ Troubleshooting Common Issues: ScrapeGraphAI connection errors: Verify API key and account status Google Sheets permission errors: Check OAuth2 scope and permissions Sentiment analysis errors: Review the Code node's JavaScript logic Rate limiting: Adjust monitoring frequency and implement delays Alert delivery failures: Check notification service credentials Support Resources: ScrapeGraphAI documentation and API reference n8n community forums for workflow assistance Google Sheets API documentation for advanced configurations Social media platform API documentation Sentiment analysis best practices and guidelines