by Rahi
Workflow 1: Domain and Email Health 🩺 This part of the workflow is triggered every 5 hours by the Schedule Trigger1 node. Its purpose is to pull health metrics for both email domains and individual email addresses. How it Works: ++Schedule Trigger:++ The Schedule Trigger1 node initiates the workflow every 5 hours. ++API Requests:++ Two separate HTTP Request nodes, HTTP Request5 and HTTP Request6, make API calls to Smartlead. ++HTTP Request5 calls++ the endpoint for domain-wise health metrics. ++HTTP Request6 calls++ the endpoint for email-wise health metrics. Both requests use the same api_key and a date range from 2025-07-04 to the current day. ++Data Splitting:++ The Split Out5 and Split Out6 nodes take the JSON response from the API calls and split the data into individual items. This is necessary so each row of data can be processed and added to Google Sheets separately. ++Google Sheets Integration:++ Finally, the Append or update row in sheet5 and Append or update row in sheet6 nodes update two different Google Sheets: ++Append or update row in sheet5 adds++ or updates rows in the DomainHealth sheet, matching on the domain column. ++Append or update row in sheet6 adds++ or updates rows in the EmailHealth sheet, matching on the from_email column. Workflow 2: Global and Campaign-Specific Analytics 📊 This second part of the workflow is triggered every 2 hours by the Schedule Trigger node. Its goal is to get a day-by-day overview of email engagement and campaign-specific performance. How it Works: Schedule Trigger: The Schedule Trigger node starts this workflow every 2 hours. ++API Requests:++ Two HTTP Request nodes, HTTP Request and HTTP Request1, call different Smartlead API endpoints. ++HTTP Request++ retrieves day-wise overall stats for email engagement. ++HTTP Request1 ++retrieves overall stats for each campaign. ++Data Splitting:++ The Split Out and Split Out1 nodes separate the JSON responses into individual data items for processing. ++Google Sheets Integration:++ The Append or update row in sheet and Append or update row in sheet1 nodes then write the data to Google Sheets. ++Append or update row in sheet++ updates the Sheet1 sheet with day-wise metrics, using the date as a matching column. ++Append or update row in sheet1++ updates the CampaignWise sheet with campaign performance metrics, using the campaign id to match rows.
by NanaB
This n8n workflow provides a comprehensive solution for user authentication and management, leveraging Airtable as the backend database. It includes flows for user sign-up and login, aswell as the sample crud operations retrieving user details, and updating user information. Youtube Video of me explaining the flow: https://www.youtube.com/watch?v=gKcGfyq3dPM How it Works User Sign-Up Flow Receives POST request: A webhook listens for POST requests containing new user details (email, first name, last name, password). Checks for existing email: The workflow queries Airtable to see if the submitted email already exists. Handles email in use: If the email is found, it responds with {"response": "email in use"}. Creates new user: If the email is unique, the password is SHA256 hashed (Base64 encoded), and the user's information (including the hashed password) is stored in Airtable. A successful response of {"response": "success"} is then sent. User Login Flow Receives POST request: A webhook listens for POST requests with user email and password for login. Verifies user existence: It checks Airtable for a user with the provided email. If no user is found, it responds with a failure message ("wrong email"). Compares passwords: If a user is found, the submitted password is hashed (SHA256, Base64 encoded) and compared with the stored hashed password in Airtable. Responds with JWT or error: If passwords match, a JWT token containing the user's ID and email is issued. If they don't match, a "wrong password" response is sent. Flows for a Logged-In User These flows require a JWT-authenticated request. Get User Details:** Webhook (GET): Receives a JWT-authenticated request. Airtable (Read): Fetches the current user’s record using the jwtPayload.id. Set Node ("Specify Current Details"): Maps fields like "First Name," "Last Name," "Email," and "Date" from Airtable to a standard output format. Update User Details:** Webhook (POST): Receives updated user data (email, name, password). Airtable (Upsert): Updates the record matching jwtPayload.id using the submitted fields. Set Node ("Specify New Details"): Outputs the updated data in a standard format. Set Up Steps (Approx. 5 Minutes) Step 1: Set up your Airtable Base and Table You'll need an Airtable Base and a table to store your user data. Ensure your table has at least the following columns: Email** (Single Line Text) First Name** (Single Line Text) Last Name** (Single Line Text) Password** (Single Line Text - this will store the hashed password) Date** (Date - optional, for user sign-up date) Step 2: Obtain an Airtable Personal Access Token Go to the Airtable website and log in to your account. Navigate to your personal access token page (usually found under your developer settings or by searching for "personal access tokens"). Click "Create new token." Give your token a name (e.g., "n8n User Management"). Grant necessary permissions: Scope: data.records:read, data.records:write for the specific base you will be using. Base: Select the Airtable base where your user management table resides. Generate the token and copy it immediately. You won't be able to see it again. Store it securely. Step 3: Create a JWT Auth Credential in n8n In your n8n instance, go to "Credentials" (usually found in the left-hand sidebar). Click "New Credential" and search for "JWT Auth". Give the credential a name (e.g., "UserAuthJWT"). For the "Signing Secret," enter a strong, random string of characters. This secret will be used to sign and verify your JWT tokens. Keep this secret highly confidential. Save the credential. Customization Options This workflow is designed to be highly adaptable: Database Integration**: Easily switch from Airtable to other databases like PostgreSQL, MySQL, MongoDB, or even Google Sheets by replacing the Airtable nodes with the appropriate database nodes in n8n. Authentication Methods**: Extend the authentication to include multi-factor authentication (MFA), social logins (Google, Facebook), or integrate with existing identity providers (IdP) by adding additional nodes. User Profile Fields**: Add or remove user profile fields (e.g., phone number, address, user roles) by adjusting the Airtable table columns and the Set nodes in the workflow. Notification System**: Integrate notification systems (e.g., email, SMS) for events like new user sign-ups, password resets, or account changes. Admin Panel**: Build an admin panel using n8n to manage users directly, including functionalities for adding, deleting, or updating user records, and resetting passwords. This workflow provides a solid foundation for building robust user management systems, adaptable to a wide range of applications and security requirements. Need Assistance or Customization? Do you have specific integrations in mind, or are you looking to add more user management features to this workflow? If you need help setting this up, or want to adapt it for a unique use case, don't hesitate to reach out! You can contact me directly at nanabrownsnr@gmail.com. I'd be glad to assist you.
by Hugues Stock
What does this template do? This workflow sets a small "lock" value in Redis so that only one copy of a long job can run at the same time. If another trigger fires while the job is still busy, the workflow sees the lock, stops early, and throws a clear error. This protects your data and keeps you from hitting rate limits. Because the workflow also stores simple progress flags ("working", "loading", "finishing"), you can poll the current status and show live progress for very long jobs. Use Case Great when the same workflow can be called many times in parallel (for example by webhooks, cron jobs, or nested Execute Workflow calls) and you need an "only run once at a time" guarantee without building a full queue system. What the Workflow Does ⚡ Starts through Execute Workflow Trigger called by another workflow 🔄 A Switch sends the run to Get, Set, or Unset actions 💾 Redis reads or writes a key named process_status_<key> with a time‑to‑live (default 600 s) 🚦 If nodes check the key and decide to continue or stop ⏱️ Wait nodes stand in for the slow part of your job (replace these with your real work) 📈 Updates the key with human‑readable progress values that another workflow can fetch with action = get 🏁 When done, the lock is removed so the next run can start Apps & Services Used Redis Core n8n nodes (Switch, If, Set, Wait, Stop and Error) Pre‑requisites A Redis server that n8n can reach Redis credentials stored in n8n A second workflow that calls this one and sends: action set to get, set, or unset key set to a unique name for the job Optional timeout in seconds Customization Tips Increase or decrease the TTL in the Set Timeout node to match how long your job usually runs Add or rename status values ("working", "loading", "finishing", and so on) to show finer progress Replace Stop and Error with a Slack or email alert, or even push the extra trigger into a queue if you prefer waiting instead of failing Use different Redis keys if you need separate locks for different tasks Build a small "status endpoint" workflow that calls this one with action = get to display real‑time progress to users Additional Use Cases 🛑 Telegram callback spam filter If a Telegram bot sends many identical callbacks in a burst, call this workflow first to place a lock. Only the first callback will proceed; the rest will exit cleanly until the lock clears. This keeps your bot from flooding downstream APIs. 🧩 External API rate‑limit protection Run heavy API syncs one after the other so parallel calls do not break vendor rate limits. 🔔 Maintenance window lock Block scheduled maintenance tasks from overlapping, making sure each window finishes before the next starts.
by Grigory Frolov
WordPress Blog to Google Sheets Sync Posts • Categories • Tags • Media 🧩 Overview This n8n workflow automatically syncs your WordPress website content — including posts, categories, tags, and media — into Google Sheets. It helps automate content reporting, SEO analysis, and data backups. The workflow can run on schedule or on demand via a webhook. 💡 Use cases Maintain a live database of blog posts in Google Sheets. Create dashboards in Google Data Studio or Looker Studio. Track new articles for newsletters or social media scheduling. Backup all WordPress content and media outside of your CMS. ⚙️ Prerequisites Before importing the workflow, ensure you have: A WordPress website with the REST API enabled (default in WP 4.7+). Authentication: either Application Passwords or Basic Auth credentials. A Google Sheet with the following tabs: Posts Categories Tags Media The following credentials configured in n8n: HTTP Basic Auth (for WordPress) Google Sheets OAuth2 🚀 Setup instructions Import the workflow into your n8n instance. Replace all example WordPress API URLs with your domain, for example: https://yourdomain.com/wp-json/wp/v2/ Connect your HTTP Basic Auth credentials (WordPress username + Application Password). Connect your Google Sheets OAuth2 account. Update the spreadsheet ID in each Google Sheets node with your own. Adjust the Schedule Trigger (e.g. run daily at 2:00 AM). Run once manually to verify data sync. 🧠 Workflow structure | Section | Description | |----------|--------------| | Schedule / Webhook Trigger | Starts the workflow manually or automatically | | Variables & Loop Vars | Initialize pagination for REST API requests | | Get Posts → Split Out → Update Posts | Fetch and update all WordPress posts | | Get Categories → Update Categories | Sync WordPress categories | | Get Tags → Update Tags | Sync WordPress tags | | Get Media → Split Out → Update Media | Sync media library (images, videos, etc.) | | IF Loops | Handles pagination logic until all items are retrieved | ⚠️ Notes & Limitations Works with standard WordPress REST API endpoints only. Custom post types require editing endpoint URLs. The per_page value defaults to 10; increase for faster syncs. For large sites, consider increasing n8n memory or adding execution logs. Avoid running the workflow too frequently to prevent API rate limits. 🎥 Video Tutorial A step-by-step setup guide is available here: 👉 https://www.youtube.com/watch?v=czSMWyD6f-0 Please subscribe to my YouTube channel to support me: 👉 https://www.youtube.com/@gregfrolovpersonal 👨💻 Author Created by: Grigory Frolov SEO & Automation Specialist — helping businesses integrate WordPress, AI, and data tools with n8n. 🧾 License This workflow is provided under the MIT License. Feel free to use, modify, and share improvements with the community.
by Shrey
This workflow can be used to save all of your workflows in: a raw state (as a json file in Dropbox) an Airtable base, in a pre-designed format. It runs periodically (currently, every 30 minutes) and either updates (if already existing in Airtable) or creates a new record in Airtable for each workflow. Here's the Airtable base to give you an idea: View Airtable base Note: This workflows uses the "http://localhost:5678/rest" API which the UI editor uses but is still not officially supported. Hence, it may suffer breaking changes at some point in the future and the workflow might become dysfunctional then.
by Oneclick AI Squad
AI-Driven Tax Compliance & Deadline Management System Description Automate tax deadline monitoring with AI-powered insights. This workflow checks your tax calendar daily at 8 AM, uses GPT-4 to analyze upcoming deadlines across multiple jurisdictions, detects overdue and critical items, and sends intelligent alerts via email and Slack only when immediate action is required. Perfect for finance teams and accounting firms who need proactive compliance management without manual tracking. 🏛️🤖📊 Good to Know AI-Powered**: GPT-4 provides risk assessment and strategic recommendations Multi-Jurisdiction**: Handles Federal, State, and Local tax requirements automatically Smart Alerts**: Only notifies executives when deadlines are overdue or critical (≤3 days) Priority Classification**: Categorizes deadlines as Overdue, Critical, High, or Medium priority Dual Notifications**: Critical alerts to leadership + daily summaries to team channel Complete Audit Trail**: Logs all checks and deadlines to Google Sheets for compliance records How It Works Daily Trigger - Runs at 8:00 AM every morning Fetch Data - Pulls tax calendar and company configuration from Google Sheets Analyze Deadlines - Calculates days remaining, filters by jurisdiction/entity type, categorizes by priority AI Analysis - GPT-4 provides strategic insights and risk assessment on upcoming deadlines Smart Routing - Only sends alerts if overdue or critical deadlines exist Critical Alerts - HTML email to executives + Slack alert for urgent items Team Updates - Slack summary to finance channel with all upcoming deadlines Logging - Records compliance check results to Google Sheets for audit trail Requirements Google Sheets Structure Sheet 1: TaxCalendar DeadlineID | DeadlineName | DeadlineDate | Jurisdiction | Category | AssignedTo | IsActive FED-Q1 | Form 1120 Q1 | 2025-04-15 | Federal | Income | John Doe | TRUE Sheet 2: CompanyConfig (single row) Jurisdictions | EntityType | FiscalYearEnd Federal, California | Corporation | 12-31 Sheet 3: ComplianceLog (auto-populated) Date | AlertLevel | TotalUpcoming | CriticalCount | OverdueCount 2025-01-15 | HIGH | 12 | 3 | 1 Credentials Needed Google Sheets - Service Account OAuth2 OpenAI - API Key (GPT-4 access required) SMTP - Email account for sending alerts Slack - Bot Token with chat:write permission Setup Steps Import workflow JSON into n8n Add all 4 credentials Replace these placeholders: YOUR_TAX_CALENDAR_ID - Tax calendar sheet ID YOUR_CONFIG_ID - Company config sheet ID YOUR_LOG_ID - Compliance log sheet ID C12345678 - Slack channel ID tax@company.com - Sender email cfo@company.com - Recipient email Share all sheets with Google service account email Invite Slack bot to channels Test workflow manually Activate the trigger Customizing This Workflow Change Alert Thresholds: Edit "Analyze Deadlines" node: Critical: Change <= 3 to <= 5 for 5-day warning High: Change <= 7 to <= 14 for 2-week notice Medium: Change <= 30 to <= 60 for 2-month lookout Adjust Schedule: Edit "Daily Tax Check" trigger: Change hour/minute for different run time Add multiple trigger times for tax season (8 AM, 2 PM, 6 PM) Add More Recipients: Edit "Send Email" node: To: cfo@company.com, director@company.com CC: accounting@company.com BCC: archive@company.com Customize Email Design: Edit "Format Email" node to change colors, add logo, or modify layout Add SMS Alerts: Insert Twilio node after "Is Critical" for emergency notifications Integrate Task Management: Add HTTP Request node to create tasks in Asana/Jira for critical deadlines Troubleshooting | Issue | Solution | |-------|----------| | No deadlines found | Check date format (YYYY-MM-DD) and IsActive = TRUE | | AI analysis failed | Verify OpenAI API key and account credits | | Email not sending | Test SMTP credentials and check if critical condition met | | Slack not posting | Invite bot to channel and verify channel ID format | | Permission denied | Share Google Sheets with service account email | 📞 Professional Services Need help with implementation or customization? Our team offers: 🎯 Custom workflow development 🏢 Enterprise deployment support 🎓 Team training sessions 🔧 Ongoing maintenance 📊 Custom reporting & dashboards 🔗 Additional API integrations Discover more workflows – Get in touch with us
by sato rio
This workflow streamlines the entire inventory replenishment process by leveraging AI for demand forecasting and intelligent logic for supplier selection. It aggregates data from multiple sources—POS systems, weather forecasts, SNS trends, and historical sales—to predict future demand. Based on these predictions, it calculates shortages, requests quotes from multiple suppliers, selects the optimal vendor based on cost and lead time, and executes the order automatically. 🚀 Who is this for? Retail & E-commerce Managers** aiming to minimize stockouts and reduce overstock. Supply Chain Operations** looking to automate procurement and vendor selection. Data Analysts** wanting to integrate external factors (weather, trends) into inventory planning. 💡 How it works Data Aggregation: Fetches data from POS systems, MySQL (historical sales), OpenWeatherMap (weather), and SNS trend APIs. AI Forecasting: Formats the data and sends it to an AI prediction API to forecast demand for the next 7 days. Shortage Calculation: Compares the forecast against current stock and safety stock to determine necessary order quantities. Supplier Optimization: For items needing replenishment, the workflow requests quotes from multiple suppliers (A, B, C) in parallel. It selects the best supplier based on the lowest total cost within a 7-day lead time. Execution & Logging: Places the order via API, updates the inventory system, and logs the transaction to MySQL. Anomaly Detection: If the AI's confidence score is low, it skips the auto-order and sends an alert to Slack for manual review. ⚙️ Setup steps Configure Credentials: Set up credentials for MySQL and Slack in n8n. API Keys: You will need an API key for OpenWeatherMap (or a similar service). Update Endpoints: The HTTP Request nodes use placeholder URLs (e.g., pos-api.example.com, ai-prediction-api.example.com). Replace these with your actual internal APIs, ERP endpoints, or AI service (like OpenAI). Database Prep: Ensure your MySQL database has a table named forecast_order_log to store the order history. Schedule: The workflow is set to run daily at 03:00. Adjust the Schedule Trigger node as needed. 📋 Requirements n8n** (Self-hosted or Cloud) MySQL** database Slack** workspace External APIs for POS, Inventory, and Supplier communication (or mock endpoints for testing).
by Ranjan Dailata
This n8n workflow automates backlink monitoring, analysis, and AI-driven interpretation for any domain or URL. It combines backlink intelligence from SE Ranking with structured reasoning and summarization powered by OpenAI GPT 4.1-mini. Instead of manually reviewing backlink reports, this workflow transforms raw backlink metrics into clear, human-readable SEO insights and persists them to multiple storage layers for reporting and tracking. Who this is for? This workflow is ideal for: SEO professionals and technical SEO teams Digital marketing agencies managing multiple domains Growth and content teams tracking backlink quality Developers building SEO intelligence pipelines Data teams using n8n for enrichment and reporting What this workflow does? Accepts a backlink query (domain, host, or URL) Uses multiple SE Ranking Backlinks API endpoints to retrieve: Backlink summary metrics Referring domains, IPs, and subnets Authority and backlink quality indicators Raw backlink lists Routes the data through an AI Agent powered by GPT-4.1-mini that: Selects the appropriate backlink dataset automatically Normalizes noisy SEO data Generates structured summaries without subjective opinions Produces a clean backlink intelligence summary Persists results to: n8n DataTables Google Sheets CSV / JSON exports Setup If you are new to SE Ranking, please signup on https://seranking.com Prerequisites Active SE Ranking API access OpenAI API key with GPT-4.1-mini enabled n8n instance (self-hosted or cloud) Basic understanding of backlink and authority metrics Import the workflow JSON into n8n Configure credentials: SE Ranking** using HTTP Header Authentication. Please make sure to set the header authentication as below. The value should contain a Token followed by a space with the SE Ranking API Key. OpenAI API (GPT-4.1-mini) Google Sheets OAuth (optional, for reporting) Open the Set Input Fields node and define: query (e.g. Backlinks Summary for https://example.com) Verify storage destinations: Google Sheet ID and sheet name n8n DataTable File export nodes (CSV / JSON) Click Execute Workflow How to customize this workflow to your needs? You can easily extend or adapt this workflow by: Switching analysis mode (domain, host, or URL) Adding historical backlink trend analysis Enhancing the AI prompt to generate: Toxic backlink alerts Link-building opportunities Competitor backlink gap analysis Replacing storage with: Databases or data warehouses Slack / Email notifications BI dashboards Scheduling the workflow for continuous backlink monitoring Summary This n8n template delivers an end-to-end backlink intelligence system from raw backlink retrieval to AI-powered interpretation and structured storage. By combining SE Ranking’s backlink data with OpenAI-driven reasoning, it eliminates manual SEO analysis and enables scalable, repeatable backlink monitoring.
by Chandan Singh
This workflow creates a daily, automated backup of all workflows in a self-hosted n8n instance and stores them in Google Drive. Instead of exporting every workflow on every run, it uses content hashing to detect meaningful changes and only updates backups when a workflow has actually been modified. To keep Google Drive clean and predictable, the workflow intentionally deletes the existing backup file before uploading the updated version. This avoids duplicate files and ensures there is always one authoritative backup per workflow. A Data Table is used as an index to track workflow IDs, hash values, and timestamps. This allows the workflow to quickly determine whether a workflow already exists, whether its content has changed, or whether it should be skipped entirely. How it works Runs daily using a Cron Trigger. Fetches all workflows from the n8n API. Processes workflows one-by-one for reliability. Generates a SHA-256 hash for each workflow. Compares hashes against a stored Data Table. Deletes existing Google Drive backups when changes are detected. Uploads updated workflows and skips unchanged ones. Store new or updated workflows details in Data Table. Filters workflows based on the configured backup scope (all | active | tagged ). Backs up all workflows, only active workflows, or only workflows matching a specific tag. Applies the scope filter before hashing and comparison, ensuring only relevant workflows are processed. Setup steps Set the Cron schedule** Open the Cron Trigger node and choose the time you want the backup to run (for example, once daily during off-peak hours). Create a Data Table** Create a new n8n Data Table with the title defined in dataTableTitle. This table stores workflowId, workflowName, hashCode, and DriveFiveId. Configure the Set node** In the Set Backup Configuration node, provide the following values: { "n8nHost": "https://your-n8n-domain", "apiKey": "your-n8n-api-key", "backupFolder": "/n8n/workflow-backups", "hashAlgorithm": "sha256", "dataTableTitle": "n8n_workflow_backup_index", "backupScope" : "", "requiredTag" : "" } In the Set Backup Configuration node, choose how workflows should be selected for backup: all – backs up every workflow (default) active – backs up only enabled workflows tagged – backs up only workflows containing a specific tag If using the tagged option, provide the required tag name to match. { "backupScope": "tagged", "requiredTag": "production" } Connect Google Drive credentials** Authorize your Google Drive account and ensure the backup folder exists. Activate the workflow** Once enabled, backups run automatically with no further action required.
by rana tamure
This n8n workflow automates the creation of high-quality, SEO-optimized blog posts using AI. It pulls keyword data from Google Sheets, conducts research via Perplexity AI, generates structured content (title, introduction, key takeaways, body, conclusion, and FAQs) with OpenAI and Anthropic models, assembles the post, performs final edits, converts to HTML, and publishes directly to WordPress. Ideal for content marketers, bloggers, or agencies looking to scale content production while maintaining relevance and engagement. Key Features Keyword-Driven Generation: Fetches primary keywords, search intent, and related terms from a Google Sheets spreadsheet to inform content strategy. AI Research & Structuring: Uses Perplexity for in-depth topic research and OpenAI/Anthropic for semantic analysis, outlines, and full content drafting. Modular Content Creation: Generates sections like introductions, key takeaways, outlines, body, conclusions, and FAQs with tailored prompts for tone, style, and SEO. Assembly & Editing: Combines sections into a cohesive Markdown post, adds internal/external links, and applies final refinements for readability and flow. Publishing Automation: Converts Markdown to styled HTML and posts drafts to WordPress. Customization Points: Easily adjust AI prompts, research depth, or output formats via Code and Set nodes. Requirements Credentials: OpenAI API (for GPT models), Perplexity API (for research), Google Sheets OAuth2 (for keyword input), WordPress API (for publishing). Setup: Configure your Google Sheets with columns like "keyword", "search intent", "related keyword", etc. Ensure the sheet is shared with your Google account. Dependencies: No additional packages needed; relies on n8n's built-in nodes for AI, HTTP, and data processing. How It Works Trigger & Input: Start manually or schedule; pulls keyword data from Google Sheets. Research Phase: Uses Perplexity to gather topic insights and citations from reputable sources. Content Generation: AI nodes create title, structure, intro, takeaways, outline, body, conclusion, and FAQs based on research and SEO guidelines. Assembly & Refinement: Merges sections, embeds links, edits for polish, and converts to HTML. Output: Publishes as a WordPress draft or outputs the final HTML for manual use. Benefits Time Savings: Automate 80-90% of content creation, reducing manual writing from hours to minutes. SEO Optimization: Incorporates primary/related keywords naturally, aligns with search intent, and includes semantic structures for better rankings. Scalability: Process multiple keywords in batches; perfect for content calendars or high-volume blogging. Quality Assurance: Built-in editing ensures engaging, error-free content with real-world examples and data-backed insights. Versatility: Adaptable for any niche (e.g., marketing, tech, finance) by tweaking prompts or sheets. Potential Customizations Add more AI models (e.g., via custom nodes) for varied tones. Integrate image generation or social sharing for full content pipelines. Filter sheets for specific topics or add notifications on completion.
by gotoHuman
Auto-detect news from n8n and turn into a human-approved LinkedIn post. gotoHuman is used to keep a human in the loop. There you can manually edit the AI draft of the post or request to regenerate it. How it works The workflow is triggered each day to fetch the latest version of https://blog.n8n.io. It then fetches each article, checks if it was published in the last 24 hours and uses an LLM to summarize it. An LLM then drafts a related LinkedIn post which is sent to gotoHuman for approval. In gotoHuman, the reviewer can manually edit it or ask to regenerate it with the option to even edit the prompt (Retries loop back to the AI Draft LinkedIn Post node) Approved Posts are automatically published to LinkedIn How to set up Most importantly, install the gotoHuman node before importing this template! (Just add the node to a blank canvas before importing) Set up your credentials for gotoHuman, OpenAI, and LinkedIn In gotoHuman, select and create the pre-built review template "Blog scraper agent" or import the ID: sMxevC9tSAgdfWsr6XIW Select this template in the gotoHuman node Requirements You need accounts for gotoHuman (human supervision) OpenAI (summary, draft) LinkedIn How to customize Change the blog URL to monitor. Adapt to its' HTML structure Provide the AI Draft LinkedIn Post with examples of previous posts so it picks up your writing style (consider adding gotoHuman's dataset of approved examples) Use the workflow to target other publications, like your newsletter, blog or other socials
by Sk developer
📊 Automated Website Traffic Tracker with Google Sheets Logging Track website traffic and backlinks effortlessly using the Website Traffic Checker - Ahref API. This n8n workflow automates data retrieval and logging into Google Sheets, making it perfect for SEO professionals and digital marketers. 🧩 What This Workflow Does (Summary) Accepts a domain via a simple web form. Sends the domain to Website Traffic Checker - Ahref API. If successful: Extracts backlink and traffic data. Appends the results to two separate Google Sheets. If failed: Sends an email alert with domain and status code. 🔧 Node-by-Node Explanation | Node | Purpose | | ---------------------------------- | ---------------------------------------------------------------------------------------------------------------- | | 🟢 Form Trigger | Starts the workflow when a domain is submitted via form. | | 🟩 Set Domain Value | Stores the submitted domain into a variable. | | 🌐 HTTP Request | Calls Website Traffic Checker - Ahref API. | | ✅ IF Node | Checks if the API responded with statusCode = 200. | | ❌ Email Node (Fail) | Sends an alert email if API fails. | | 📦 Code (Backlink Info) | Extracts backlink data from API response. | | 📄 Google Sheet: Backlink Info | Appends backlink data to a sheet. | | 📦 Code (Traffic Info) | Extracts traffic data from API response. | | 📄 Google Sheet: Traffic Data | Appends traffic metrics to another sheet. | 📁 Google Sheet Columns Backlink Info Sheet | Column | Description | | ------------------ | --------------------------- | | website | Domain submitted | | ascore | Authority score | | referring domain | Number of referring domains | | total backlinks | Total backlinks | Traffic Data Sheet | Column | Description | |----------------------|---------------------------------------------| | accuracy | Accuracy level of the traffic data | | bounce_rate | Bounce rate percentage | | desktop_share | Percentage of traffic from desktop devices | | direct | Direct traffic sources | | display_ad | Display ad traffic sources | | display_date | Date when traffic data was captured | | mail | Traffic from email campaigns | | mobile_share | Percentage of traffic from mobile devices | | pages_per_visit | Average number of pages per visit | | paid | Paid traffic sources | | prev_bounce_rate | Bounce rate in the previous period | | prev_direct | Previous period's direct traffic | | prev_display_ad | Previous period's display ad traffic | | prev_mail | Previous period's email traffic | | prev_pages_per_visit | Previous period's pages per visit | | prev_referral | Previous period's referral traffic | | prev_search_organic | Previous organic search traffic | | prev_search_paid | Previous paid search traffic | | prev_social_organic | Previous organic social traffic | | prev_social_paid | Previous paid social traffic | | prev_time_on_site | Previous time spent on site | | prev_users | Number of users in the previous period | | prev_visits | Visits in the previous period | | rank | Global rank of the website | | referral | Referral traffic | | search | Total search traffic | | search_organic | Organic search traffic | | search_paid | Paid search traffic | | social | Total social traffic | | social_organic | Organic social traffic | | social_paid | Paid social traffic | | target | Targeted country or demographic | | time_on_site | Average time spent on site | | unknown_channel | Traffic from unknown sources | | users | Number of unique users | | visits | Total number of visits | 🔐 How to Configure 🔑 Get API Key Go to Website Traffic Checker - Ahref API on RapidAPI. Sign in or create a free RapidAPI account. Subscribe to the API plan. Copy your x-rapidapi-key from the Endpoints tab. 📝 Add Key in n8n Go to your HTTP Request node. Under Headers, set: x-rapidapi-host = website-traffic-checker-ahref.p.rapidapi.com x-rapidapi-key = your API key 📄 How to Setup Google Sheets in n8n Connect a Google account via Google Sheets credentials in n8n. Use the full Google Sheet URL in the documentId field. Set correct Sheet name or GID (e.g., "Traffic Data"). Use Auto Map or Custom Map to define columns. > Make sure your Google Sheet has edit access and headers already created. 🧠 Use Case & Benefits 👤 Ideal For: SEO analysts Digital marketers Agencies managing multiple clients Web analytics consultants ✅ Benefits: Fully automated data collection. No manual copy-paste** from tools. Real-time insights delivered to Google Sheets. Easy monitoring of backlinks and traffic trends.