by Łukasz
Who is it for This workflow is for anyone who is using N8N. It's especially helpful if you are a DevOps and your N8N instance is self hosted. If you carea lot about security and number of failed executions and at the same time you are using InfluxDB to monitor status of your systems, this will perfectly fit in your stack. How it works This automation is fairly simple. It uses native N8N nodes to gather data from itself. Then it is parsing this data to be compatible with InfluxDB input. And finally it is sending this data to InfluxDB for further processing. Remember to set up Setup is really simple and you just need to provide just three variables. First is your InfluxDB URL, second is your InfluxDB organization, and third is your InfluxDB bucket name. Of course, to set up N8N nodes and gather data from them, you will need your instance API key. And that's all. How it looks in InfluxDB? See below Schedule Audits Audits don't need to be run often, but I would recommend it to be run on regular basis. This way you can see real data series in InfluxDB. I think that once a day should be enough, but it depends on your N8N usage of course Thank you, perfect! Glad I could help. Visit my profile for other automations for businesses. And if you are looking for dedicated software development, do not hesitate to reach out! You can also see automations on my Sailing Byte's GitHub N8N repository.
by Mohamed Abdelwahab
Automates the process of generating, storing, and publishing engaging LinkedIn posts derived from books (PDFs) using AI and vector search. 🧠 Overview This workflow: Watches a Google Drive folder for new or updated book PDFs. Extracts and embeds the content using OpenAI. Stores the data in a Pinecone vector database. Uses a LangChain agent to generate post ideas. Creates concise LinkedIn posts with hook, insight, CTA. Updates a Google Sheet and posts to LinkedIn. 🛠 Workflow Breakdown 📥 1. Google Drive Trigger Trigger:** Watches a folder for new or updated PDF files. Action:** Downloads the updated PDF. 📄 2. Extract and Embed Content Extract from File:** Parses PDF to extract text. Text Splitter:** Breaks text into chunks. Embeddings (OpenAI):** Converts chunks into vector embeddings. Pinecone Vector Store:** Saves the embeddings with the book name as namespace. 🧠 3. Post Idea Generation (LangChain Agent) Uses a prompt to: Search Pinecone DB Extract insights Format into 5 LinkedIn post ideas with: Hook Insight CTA Memory buffer** and structured output parser are used for clean AI interaction. ✍️ 4. Post Creation Each idea is: Split Rewritten with a GPT model prompt to match LinkedIn tone Styled for under 600 characters Includes emojis, hashtags, and tone guidelines 📊 5. Google Sheet Integration Saves all generated posts to a Google Sheet. Marks status: "published" or "no". 🔁 6. Scheduled Publishing Every day: Pulls an unpublished post Publishes it to LinkedIn Updates the post's status and timestamp in the Google Sheet ⚙️ Setup Guide 📂 Google Drive Create a folder for book PDFs Connect your Google Drive account to n8n Provide access token with file read permission 📊 Google Sheets Create a Google Sheet with columns: bookname, hook, insight, cta, postContent, published, date Add credentials in n8n with read/write permission 🧠 Pinecone Set up a Pinecone project and index (linkdenpost) Namespace will be auto-named using the book filename 🔑 API Credentials Required OpenAI API** (for embeddings and post generation) Pinecone API** (for vector storage and retrieval) LinkedIn OAuth2** (to publish posts) Google Drive & Sheets** credentials 🔁 Flow Summary graph TD A[Google Drive Trigger] --> B[Download PDF] B --> C[Extract Text] C --> D[Text Splitter] D --> E[Create Embeddings] E --> F[Pinecone Vector Store] F --> G[LangChain Agent] G --> H[Structured Output (5 Post Ideas)] H --> I[Split Ideas] I --> J[Format as LinkedIn Post (GPT)] J --> K[Store in Google Sheet] L[Schedule Trigger] --> M[Get Unpublished Post] M --> N[Post to LinkedIn] N --> O[Mark as Published] 🧪 Prompt Example (Used in LangChain Agent) You are a content strategist. Search the Pinecone vector DB containing a book. Generate 5 unique LinkedIn post ideas with: A Hook (curiosity driven) Insight (summary < 100 words) CTA ("Agree or disagree?", etc.) Respond in structured JSON: [ { "Hook": "...", "Insight": "...", "CTA": "..." }, ... ] ✅ Output Sample { "Hook": "Why your lab's results might be invalid 😱", "Insight": "ISO/IEC 17025 stresses that labs must plan and address risks to impartiality and validity.", "CTA": "Does your lab audit for these risks?" } 📆 Schedule Control Uses Schedule Trigger to post daily at a set time. Ensures automation with LinkedIn and accurate Google Sheet syncing. 📝 Notes Posts remain professional and concise for a LinkedIn audience Works with any PDF book Supports multi-book pipelines You can filter and tag books by filename or folder for segmenting post styles
by Alex Gurinovich
AI powered Automated Crypto Insights with Chart-img and BrowserAI Tired of paying for costly crypto updates? Or reading long analyses? This n8n workflow automates the delivery of personalized crypto insights, using Chart-img for capturing coin graphs of BTC, ETH, SOL, and XRP as base64 images, and BrowserAI for web scraping and information gathering of news and articles. This setup ensures thorough market coverage and timely updates, without breaking the bank. Overview Designed for crypto enthusiasts, traders, and analysts, this workflow automates the process of collecting and distributing valuable crypto information. It’s perfect for anyone wanting consistent and accurate updates conveniently. Setup Instructions Pre-conditions Chart-img Account: Register for a Chart-img account and obtain an API key here. BrowserAI Account: Sign up for BrowserAI and get your API key from your BrowserAI dashboard. Step-by-Step Setup 🗓️ Schedule and Date Calculation Triggers twice daily at 8AM and 8PM to ensure up-to-date insights, and can be changed to your like. Calculates yesterday’s date dynamically for accurate data retrieval. 📊 Coin Graph Capture with Chart-img Uses Chart-img API to capture 24-hour graphs for BTC, ETH, SOL, and XRP. Converts images to base64 strings for easy integration into analysis. 🌐 Web Scraping with BrowserAI Creates tasks in BrowserAI to gather the latest crypto news and insights. Automates data extraction for comprehensive market analysis. ⌛ Monitor and Complete Tasks Incorporates status checks to ensure BrowserAI tasks complete successfully before proceeding. ✏️ Analyze and Synthesize Information Combines graph data with web-scraped insights for an enriched summary. Uses AI to generate simple, informative descriptions under 60 words to not overload you. 📩 Deliver Insights Efficiently Sends the compiled analysis to your Telegram, with easy options to switch to WhatsApp, email, or any other communication channel. Customization Guidance Content Personalization:** Customize the datasets and keywords for tailored updates. Modify Schedule:** Adjust triggering times according to your needs using n8n’s scheduling options. This workflow delivers a seamless and cost-effective approach to staying informed about crypto market trends, combining the latest technology for superior insights. ++WARNING:++ This template is intended for personal use only and does not constitute financial advice. Any actions taken using this tool are solely the user's responsibility.
by Adnan
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 👥 Who is this for? This workflow is designed for a variety of professionals who manage vendor relationships and data security. It is especially beneficial for: 🛡️ GRC (Governance, Risk, and Compliance) Professionals**: Streamline your risk assessment processes 🔒 Information Security Teams**: Quickly evaluate the security posture of third-party vendors 📋 Procurement Departments**: Enhance due diligence when onboarding new service providers 🚀 Startup Founders**: Efficiently assess vendors without a dedicated security team This tool is perfect for anyone looking to automate the manual review of vendor websites, policies, and company data. ✨ 🎯 What problem is this workflow solving? Manual vendor due diligence is a time-consuming process that can take hours for a single vendor. This workflow automates over 80% of these manual tasks, which typically include: 🔍 Finding and organizing basic vendor information 🏢 Researching the company's background 📄 Collecting links to key documents like Privacy Policies, Terms of Service, and Trust Pages 📖 Manually reviewing each document to extract risk-relevant information 📊 Compiling all findings into a formatted report or spreadsheet for record-keeping By leveraging Gemini for structured parsing and web scraping with live internet data, this workflow frees you up to focus on critical analysis and final review. ⚡ ⚙️ What this workflow does This end-to-end automated n8n workflow performs the following steps: 📝 Intake: Begins with a simple form to capture the vendor's name, the business use case, and the type of data they will handle 🔎 Background Research: Gathers essential background information on the company ⚠️ Risk Analysis: Conducts comprehensive research on various risk-related topics 🔗 URL Extraction: Finds and validates public URLs for privacy policies, security pages, and trust centers 📈 Risk Assessment: Generates a structured risk score and a detailed assessment based on the collected content and context 📤 Export: Exports the final results to a Google Sheet for easy access and record-keeping 🚀 Setup To get started with this workflow, follow these steps: 🔑 Configure Credentials: Set up your API credentials for Gemini and Jina AI 📊 Connect Google Sheets: Authenticate your Google Sheets account and configure the the Sheet where you want to store the results 🔗 Download the Google Sheet template for your assessment ouput from here ⚙️ (Optional) Customize Prompts: Adjust the prompts within the workflow to better suit your specific needs 🎯 (Optional) Align Risk Framework: Modify the risk questions to align with your organization's internal vendor risk framework
by Gleb D
This n8n workflow automates the enrichment of a company list by discovering and extracting each company’s official LinkedIn URL using Bright Data’s search capabilities and Google Gemini AI for HTML parsing and result interpretation. Who is this template for? This workflow is ideal for sales, business development, and data research professionals who need to collect official LinkedIn company profiles for multiple organizations, starting from a list of company names in Google Sheets. It’s especially useful for teams who want to automate sourcing LinkedIn URLs, enrich their prospect database, or validate company data at scale. How it works Manual Trigger: The workflow is started manually (useful for controlled batch runs and testing). Read Company Names: Company names are loaded from a specified Google Sheets table. Loop Over Each Company: Each company is processed one-by-one: A custom Google Search URL is generated for each name. A Bright Data Web Unlocker request is sent to fetch Google search results for “site:linkedin.com [company name]”. Parse LinkedIn Profile URL Using AI: Google Gemini (or your specified LLM) analyzes the fetched search page and extracts the most likely official LinkedIn company profile. Result Handling: If a profile is found, it’s stored in the results. If not, an empty result is created, but you can add custom logic (notifications, retries, etc.). Batch Data Enrichment: All found company URLs are bundled into a single request for further enrichment from a Bright Data dataset. Export: The workflow appends the final, structured data for each company to another sheet in your Google Sheets file. Setup instructions 1. Replace API Keys: Insert your Bright Data API key in these nodes: Bright Data Web Request - Google Search for Company LinkedIn URL HTTP Request - Post API call to Bright Data Snapshot Progress HTTP Request - Getting data from Bright Data 2. Connect Google Sheets: Set up your Google Sheets credentials and specify the sheet for reading input and writing output. 3. Customize Output Structure: Adjust the Python code node (see sticky note in the template) if you want to include additional or fewer fields in your output. 4. Adjust for Scale or Error Handling: You can modify the logic for “not found” results (e.g., to notify a Slack channel or retry failed companies). 5. Run the Workflow: Start manually, monitor the run, and check your Google Sheet for results. Customization guidance Change Input/Output Sheets: Update the sheet names or columns if your source/target spreadsheet has a different structure. Use Another AI Model: Replace the Google Gemini node with another LLM node if preferred. Integrate Alerts: Add Slack or email nodes to notify your team when a LinkedIn profile is not found or when the process is complete.
by Alex Kim
Weather via Slack 🌦️💬 Overview This workflow provides real-time weather updates via Slack using a custom Slack command: /weather [cityname] Users can type this command in Slack (e.g., /weather New York), and the workflow will fetch and post the latest forecast, including temperature, wind conditions, and a short weather summary. While this workflow is designed for Slack, users can modify it to send weather updates via email, Discord, Microsoft Teams, or any other communication platform. How It Works Webhook Trigger – The workflow is triggered when a user runs /weather [cityname] in Slack. Geocoding with OpenStreetMap – The city name is converted into latitude and longitude coordinates. Weather Data from NOAA – The coordinates are used to retrieve detailed weather data from the National Weather Service (NWS) API. Formatted Weather Report – The workflow extracts relevant weather details, such as: Temperature (°F/°C) Wind speed and direction Short forecast summary Slack Notification – The weather forecast is posted back to the Slack channel in a structured format. Requirements A custom Slack app with: The ability to create a Slash Command (/weather) OAuth permissions to post messages in Slack An n8n instance to host and execute the workflow Customization Replace Slack messaging with email, Discord, Microsoft Teams, Telegram, or another service. Modify the weather data format for different output preferences. Set up scheduled weather updates for specific locations. Use Cases Instantly check the weather for any location directly in Slack. Automate weather reports for team members or projects. Useful for remote teams, outdoor event planning, or general weather tracking. Setup Instructions Create a custom Slack app: Go to api.slack.com/apps and create a new app. Add a Slash Command (/weather) with the webhook URL from n8n. Enable OAuth scopes for sending messages. Deploy the webhook – Ensure it can receive and process Slack commands. Run the workflow – Type /weather [cityname] in Slack and receive instant weather updates.
by Huzaifa Tahir
🎬 What it does This workflow creates an engaging YouTube Short with a single click — from script to voiceover, to visuals and background music. It combines several AI tools to automate content creation and final video assembly. ⚙️ How it works Accepts an input prompt or topic Generates script using GPT Converts script to voiceover using ElevenLabs Generates b-roll style images via Leonardo.Ai Matches background music Assembles a vertical 1080×1920 MP4 video using JSON render config Optionally uploads to YouTube or saves to Cloudinary 🧰 Setup steps Add your credentials: Leonardo API (image generation) ElevenLabs (voiceover) Cloudinary (upload destination) Any GPT-based text generator Drop your audio/music file in the right node Replace API expressions with your own credentials > 🟨 Full step-by-step instructions are in sticky notes inside the workflow.
by DUBCOM
Workflow: Snapshot Contabo How it Works This workflow automates daily backups (snapshots) of VPS instances hosted on Contabo. Each day at midnight, it checks for existing snapshots and ensures that only the latest backups are retained by removing older ones. It provides a seamless, hands-off backup process to keep your data secure. Setup Steps Setting up this workflow is quick, typically taking about 10-15 minutes. The essential part of the setup is providing the necessary credentials, which you can easily retrieve from your Contabo control panel. Import the Workflow: Download and upload the workflow JSON into n8n. Configure Credentials: Add CLIENT_ID, CLIENT_SECRET, API_USER, and API_PASSWORD in the credential node. Activate the Workflow: Enable it to run automatically at midnight every day. Flow Overview Schedule Trigger (00:00 daily):** Automatically initiates the workflow. Formatted Date:** Prepares a timestamp for naming the snapshot. List Snapshots:** Verifies if an existing snapshot is available for each VPS. Conditional Logic:** No Snapshot? Proceeds to create a new one. Snapshot Found? Deletes the old snapshot before creating a new one. Key Points Snapshot Retention:** Old snapshots are deleted to ensure only the latest backups are stored. Unique Identifiers:** UUIDs are used to track and guarantee unique operations.
by Niranjan G
How it works This workflow acts like your own personal AI assistant, automatically fetching and summarizing the most relevant Security, Privacy, and Compliance news from curated RSS feeds. It processes only the latest articles (past 24 hours), organizes them by category, summarizes key insights using AI, and delivers a clean HTML digest straight to your inbox—saving you time every day. Key Highlights Handles three independent tracks: Security, Privacy, and Compliance Processes content from customizable RSS sources (add/remove easily) Filters fresh articles, removes duplicates, and sorts by recency Uses AI to summarize and format insights in a digestible format Sends polished HTML digests via Gmail—one per category Fully modular and extensible—adapt it to your needs Personalization You can easily tailor the workflow: 🎯 Customize feeds: Add or remove sources in the following Code nodes: Fetch Security RSS, Fetch Privacy Feeds, and Fetch Compliance Feeds 🔧 Modify logic: Adjust filters, sorting, formatting, or even AI prompts as needed 🧠 Bring your own LLM: Works with Gemini, but easily swappable for other LLM APIs Setup Instructions Requires Gmail and LLM (e.g., Gemini) credentials Prebuilt with placeholders for RSS feeds and email output Designed to be readable, maintainable, and fully adaptable
by KlickTipp
Community Node Disclaimer: This workflow uses KlickTipp community nodes. How It Works AI Agent and KlickTipp Tools Integration via Telegram: This component connects a large language model (LLM), such as Claude or OpenAI, to the KlickTipp contact management platform through Telegram messaging. The AI Agent interprets natural language queries received from Telegram and dynamically maps them to KlickTipp API operations, enabling intuitive and automated contact handling through a familiar messaging interface. Key Features Telegram & LLM Interaction Setup: Captures messages received via Telegram bot as an alternative to the chat message node. Maintains conversation state using a memory buffer tied to Telegram chat IDs. Interprets user input using an LLM (Claude or OpenAI). Routes interpreted commands to specific KlickTipp tools based on detected intent. Sends responses back to Telegram users with operation results. KlickTipp Integration: Complete set of KlickTipp API endpoints included: Contact Management:** Add, update, get, list, delete, and unsubscribe contacts. Contact Tagging:** Tag, untag, list tagged contacts. Tag Operations:** Create, get, update, delete, list tags. Opt-In Processes:** List and retrieve opt-in process details. Data Fields:** List and get custom data fields. Redirects:** Retrieve redirect URLs. Use Cases Supported: Query contact information via email or name through Telegram messages. Identify and segment contacts by city, region, or behavior via Telegram commands. Create or update contacts from data provided in Telegram messages. Dynamically apply or remove tags to initiate campaigns through Telegram bot interactions. Automate targeted outreach based on contact attributes using Telegram as the control interface. Setup Instructions Install and Configure Nodes: Set up a Telegram bot using BotFather and obtain the bot token. Configure the Telegram Trigger node in n8n with your bot token. Configure the LLM model (e.g., OpenAI or Claude) and memory node if used. Connect all required KlickTipp nodes and authenticate using valid API credentials. Activate the workflow. Define Tagging and Field Mapping: Identify which fields and tags are relevant to your use cases. Ensure necessary tags and custom fields are already created in KlickTipp. Workflow Logic: Trigger via Telegram: A message is received by the Telegram bot and passed to the AI Agent. Query Handling via LLM Agent: AI interprets the natural language input and determines the action. Contact Search & Segmentation: Searches contacts using identifiers (email, address) or criteria. Data Operations: Retrieves, updates, or manages contact and tag data based on interpreted command. Campaign Preparation: Applies tags or sends campaign triggers depending on query results. Response via Telegram: Sends formatted results back to the Telegram user. Benefits: Mobile-First Interface:** Users can manage KlickTipp contacts directly from Telegram on any device. AI-Powered Automation:** Reduces manual contact search and tagging efforts through intelligent processing. Scalable Integration:** Built-in support for full range of KlickTipp operations allows diverse use-case handling. Data Consistency:** Ensures structured data flows between Telegram, AI, and KlickTipp, minimizing errors. Testing and Deployment: Use defined Telegram messages such as: “Tell me something about the contact with email address X” “Tag all contacts from region Y” “Send campaign Z to customers in area A” Validate expected actions in KlickTipp after message execution and confirm responses in Telegram. Notes: Customization:** Adjust tag logic, AI prompts, and contact field mappings based on project needs. Extensibility:** The template can be expanded with further logic for Google Sheets input or campaign feedback loops Resources: Use KlickTipp Community Node in n8n Automate Workflows: KlickTipp Integration in n8n
by Don Jayamaha Jr
Instantly fetch real-time Bitget spot market data directly in Telegram! This workflow integrates the Bitget REST v2 API with Telegram (plus optional AI-powered formatting) to deliver the latest crypto price, order book, candles, and recent trades. Perfect for crypto traders, analysts, and investors who need reliable market data at their fingertips—no API key required.  Sign-up for Bitget for 6,200 USDT in rewards to trade: Collect Now How It Works A Telegram bot listens for user requests (e.g., BTCUSDT). The workflow connects to Bitget public endpoints to fetch: Ticker (latest price & 24h stats) Order book depth (top bids/asks) Recent trades (price, side, volume, timestamp) Candlestick data (1m, 15m, 1h, 4h, 1d) Historical candles (optional, for backfill before endTime) A Calculator node derives useful metrics like mid-price and spread. A Think node reshapes raw JSON into Telegram-ready text. A splitter ensures reports over 4000 characters are chunked safely. The final market insights are delivered instantly back to Telegram. What You Can Do with This Agent ✅ Track live prices & 24h stats for any Bitget spot pair. ✅ Monitor order book liquidity and spreads in real-time. ✅ Analyze candlesticks across multiple timeframes. ✅ Review recent trades to see execution flow. ✅ Fetch historical candles for extended market context. ✅ Receive clean, structured reports with optional AI-enhanced formatting. Set Up Steps Create a Telegram Bot Use @BotFather to generate a bot token. Configure in n8n Import Bitget AI Agent v1.02.json into your n8n instance. Add your Telegram credentials (bot token + your Telegram ID in the User Authentication node). Add an OpenAI key if you want AI-powered formatting. (Optional) Add an *Bitget api key** . Deploy and Test Send BTCUSDT to your bot. Get live Bitget spot data instantly in Telegram! 🚀 Unlock powerful, real-time Bitget insights in Telegram—zero setup, zero API keys required! 📺 Setup Video Tutorial Watch the full setup guide on YouTube: 🧾 Licensing & Attribution © 2025 Treasurium Capital Limited Company Architecture, prompts, and trade report structure are IP-protected. No unauthorized rebranding permitted. 🔗 For support: Don Jayamaha – LinkedIn
by Robert Breen
Chat to write or reword a blog post. The workflow stores each result in Google Sheets and uses a sub-workflow “Google tool” to count rows per session (your running context). If a session exceeds a row threshold, the flow can branch (e.g., stop or notify). ⚙️ Setup Instructions 1️⃣ Set Up OpenAI Connection Go to OpenAI Platform Navigate to OpenAI Billing Add funds to your billing account Copy your API key into the OpenAI credentials in n8n 2️⃣ Prepare Your Google Sheet Connect your Data in Google Sheets Use this format: Sample Sheet Row 1 = column names (e.g., session, Rows, output) Data in rows 2–100 (or more if you prefer) In n8n, use Google Sheets OAuth2 → pick your Spreadsheet and Worksheet (Optional) You can adapt this to Airtable, Notion, or a Database 🧠 How It Works Chat Trigger**: Provide a topic (write) or paste existing text (reword). Code Node (“Choose to Write or Edit Blog”)**: Builds a system_prompt + user_prompt Instructs the agent to call the Google tool (sub-workflow) with only the sessionid to count existing rows. Tool Workflow (“google”)**: Fetches rows from the sheet → filters by session → summarizes row count. Agent (“Blog Writer & Editor”)**: Returns structured JSON (items/rows, session, blog body). Store (Google Sheets)**: Appends { session, Rows, output } to the sheet. If Node**: Example rule: Rows > 3 → branch/limit/notify as needed. 💬 Example Prompts “Write a 600-word blog about n8n agents with 3 bullet takeaways. Session: abc123.” “Reword this post into a concise LinkedIn article. Session: launchQ3:\n<your text here>” “Draft a blog intro and 5 SEO headlines on marketing automation. Session: mkt-01.” 📬 Contact Need help tailoring this to Airtable/Notion/DB, or adding auto-publishing? 📧 rbreen@ynteractive.com 🔗 Robert Breen 🌐 ynteractive.com