by WeWeb
This n8n template helps you build a full AI-powered LinkedIn content generator with just a few clicks. Paired with the free WeWeb UI template, it becomes a ready-to-use web app where users can: Add their own OpenAI API key Customize the prompt and define 6 content topics Edit the AI-generated topics Choose when to generate LinkedIn posts, complete with hashtags and an optional image Who This Is For Perfect for marketers, indie hackers, and solopreneurs who want to build their personal brand on LinkedIn while staying in control of what gets posted. 🧠 What Makes This Different Unlike most AI agents, you stay fully in control: You define the tone and focus via the prompt. You choose which topics to keep or modify. You decide when to generate a post. You can build on top of this and create your own SaaS product. It’s also modular and extendable—hook it up to your backend, add user login, or feed AI improvements based on user input. ⚙️ How It Works Triggering Events: The app includes 3 pre-configured triggers, ready to be hooked into your WeWeb frontend. Just update the webhook URLs after duplicating the n8n workflow. Topic Generation: A call is made to OpenAI (GPT-4) to generate topic ideas based on your prompt. Post Creation: Once topics are approved or edited, GPT-4 writes full posts with suggested hashtags. Image Generation (Optional): If enabled, a DALL·E call generates a relevant image. Everything Stays Local: All data and images are handled locally, no cloud storage setup needed. 🧪 Requirements & Setup No fancy infrastructure required. Here’s what helps you get started: Free WeWeb account** (recommended) to use the frontend UI template OpenAI account** with API access (for GPT-4 and DALL·E) n8n account** (self-hosted or cloud) to run the backend workflow The template is completely free to use. Since each user adds their own OpenAI API key, you don't need to worry about usage costs or rate limits on your end. 🔧 Want to Go Further? This setup is beginner-friendly, but developers can: Add user accounts Save post history Feed user feedback back into the prompt logic Launch their own branded version as a SaaS
by Muhammad Bello
Email Inbox Manager System Categories Email Automation AI-Powered Operations Internal Productivity Tools This workflow builds a fully automated AI-powered email categorization and response assistant. It intelligently processes, categorizes, labels, and drafts replies to incoming Gmail messages in real time using AI with zero manual involvement. Perfect for support, sales, finance, and internal operations. Benefits Automated Email Triage** – Every unread email is instantly read, analyzed, and classified AI-Powered Categorization** – Uses GPT-4 to understand content and apply correct labels Smart Response Generation** – Automatically drafts accurate replies based on category Slack Notifications** – Instantly notifies your team about internal or sales-related messages Seamless Gmail Integration** – Labels, drafts, and marks emails directly in your inbox Custom Classification Rules** – Tailored to internal, support, sales, finance, and promotional needs How It Works Gmail Trigger: Monitors Gmail inbox in real-time for new unread messages Triggers the workflow every minute with no need for manual refresh Smart Classification: Feeds email body to AI-powered Text Classifier Categories include: Internal, Customer Support, Promotions, Admin/Finance, and Sales Opportunity Classification based on sender domain, keywords, and context AI-Based Labeling: Applies appropriate Gmail label based on classification result Helps keep inbox clean, organized, and easily searchable AI Reply Generation: Specialized GPT-4 agents generate replies tailored to each category: Internal:** Polished team replies Customer Support:** Clear and professional customer responses Promotions:** Summarizes and evaluates promotional value Admin/Finance:** Extracts invoice/payment information Sales Opportunity:** Drafts personalized replies and sales notifications Auto-Drafting + Slack Alerts: Replies are saved as Gmail drafts, ready for review or direct send Sends Slack notifications for Internal or Sales Opportunities Includes subject lines and quick message previews Smart Decision-Making: Promotional emails are evaluated with AI for usefulness Only valuable offers are flagged or responded to Automatically marks emails as read after processing Business Use Cases Customer Support Teams** – Automatically categorize and prep replies to client messages Sales Reps** – Instantly receive drafted responses to new inquiries Operations Managers** – Keep internal comms clear and responsive Finance Departments** – Auto-extract and review payment/invoice messages Founders** – Never miss an important email while your AI sorts and replies for you Difficulty Level: Intermediate Estimated Build Time: 2–4 hours Monthly Operating Cost: $20–80 (depending on OpenAI usage and Slack volume) Required Setup Gmail Integration Set up Gmail OAuth2 connection Create labels: Internal, Customer Support, Promotions, Admin/Finance, Sales Opportunity OpenAI Integration Connect GPT-4 or GPT-4o account Configure role-based prompts for each email category Output structured response data (subject, body, notification) Slack Integration Set up Slack OAuth2 connection Configure target Slack channel Notify team when important categories are triggered System Architecture The workflow follows a powerful 6-stage automation: Trigger – Poll Gmail for new unread emails Classify – AI categorizes the email into the right bucket Label – Applies Gmail label for search and visibility Generate Reply – GPT-4 crafts draft email reply Draft Email – Saves response in Gmail Notify – Optional Slack message alerts for priority emails Why This System Works Inbox Clarity – Keeps your inbox categorized and organized Human Quality Replies – AI-generated messages sound professional and personalized Time-Saving Automation – Handles support, sales, internal ops, and finance without touching your inbox Multi-Agent Architecture – Each email type is handled by a specialized GPT-4 prompt Real Time Reactions – From email receipt to Slack notification in under a minute
by Adam Bertram
LintGuardian: Automated PR Linting with n8n & AI What It Does LintGuardian is an n8n workflow template that automates code quality enforcement for GitHub repositories. When a pull request is created, the workflow automatically analyzes the changed files, identifies linting issues, fixes them, and submits a new PR with corrections. This eliminates manual code style reviews, reduces back-and-forth comments, and lets your team focus on functionality rather than formatting. How It Works The workflow is triggered by a GitHub webhook when a PR is created. It fetches all changed files from the PR using the GitHub API, processes them through an AI-powered linting service (Google Gemini), and automatically generates fixes. The AI agent then creates a new branch with the corrected files and submits a "linting fixes" PR against the original branch. Developers can review and merge these fixes with a single click, keeping code consistently formatted with minimal effort. Prerequisites To use this template, you'll need: n8n instance: Either self-hosted or using n8n.cloud GitHub repository: Where you want to enforce linting standards GitHub Personal Access Token: With permissions for repo access (repo, workflow, admin:repo_hook) Google AI API Key: For the Gemini language model that powers the linting analysis GitHub webhook: Configured to send PR creation events to your n8n instance Setup Instructions Import the template into your n8n instance Configure credentials: Add your GitHub Personal Access Token under Credentials → GitHub API Add your Google AI API key under Credentials → Google Gemini API Update repository information: Locate the "Set Common Fields" code node at the beginning of the workflow Change the gitHubRepoName and gitHubOrgName values to match your repository const commonFields = { 'gitHubRepoName': 'your-repo-name', 'gitHubOrgName': 'your-org-name' } Configure the webhook: Create a file named .github/workflows/lint-guardian.yml in your repository replacing the Trigger n8n Workflow step with your webhook: name: Lint Guardian on: pull_request: types: [opened, synchronize] jobs: trigger-linting: runs-on: ubuntu-latest steps: name: Trigger n8n Workflow uses: fjogeleit/http-request-action@v1 with: url: 'https://your-n8n-instance.com/webhook/1da5a6e1-9453-4a65-bbac-a1fed633f6ad' method: 'POST' contentType: 'application/json' data: | { "pull_request_number": ${{ github.event.pull_request.number }}, "repository": "${{ github.repository }}", "branch": "${{ github.event.pull_request.head.ref }}", "base_branch": "${{ github.event.pull_request.base.ref }}" } preventFailureOnNoResponse: true Customize linting rules (optional): Modify the AI Agent's system message to specify your team's linting preferences Adjust file handling if you have specific file types to focus on or ignore Security Considerations When creating your GitHub Personal Access Token, remember to: Choose the minimal permissions needed (repo, workflow, admin:repo_hook) Set an appropriate expiration date Treat your token like a password and store it securely Consider using GitHub's fine-grained personal access tokens for more limited scope As GitHub documentation notes: "Personal access tokens are like passwords, and they share the same inherent security risks." Extending the Template You can enhance this workflow by: Adding Slack notifications when linting fixes are submitted Creating custom linting rules specific to your team's needs Expanding it to handle different types of code quality checks Adding approval steps for more controlled environments This template provides an excellent starting point that you can customize to fit your team's exact workflow and code style requirements.
by David Roberts
This workflow allows you to ask questions about the data in a Google Sheet over a chat interface. It uses n8n's built-in chat, but could be modified to work with Slack, Teams, WhatsApp, etc. Behind the scenes, the workflow uses GPT4, so you'll need to have an OpenAI API key that supports it. How it works The workflow uses an AI agent with custom tools that call a sub-workflow. That sub-workflow reads the Google Sheet and returns information from it. Because models have a context window (and therefore a maximum number of characters they can accept), we can't pass the whole Google Sheet to GPT - at least not for big sheets. So we provide three ways of querying less data, that can be used in combination to answer questions. Those three functions are: List all the columns in the sheet Get all values of a single column Get all values of a single row Note that to use this template, you need to be on n8n version 1.19.4 or later.
by Naveen Choudhary
This workflow automatically enriches company domain lists with comprehensive business information scraped from ZoomInfo, organizing the data in Google Sheets for sales teams and researchers. Who's it for Sales teams** building prospect databases with accurate company information Marketing professionals** researching target companies for outreach campaigns Business development teams** qualifying leads with revenue and employee data Researchers** collecting structured company data for market analysis Lead generation specialists** enriching domain lists with contact details How it works The workflow processes unprocessed domains from a Google Sheet, searches for their ZoomInfo profiles using Serper API, scrapes the company pages through Oxylabs proxy service, and extracts structured business data. Each domain is marked as processed to prevent duplicates, and the workflow includes proper rate limiting to respect API limits. What it does Loads unprocessed domains from your Google Sheets database Searches ZoomInfo using targeted queries via Serper API for each domain Validates search results and extracts relevant ZoomInfo profile URLs Scrapes company pages using Oxylabs to bypass anti-scraping protection Extracts structured data including company details, address, revenue, and employee count Updates Google Sheets with enriched company information Tracks processing status to prevent reprocessing the same domains Requirements Serper API account** with search credits (Get API key) Oxylabs subscription** for web scraping proxy service (Sign up here) Google Sheets API access** with OAuth2 authentication Google Sheets template** - Make a copy of this template sheet with pre-configured columns How to set up Make a copy of the Google Sheets template - Click here to copy the template to your Google Drive Configure API credentials in the respective HTTP Request nodes: Add Serper API key in the search node Set up Oxylabs username/password in the scraping node Set up Google Sheets authentication using OAuth2 Update the Google Sheets document ID in all Google Sheets nodes to point to your copied template Add your domain list to the sheet with 'processed' column empty or false Run the workflow using the manual trigger How to customize the workflow Search query modification**: Update the search query in the Serper node for different geographic focus (currently set for Czech Republic) Data extraction fields**: Modify the Google Sheets column mapping to include/exclude specific company data points Rate limiting**: Adjust wait times between requests to match your API rate limits Batch processing**: Configure the split batch size for processing domains in smaller groups Error handling**: Customize the continue-on-error settings based on your data quality requirements Scheduling**: Replace Manual Trigger with Schedule Trigger for automated daily/weekly runs Output data includes Complete company name and official address Phone numbers and contact information Revenue figures and employee headcount Industry classifications and business categories LinkedIn company profile URLs Geographic location details (city, state, country, postal code) Processing status tracking for workflow management Note: This workflow includes comprehensive error handling to ensure domains are always marked as processed, preventing infinite loops while maintaining data integrity. Rate limiting is built-in to respect API quotas and avoid service interruptions.
by Blockia Labs
Time Logging on Clockify Using Slack How it works This workflow simplifies time tracking for teams and agencies by integrating Slack with Clockify. It enables users to log, update, or delete time entries directly within Slack, leveraging an AI-powered assistant for seamless and conversational interactions. Key features include: Effortless Time Logging**: Create and manage time entries in Clockify without leaving Slack. AI-Powered Assistant**: Get step-by-step guidance to ensure accurate and efficient time logging. Project and Client Management**: Retrieve project and client information from Clockify effortlessly. Overlap Prevention**: Avoid overlapping entries with built-in time validation. Automated Descriptions**: Generate ethical, grammatically correct descriptions for time logs. Set up steps 1. Prepare your integrations Ensure you have active accounts for both Slack and Clockify. Generate your Clockify API credentials for integration. 2. Import the workflow Download and import the workflow template into your n8n instance. Configure the workflow to connect with your Slack and Clockify accounts. 3. Configure the workflow Add your Clockify API credentials in the workflow settings. Set up the Slack Trigger to listen for app mentions or specific commands. 4. Test the workflow Use Slack to create a time entry and verify it in Clockify. Test updating and deleting existing entries to ensure smooth functionality. Check for any overlapping time logs or incorrect data entries. Why use this workflow? Efficiency**: Eliminate the need to switch between tools for time tracking. Accuracy**: AI-driven validation ensures error-free entries. Automation**: Simplify repetitive tasks like updating or deleting time logs. Proactive Guidance**: Conversational assistant ensures smooth operations.
by Elegant Biztech
Automated QuickBooks Invoice to Custom PDF & Email Tired of the standard, boring invoices from QuickBooks Online? This workflow completely automates the process of creating beautiful, custom-branded PDF invoices and emailing them directly to your clients, saving you time and elevating your brand's professionalism. The moment you create an invoice in QuickBooks, this workflow triggers, fetches all the necessary data, and generates a lavish, multi-page-aware PDF invoice complete with your company logo and signature. Key Features Fully Automated:** Runs instantly when a new invoice is created in QuickBooks. Custom Branding:** Automatically fetches your company logo and signature from a URL to place on the invoice. Modern & Professional Design:** Uses a premium, multi-column HTML template that is clean, easy to read, and far superior to the default QBO templates. Multi-Page Ready:** If an invoice has many line items, the template will intelligently create multiple pages and add a "Page X of Y" footer automatically. Smart Layout:** The totals and summary block are designed to never break across pages, ensuring a professional look no matter the length. Automatic Emailing:** The final PDF is attached to a beautifully formatted email and sent directly to the customer's email address on file. Prerequisites Before you start, you will need a few things: A running n8n instance. A QuickBooks Online account with API access. A running Gotenberg instance. This is a powerful, open-source tool for converting HTML to PDF. This workflow is designed to connect to its API. You can learn more about it here. Publicly accessible URLs for your company logo and signature image (e.g., hosted on your website or a service like Imgur). Setup Guide Follow these steps carefully to configure the workflow for your own use. Nodes that need your attention are marked with a [!!] prefix. Step 1: Configure the QuickBooks Webhook The workflow starts with a webhook. You need to tell QuickBooks to send information to this webhook. Open the [!!] Listen for New QuickBooks Invoice node. You will see a Webhook URL. Copy the Production URL. Go to your QuickBooks Developer dashboard, select your app, and navigate to the Webhooks section. Paste the n8n URL into the Endpoint URL field and select the Invoice event to subscribe to. Step 2: Connect Your QuickBooks Account Open the [!!] Get Invoice Data from QuickBooks node. In the "Credentials" field, select your existing QuickBooks Online credentials or create a new set. Step 3: Add Your Branding Open the [!!] Fetch Company Logo Image node. In the URL field, replace the placeholder with the public URL of your company's logo. Open the [!!] Fetch Company Signature Image node. In the URL field, replace the placeholder with the public URL of your signature image. Step 4: Update the PDF Generation Service Open the [!!] Generate PDF via Gotenberg node. In the URL field, replace the placeholder http://YourGotenBergInstanceURL/... with the real URL of your running Gotenberg instance. Step 5: Configure Your Email Open the [!!] Email PDF Invoice to Customer node. In the "Credentials" field, select your SMTP or email service credentials. Customize the From Email and Subject fields. You can also edit the beautiful HTML email body to match your company's tone of voice. Step 6: Activate Your Workflow You're all set! Save the workflow and activate it using the toggle at the top-right of the screen. Now, when you create a new invoice in QuickBooks, this automation will handle the rest. A Note from the Creator Thank you for using this workflow! I believe that professional and automated invoicing is a cornerstone of a great business. This tool was designed to save you time and help you put your best foot forward with every client interaction. If you have any questions or need assistance, feel free to reach out. Website:** https://www.elegantbiztech.com/ Email:** sales@elegantbiztech.com
by Cyril Nicko Gaspar
📌 AI Agent Template with Bright Data MCP Tool Integration This template enables natural-language-driven automation using Bright Data MCP tools. It extracts all available tools from MCP, processes the user’s query through an AI agent, then dynamically selects and executes the appropriate tool. ❓ Problem It Solves Traditional automation often requires users to understand APIs, interfaces, or scripts to perform backend tasks. The Bright Data MCP integration solves this by allowing natural language interaction, intelligently classifying user intent, and managing context-aware execution of complex operations—ideal for data extraction, customer support, and workflow orchestration. 🧰 Pre-requisites Before deploying this template, make sure you have: An active N8N instance (self-hosted or cloud). A valid OpenRouter API key (or another compatible AI model). Telegram bot and its API token Access to the Bright Data MCP API with credentials. Basic familiarity with N8N workflows and nodes. ⚙️ Setup Instructions Setup and obtain API token and other necessary information from Bright Data In your Bright Data account, obtain the following information: API token Web Unlocker zone name (optional) Browser Zone name (optional) Host SSE server from STDIO command The methods below will allow you to receive SSE (Server-Sent Events) from Bright Data MCP via a local Supergateway or Smithery ** Method 1: Run Supergateway in a separate web service (Recommended) This method will work for both cloud version and self-hosted N8N. Signup to any cloud services of your choice (DigitalOcean, Heroku, Hetzner, Render, etc.). For NPM based installation: Create a new web service. Choose Node.js as runtime environment and setup a custom server without repository. In your server’s settings to define environment variables or .env file, add: `API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_ZONE=optional_browser_zone_name` Paste the following text as a start command: npx -y supergateway --stdio "npx -y @brightdata/mcp" --port 8000 --baseUrl http://localhost:8000 --ssePath /sse --messagePath /message Deploy it and copy the web server URL, then append /sse into it. Your SSE server should now be accessible at: https://your_server_url/sse For Docker based installation: Create a new web service. Choose Docker as the runtime environment. Set up your Docker environment by pulling the necessary images or creating a custom Dockerfile. In your server’s settings to define environment variables or .env file, add: `API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_AUTH=optional_browser_auth` Use the following Docker command to run Supergateway: `docker run -it --rm -p 8000:8000 supercorp/supergateway \ --stdio "npx -y @brightdata/mcp /" \ --port 8000` Deploy it and copy the web server URL, then append /sse into it. Your SSE server should now be accessible at: https://your_server_url/sse For more installation guides, please refer to https://github.com/supercorp-ai/supergateway.git. ** Method 2: Run Supergateway in the same web service as the N8N instance This method will only work for self-hosted N8N. a. Set Required Environment Variables In your server's settings to define environment variables or .env file, add: API_TOKEN=your_brightdata_api_token WEB_UNLOCKER_ZONE=optional_zone_name BROWSER_ZONE=optional_browser_zone_name b. Run Supergateway in Background npx -y supergateway --stdio "npx -y @brightdata/mcp" --port 8000 --baseUrl http://localhost:8000 --ssePath /sse --messagePath /message Use the command above to execute it through the cloud shell or set it as a pre-deploy command. Your SSE server should now be accessible at: http://localhost:8000/sse For more installation guides, please refer to https://github.com/supercorp-ai/supergateway.git. * *Method 3: Configure via Smithery.ai* (Easiest) If you don't want additional setup and want to test it right away, follow these instructions: Visit https://smithery.ai/server/@luminati-io/brightdata-mcp/tools to: Signup (if you are new to Smithery) Create an API key Define environment variables via a profile Retrieve your SSE server HTTP URL Import the Workflow Open N8N. Import the JSON workflow file included with this template. Update any nodes referencing external services (e.g., OpenRouter, Telegram). Setup Telegram Integration If you haven't setup a bot in Telegram, below is the instruction how to create one using BotFather: Search for @BotFather in Telegram and start a conversation with it. Send the command /newbot to create a new bot. You'll be prompted to enter a name and a unique username for your bot. BotFather will provide you with an access token, which you'll need to use to interact with the bot's API. Edit the HTTP Request node in the workflow. Configure the URL as follows: https://api.telegram.org/bot+your_telegram_bot_token+/setWebhook?url=+your_webhook_url Replace +your_telegram_bot_token+ with your actual Telegram bot token. Replace +your_webhook_url+ with the URL from the Webhook Trigger node in the workflow. This will set up Telegram to forward messages to your n8n agent. 🔄 Workflow Functionality (Summary) The user submits a message via chat. Memory** nodes retain context for multi-turn conversations. The mapped tool is executed and results are returned contextually. 🧠 Optional memory buffers and memory manager nodes keep the interaction context-aware. 🧩 Use Cases Data Scraping on Demand**: Launch scraping tasks via chat. Lead Generation Bots**: Enrich or validate leads with MCP tools. AI-Powered Customer Support**: Classify and answer queries with real-time data tools. Workflow Assistants**: Let teams run backend processes like lookups or report generation using plain language. 🛠️ Customization Classifier Prompt & Logic**: Tweak the AI’s prompt and tool-matching schema to better fit your use case. Memory Configuration**: Adjust retention policies and context depth. Tool Execution Sub-Workflow**: Extend for retries, logging, or chaining actions. Omni-Channel Support**: Connect via webhooks to chat interfaces like Slack, WhatsApp, Telegram, or custom UIs. ✅ Summary This template equips you with a powerful no-code/low-code AI agent that translates conversation into real-world action. Using Bright Data’s MCP tools through natural language, it enables teams to automate and scale data-driven tasks effortlessly.
by Nasser
For Who? Content Creators Youtube Automation Marketing Team How it works? 1 - Every week, retrieve the keywords you want to track 2 - Thanks to Apify, scrape videos from YouTube Search related to these keywords, filtered by relevance 3 - Wait until the dataset is completed 4 - Get the information contained in the dataset 5 - For each video, clean and summarize the script 6 - Upload everything to your Airtable database 📺 YouTube Video Tutorial: Setup (~5min) Scheduled Trigger: Select the frequency you want. If you change it, update the data accordingly in the "Create Videos Dataset" HTTP Request node in Body ➡️ JSON ➡️ dateFilter. Setup Keywords: Enter keywords related to the niche you want. If you change the number of keywords, update the data accordingly in the "Create Videos Dataset" HTTP Request node in Body ➡️ JSON ➡️ searchQueries. Create Videos Dataset: Refer to the Apify documentation for more: https://docs.apify.com/api/v2/getting-started APIs: For all HTTP Request nodes in the URL field, replace [YOUR_API_TOKEN] with your API token. 👨💻 More Workflows : https://n8n.io/creators/nasser/
by Abhishek Patoliya
This n8n automation lets you build a complete AI-powered task management system that integrates Telegram, Google Sheets, and GPT-4o mini to help users easily manage to-do lists and receive daily task reminders. Users can interact with the system via Telegram, while the AI assistant (powered by GPT-4o mini) processes commands naturally, updates a central Google Sheet, and ensures scheduled reminders are sent for pending tasks. ✨ Key Features ✅ Add, list, update, complete, or delete tasks via Telegram ✅ AI-powered conversational responses using GPT-4o ✅ All tasks stored and synced in Google Sheets ✅ Daily scheduled task summary and pending reminders sent to Telegram ✅ Friendly, human-like assistant responses ✅ Fully configurable and easy to set up 🛠️ Workflow Functionality Breakdown 1. User Interacts on Telegram Sends commands like: add buy groceries list tasks complete submit report delete dentist appointment 2. AI-Powered Processing A GPT-4o agent processes user messages Ensures clear, friendly responses Determines task intent: add, update, delete, list, complete 3. Google Sheets Sync Every operation is logged to Google Sheets Google Sheets acts as the master task database Sheet structure includes: Task Status (pending or done) Created At (timestamp) Due Date (optional) Notes (optional) 4. Scheduled Daily Task Summary At 9 PM daily, the workflow: Fetches tasks from Google Sheets Generates a warm, conversational summary via GPT-4o Sends the summary to the user on Telegram 5. Automated Reminders Checks for pending tasks due today Sends reminder messages to Telegram ✅ Prerequisites Before setting up the workflow, ensure you have: ✔️ An n8n instance (Cloud or self-hosted) ✔️ A Telegram Bot Token ✔️ Access to Google Sheets API (OAuth2 credentials) ✔️ An OpenAI API Key with GPT-4o access ✔️ A Google Sheet structured as per the specification below 📝 Google Sheet Structure Your Google Sheet should have the following columns: | Column Name | Description | | -------------- | ---------------------------------------------------- | | Task | Short task description | | Status | pending or done | | Created At | Date & time task was created (YYYY-MM-DD HH:mm:ss) | | Due Date | (Optional) When task is due (YYYY-MM-DD HH:mm:ss) | | Notes | (Optional) Additional details | Important: The first row should be the header row with these exact column names. 🔧 Setup Instructions 1. Telegram Bot Setup Create a bot via BotFather Obtain the Bot Token Connect Telegram Trigger and Telegram Send nodes using your Bot Token in n8n 2. Google Sheets API Setup Follow n8n Google Sheets integration guide Set up OAuth2 credentials Provide access to your task Google Sheet 3. OpenAI API Setup Obtain an API key from OpenAI Ensure GPT-4o mini access is enabled Add OpenAI credentials to relevant nodes 4. Sheet Linking Replace the Google Sheet ID in the workflow with your own Confirm sheet names and column structure match exactly 5. Schedule Configuration (Optional) Adjust the daily summary time (Schedule Trigger node) as desired ⚙️ Configuration Options 🔧 Adjust AI prompt instructions for tone/style 🔧 Change reminder times in the schedule trigger 🔧 Customize Google Sheet columns if needed (update mappings accordingly) 🔧 Add multi-user support with chat ID checks (advanced) 📂 Files Included Full n8n JSON workflow ready to import 💡 Tips You can extend this with Slack, WhatsApp, or Email reminders Combine with Notion, ClickUp, or CRM integrations for more powerful task management Consider adding a "Priority" column for advanced sorting Ready to stay organized with AI-powered task management? Import this workflow, link your accounts, and your Telegram assistant is good to go! 🚀
by Ranjan Dailata
Who this is for The Crunchbase B2B Lead Discovery Pipeline is designed for sales teams, B2B marketers, business analysts, and data operations teams who need a reliable way to extract, structure, and summarize company information from Crunchbase to fuel lead generation and market intelligence. This workflow is ideal for: Sales Development Reps (SDRs) - Needing structured leads from Crunchbase Marketing Analysts - Generating segmented outreach lists Growth Teams - Identifying trending B2B startups RevOps Teams - Automating company research pipelines Data Teams - Consolidating insights into Google Sheets for dashboards What problem is this workflow solving? Manual extraction of company data from Crunchbase is time-consuming, inconsistent, and often lacks the contextual summary required for sales enablement or growth targeting. This workflow automates the extraction, transformation, summarization, and delivery of Crunchbase company data into structured formats, making it instantly usable for B2B targeting and analysis. It solves: The difficulty of scaling lead discovery from Crunchbase The need to summarize raw textual content for quick insights The lack of integration between web scraping, LLM processing, and storage What this workflow does Markdown to Textual Data Extractor**: Takes raw scraped markdown from Crunchbase and converts it into readable plain text using a basic LLM chain Structured Data Extraction**: Applies a parsing model (OpenAI) to extract structured fields such as company name, funding rounds, industry tags, location, and founding year Summarization Chain**: Generates an executive summary from the raw Crunchbase text using a summarization prompt template Send to Google Sheets**: Adds the structured data and summary into a Google Sheet for team access and further processing Persist to Disk**: Saves both raw and structured data files locally for archiving or further use Webhook Notification**: Sends a structured payload to a webhook endpoint (e.g., Slack, CRM, internal tools) with lead insights Pre-conditions You need to have a Bright Data account and do the necessary setup as mentioned in the "Setup" section below. You need to have an OpenAI Account. Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. In n8n, Configure the Google Sheet Credentials with your own account. Follow this documentation - Set Google Sheet Credential In n8n, configure the OpenAi account credentials. Ensure the URL and Bright Data zone name are correctly set in the Set URL, Filename and Bright Data Zone node. Set the desired local path in the Write a file to disk node to save the responses. How to customize this workflow to your needs LLM Prompt Customization : Modify the extraction prompt to include additional fields like revenue, social links, leadership team Adjust summarization tone (e.g., executive summary, sales-focused snapshot or marketing digest) File Persistence Store raw markdown, extracted JSON, and summary text separately for audit/debug Webhook Notification Connect to CRM (e.g., HubSpot, Salesforce) via webhook to automatically create leads Send Slack notifications to alert sales reps when a new high-potential company is discovered
by Trung Tran
🎙️ VoiceScribe AI: Telegram Audio Message Auto Transcription with OpenAI Whisper > Automatically transcribe Telegram voice messages and store them as structured logs in Google Sheets, while backing up the audio in Google Drive. 🧑💼 Who’s it for Journalists, content creators, or busy professionals who often record voice memos or short interviews on the go. Anyone who wants to turn voice recordings into searchable, structured notes. ⚙️ How it works / What it does User sends a voice message to a Telegram bot. n8n checks if the message is an audio voice note. If valid, it downloads the audio file and: Transcribes it using OpenAI Whisper (or your LLM of choice). Uploads the original audio to Google Drive for safekeeping. The transcript and audio metadata are merged. The workflow: Logs the data into a Google Sheet. Sends a formatted confirmation message to the user via Telegram. If the input is not audio, the bot politely informs the user that only voice messages are accepted. ✅ Features Accepts only Telegram voice messages. Transcribes via OpenAI Whisper. Logs DateTime, Duration, Transcript, and Audio URL to Google Sheets. Sends user feedback message via Telegram with download + transcript link. 🚀 How to set up Prerequisites Telegram Bot connected to n8n (via Telegram Trigger) Google Drive & Google Sheets credentials configured OpenAI or Whisper API credentials (for transcription) Steps Telegram Trigger Start the flow when a new message is sent to your bot. Check Message Type Use a conditional node to confirm it's a voice message. Download Voice Message Download the .oga file from Telegram. Transcribe Audio Send the binary audio to OpenAI Whisper or your transcription service. Upload to Google Drive Backup the original audio file. Merge Outputs Combine transcription with Drive metadata. Transform to Row Format Prepare structured JSON for Google Sheets. Append to Google Sheet Store the transcript log (DateTime, Duration, Transcript, AudioURL). Send Confirmation to User Inform the user via Telegram with their transcript and download link. Unsupported Message Handler Reply to users who send non-audio messages. 📄 Example Output in Google Sheet | DateTime | Duration | Transcript | AudioURL | |-----------------------|----------|--------------------------------------------|------------------------------------------------------------| | 2025-08-07T13:12:19Z | 27 | Dự án Outlet Activation là... | https://drive.google.com/uc?id=xxxx&export=download | 🧠 How to customize the workflow Swap Whisper with Deepgram, AssemblyAI, or other providers. Add speaker name detection or prompt-based tagging via GPT. Route transcripts into Notion, Airtable, or CRM systems. Add multi-language support or summarization steps. 📦 Requirements | Component | Required | |---------------------|----------| | Telegram API | ✅ | | Google Drive API | ✅ | | Google Sheets API | ✅ | | OpenAI Whisper API | ✅ | | n8n Cloud or Self-hosted | ✅ | Created with ❤️ using n8n