by Praveena
What is Elderwatch Elder Watch is a simple system that checks daily vitals — like heart rate, oxygen, and walking symmetry — using data from an iPhone or Apple Watch. If something looks off — say oxygen drops or heart rate spikes — it flags that as “attention required.” And depending on that status, it can either: Email a daily report to a caregiver Or if there’s an alert — trigger a phone call via Twilio Why do we need this Elder Watch can help older people living alone for children or care givers to keey an eye on without obsessively checking apps. It’s useful for clinics that run home-care programs. Requirements Self hosted or cloud N8N Apple health vis iphone/watch Twilio VOIP phone number (to place a call) Workflows Core workflow for getting health data, processing and making a phone call. Twilio workflow to invoke Calls API to place an outbound voice call. twilio workflow { "name": "Twilio Bridge Caller copy", "nodes": [ { "parameters": { "httpMethod": "POST", "path": "twilio-call", "responseMode": "responseNode", "options": {} }, "type": "n8n-nodes-base.webhook", "typeVersion": 2, "position": [ 0, 0 ], "id": "ca3e6c69-3e7f-4d28-b699-4789a6fa2a6d", "name": "Webhook", "webhookId": "eb3d63df-800c-401d-931a-c6fba7d834ae" }, { "parameters": { "respondWith": "text", "responseBody": "={{ $json.body }}", "options": { "responseCode": 200, "responseHeaders": { "entries": [ { "name": "Content-Type", "value": "text/xml" } ] } } }, "type": "n8n-nodes-base.respondToWebhook", "typeVersion": 1.1, "position": [ 580, 0 ], "id": "6587b7e2-ace8-4e2b-9f4b-ed028a363c25", "name": "Respond to Webhook" }, { "parameters": { "jsCode": "const summary = $input.first().json.query.summary || 'No summary, check mail for critical health info';\n\nreturn [\n {\n json: {\n body: <Response>\n <Say voice=\"alice\">${summary}</Say>\n</Response>\n }\n }\n];\n" }, "type": "n8n-nodes-base.code", "typeVersion": 2, "position": [ 340, 0 ], "id": "0d4abf87-daf3-4533-8811-64ae61265f5d", "name": "Voice Twilio response" } ], "pinData": { "Webhook": [ { "json": { "headers": { "host": "n8n.domain.com", "user-agent": "curl/8.7.1", "content-length": "0", "accept": "/", "accept-encoding": "gzip, br" }, "params": {}, "query": { "lead": " 44711111111111" }, "body": {}, "webhookUrl": "https://n8n.domain.com/webhook/twilio-call", "executionMode": "production" } } ] }, "connections": { "Webhook": { "main": [ [ { "node": "Voice Twilio response", "type": "main", "index": 0 } ] ] }, "Voice Twilio response": { "main": [ [ { "node": "Respond to Webhook", "type": "main", "index": 0 } ] ] } }, "active": true, "settings": { "executionOrder": "v1" }, "versionId": "b58c5a12-75be-4b1d-b144-8c7251468021", "meta": { "instanceId": "8dc0e8a0878d0086b2f46ef04bb00ae07186c936d82d0f0a67563e9652996d33" }, "id": "RHaKqf8Wqt7fIuGH", "tags": [] } Samples Resources https://www.youtube.com/watch?v=HYk5_jtMlgc Questions/Support Contact me on info@pankstr.com.
by Samir Saci
Tags*: Sustainability, Supply Chain, AI Agent, CO2 Emissions, Carbon Interface API, Logistics, Automation Context Hi! I’m Samir — a Supply Chain Engineer and Data Scientist based in Paris, and founder of LogiGreen Consulting. I help logistics teams reduce their environmental footprint by combining AI automation and carbon estimation APIs. This workflow is part of our green logistics initiative, allowing businesses to track the CO₂ emissions of last-mile or regional shipments. > Automate carbon tracking for shipping operations with n8n! 📬 For business inquiries, feel free to connect with me on LinkedIn Who is this template for? This workflow is designed for logistics coordinators, transportation planners, or sustainability officers who want to estimate and record emissions for B2B shipments. Let’s imagine your carrier sends a shipment confirmation email after a pickup is scheduled: An AI Agent reads the email and extracts structured data: addresses, distance, cargo weight, and delivery time. The Carbon Interface API is then called to calculate CO₂ emissions based on weight and distance, and the results are stored in a Google Sheet. How does it work? This workflow automates the process of tracking CO₂ emissions for scheduled shipments: 📨 Gmail Trigger captures shipment confirmation emails 🧠 AI Agent parses the shipment info (pickup, delivery, weight, distance) 🚚 Carbon Interface API estimates CO₂ emissions 📊 Google Sheets is used to store shipment metadata and carbon results Steps: 💌 Trigger on new shipment confirmation email 🧠 Extract structured shipment info with AI Agent 📋 Store metadata in Google Sheets ⚙️ Call Carbon Interface API with weight and distance 📥 Append estimated CO₂ emissions to the shipment row What do I need to get started? You’ll need: A Gmail account to receive shipment confirmation emails A Google Sheet to track shipment data and CO₂ A free Carbon Interface API key OpenAI access for using the AI Agent parser A few sample emails from your logistics provider to test Next Steps 🗒️ Use the sticky notes in the n8n canvas to: Add your Gmail and Carbon Interface credentials Try with a sample shipment confirmation email Check your Google Sheet to verify emissions and timestamps This template was built using n8n v1.93.0 Submitted: June 7, 2025
by Airtop
LinkedIn Post Engagement Data Extractor Use Case This automation is designed to extract key engagement metrics and audience data from a LinkedIn post. It's useful for analyzing the impact of content and identifying engaged users for lead generation, marketing, or research purposes. What It Does Given a LinkedIn post URL and an Airtop profile, this automation extracts: Total number of reactions Total number of comments Total number of reposts A list of users who reacted or commented, including: Their full name Their job title A link to their LinkedIn profile Input Parameters | Name | Description | Required | |------------------|------------------------------------------------------------|----------| | airtop_profile | The name of an Airtop Profile that's logged into LinkedIn | Yes | | linkedin_post_url| The full URL of the LinkedIn post you want to analyze | Yes | How It Works The workflow starts when triggered manually or from another workflow/form. It maps input fields for Airtop profile and post URL. Airtop opens a browser session and loads the LinkedIn post. An AI agent is instructed to extract engagement data via prompt-based analysis. The response is parsed and output in a structured format. Output Format The output will be a structured JSON object with the following fields: { "interactors": [ { "name": "Jane Doe", "job_title": "Marketing Director at ExampleCorp", "profile_url": "https://linkedin.com/in/janedoe" } // ... more interactors ], "reactions_count": 153, "comments_count": 21, "reposts_count": 8 } Read more about how to extract Linkedin post comments and reactions.
by Agent Circle
This N8N template demonstrates how to use our tool to collect key information from any YouTube channel - including title, description, custom URL, stats (views, subscribers, videos), creation date, country, keywords, thumbnails, and channel ID. Everything is fetched and saved directly in a linked Google Sheet. Use cases are many: Whether you're part of an MCN or YouTube channel network, a digital marketing agency, an influencer analyst, or a growth-focused YouTube consultant, this tool helps you instantly access detailed channel data for scouting, benchmarking, and strategic planning. It's perfect for competitive research, brand vetting, or building your next performance dashboard. How It Works The workflow starts when you manually click Test Workflow or Execute Workflow in N8N. It reads through the list of full channel URLs or custom channel URLs from the Channel URLs tab in a connected Google Sheet. Only the URLs marked with the Ready status will be processed. The tool loops through each row and prepares the necessary data for the YouTube API call later. A Switch node detects the URL type of each input (whether it's a full or custom channel URL) and routes it accordingly. Based on the input type, the appropriate YouTube API endpoint is triggered to fetch channel metrics. It checks whether the API call is successful before continuing: If successful, the fetched data is written back to the connected Google Sheet in the same row and the row's status is updated to Finished. If it fails, the row's status is updated to Error. How To Use Download the workflow package. Import the workflow package into your N8N interface. Duplicate the YouTube - Get Channel Information Google Sheet template into your Google Sheets account. Set up Google Cloud Console credentials in the following nodes in N8N, ensuring enabled access and suitable rights to Google Sheets and YouTube services: For Google Sheets access, ensure each node is properly connected to the correct tab in your connected Google Sheet template: Node Google Sheets - Get Channel URLs → connected to Tab Channel URLs; Node Google Sheets - Update Data → connected to Tab Channel URLs; Node Google Sheets - Update Data - Error → connected to Tab Channel URLs. For YouTube access, set up a GET method to connect to YouTube API in the following nodes: Node HTTP Request - Get Comments For Video URLs 1; Node HTTP Request - Get Comments For Video URLs 2. In your connected Google Sheet, enter the full or custom channel URLs that you want to crawl and set the rows' status to Ready. Run the workflow by clicking Execute Workflow or Test Workflow in N8N. View the results in your Google Sheet: Successful fetches will update channel data in the same row in Channel URLs tab and the row's status is updated to Finished; otherwise, the row's status will be changed to Error. Requirements Basic setup in Google Cloud Console (OAuth or API Key method enabled) with enabled access to YouTube and Google Sheets. How To Customize By default, the workflow is manually triggered in N8N. However, you can automate the process by adding a Google Sheets trigger that monitors new entries in your YouTube – Get Channel Information template and starts the workflow automatically. Need Help? Join our community on different platforms for support, inspiration and tips from others. Website: https://www.agentcircle.ai/ Etsy: https://www.etsy.com/shop/AgentCircle Gumroad: http://agentcircle.gumroad.com/ Discord Global: https://discord.gg/d8SkCzKwnP FB Page Global: https://www.facebook.com/agentcircle/ FB Group Global: https://www.facebook.com/groups/aiagentcircle/ X: https://x.com/agent_circle YouTube: https://www.youtube.com/@agentcircle LinkedIn: https://www.linkedin.com/company/agentcircle
by Samir Saci
Tags*: Crypto, Currency Exchange, Alpha Vantage API, Google Sheets Context Hi! I’m Samir — a Supply Chain Engineer and Data Scientist based in Paris, and founder of LogiGreen Consulting. I help companies automate data pipelines using APIs, AI agents, and workflow automation to improve operational visibility and decision-making. This workflow is part of our financial monitoring toolkit, allowing businesses to track real-time cryptocurrency exchange rates for assets like BTC and ETH. > Monitor crypto rates in real time and log the data to Google Sheets. 📬 For business inquiries, feel free to connect with me on LinkedIn Who is this template for? This workflow is designed for crypto analysts, finance teams, or automation enthusiasts who want to collect and track crypto exchange rates automatically for reporting or analysis. Let’s imagine you want to monitor BTC and ETH prices against the EUR: The Alpha Vantage API is called every hour to collect the exchange rate, bid/ask prices, and timestamp. This data is saved in Google Sheets and can trigger a Telegram alert. How does it work? This workflow automates the hourly logging of crypto exchange rates: ⏰ Schedule Trigger launches the workflow every hour 🌐 Alpha Vantage API fetches BTC and ETH exchange rates (to EUR) 📄 Google Sheets node appends each result to a tab (BTC or ETH) 📲 Telegram node sends a summary alert with the latest rate info Steps: ⏰ Trigger every hour via Schedule node 🌐 Call Alpha Vantage API for BTC and ETH 📊 Append rates, bid/ask prices, and timestamp to Google Sheets 📲 Send formatted Telegram message with rate summary What do I need to get started? You’ll need: A free Alpha Vantage API key A Google Sheet with BTC and ETH tabs A Telegram bot and Chat ID to send notifications Next Steps 🗒️ Use the sticky notes in the n8n canvas to: Add your Alpha Vantage API key and Google Sheet credentials Set your Telegram Chat ID and bot connection Monitor and visualize real-time crypto prices with no manual input This template was built using n8n v1.93.0 Submitted: June 12, 2025
by Tamer
Gmail Daily Summary Scheduler - Template Description Never miss important emails again! This intelligent automation runs every morning at 8 AM to analyze your Gmail inbox from the past 24 hours and delivers a comprehensive summary powered by Google Gemini AI. 🚀 What it does: Automatically triggers daily at 8 AM (customizable) Fetches Gmail emails from the last 24 hours AI analyzes all emails to identify themes and key updates Highlights priority emails with reasons why they're important Delivers structured output in consistent JSON format 💡 Perfect for: Busy professionals managing high email volumes Team leads who need quick inbox overviews Anyone wanting to prioritize email responses efficiently Remote workers staying on top of communications 🎯 Key Features: Smart filtering - Only processes emails from the last 24 hours AI-powered analysis - Uses Google Gemini for intelligent summaries Priority detection - Identifies urgent, time-sensitive emails Consistent output - Structured JSON with summary and highlights Fully automated - Set it and forget it 📋 Sample Output: Summary: "Today's emails covered 3 project updates, 2 meeting requests, and 1 urgent client issue requiring immediate attention." Important Emails: Subject: "Project Deadline Extension Request" | Sender: client@company.com | Reason: Client requesting timeline change for major deliverable due this week Subject: "Tomorrow's Board Meeting Moved" | Sender: assistant@office.com | Reason: Last-minute schedule change requiring immediate calendar adjustment 🛠 Prerequisites: Gmail account with API access Google Gemini API credentials Basic n8n knowledge for credential setup ⚙️ Easy to Customize: Change schedule time (default: 8 AM daily) Adjust time window (default: 24 hours) Modify AI analysis focus Add email filters by sender, subject, etc. Extend with notifications (Slack, email, etc.) 🔒 Privacy Note: This workflow processes email content through Google Gemini API. Ensure compliance with your organization's data policies. Save hours of manual email sorting every week! Import this template and start getting intelligent email summaries delivered automatically every morning.
by Hostinger
This workflow automates the routine check for upgradable packages on your Ubuntu server, ensuring you stay updated with the latest software patches and security improvements. By running a daily script, it efficiently monitors any available package upgrades and promptly notifies you via email, saving you time and enhancing your server’s security. How It Works: Daily Monitoring**: The workflow is configured to execute a script daily that connects to your Ubuntu server and checks for any upgradable packages. Email Notification**: If any upgradable packages are detected during the check, the workflow triggers an alert mechanism that automatically sends you a notification email detailing the available updates. Set Up Steps: SSH Credentials**: Provide the SSH login credentials for your Ubuntu server. This will allow the workflow to securely connect and perform checks for software updates. SMTP Credentials**: Provide SMTP login details for your email account. These credentials are used to configure the email notifications system, enabling it to send alerts about the upgradable packages. Benefits: Timeliness**: Receive prompt updates on critical software upgrades to maintain the optimal performance and security of your server. Automation**: Reduces the need for manual checks, allowing you to focus on other critical tasks with peace of mind. Customizable**: Easily adjust the checking frequency or update the notification settings according to your preferences.
by Afnan
This n8n workflow automates the process of finding, summarizing, and posting breaking news headlines on X (formerly Twitter). It combines Google Custom Search for finding the latest news articles with Groq's LLaMA 3 model to generate short, engaging headlines — complete with hashtags — and posts them on your X account. 🔧 Features Custom topic support (e.g., "AI", "health", "technology") Automated scheduling every few hours Google Custom Search to find the most recent news articles Groq LLaMA3-based headline generation with hashtags Auto-post to X (Twitter) Built-in credential separation for API keys and access tokens 📦 Included Nodes Schedule Trigger Set (Set Topic, Google API Key, Custom Search CX, etc.) HTTP Request (Google Search API) Code Node (Format prompt and extract article data) HTTP Request (Groq API for headline generation) Twitter Node (Post to X) ⚙️ How It Works (Step-by-Step) Trigger The workflow starts on a scheduled interval (default: every 5 hours, at a random minute within the hour). Set Topic You can define your own topic keyword (e.g., AI, mental health, climate change) by editing the Set Topic node. Build Search Query Constructs a Google search query like: latest {topic} news. Google API Config Injects your own Google API Key and Custom Search CX (replace the placeholders in the Google Config node). Search for News Performs a real-time search using Google Custom Search API and fetches the latest article result. Generate Prompt for AI A JavaScript Function node extracts the top article’s title and link, formats it into a clean prompt including instructions to append hashtags. Groq AI Request Sends the prompt to Groq’s LLaMA 3 model to generate a concise, tweet-length headline with 1–2 relevant hashtags. Post to Twitter (X) The generated headline is posted to your connected X account via the Twitter OAuth2 API. ✅ Requirements Google API Key Google Custom Search Engine (CX) Groq API Key Twitter Developer App with OAuth2 credentials 💡 Customization Tips Change the topic in the Set Topic node to anything you like. Adjust the posting frequency in the Schedule Trigger node. Modify prompt behavior in the Function node to fit a specific tone or brand voice. Add logging, filtering, or multiple post variations as needed.
by Pauline
This workflow automatically finds verified and nominative email and all information (with Dropcontact) to prospect properly from a Typeform. Typeform Trigger**: This node triggers the workflow when a new Typeform is submitted. Dropcontact**: This node fetches the new contact information and returns the data of the person and the company associated with the email address, job function, and all legal information. Airtable - List**: This will list all the entries from your table in order to check if the subscriber is new or not. If**: Does the contact exist in your Airtable yet? Set**: Set values on the Airtable values Airtable Update**: This node updates your record with new and up-to-date data Airtable - Create**: This node creates a new record when the subscriber is a new lead Slack**: This node sends an alert when the record is added or updated
by Roshan Ramani
📧 Morning Mail Summary Ai Agent Simplify your mornings with this automated email summary workflow in n8n, designed for teams and individuals who want a quick, actionable snapshot of their inbox—every day, at a glance. 📌 Features Fully Automated**: Triggers daily at 8 AM IST Targeted Extraction**: Fetches messages from specific senders with dynamic date filtering Structured Data Aggregation**: Pulls key fields (sender, recipients, snippet) AI-Powered Summarization**: Uses LangChain + OpenRouter to generate: Insightful summaries Highlighted issues Assigned action items Memory-Enhanced Context**: Maintains continuity across days Stylish HTML Email Output**: Responsive design with: Summary section Action items Branded footer Team-Ready Sharing**: Auto-distributes to team with CC support 🛠️ How It Works Trigger: Daily schedule in n8n Fetch: Retrieves last 24h emails from target sender Organize: Aggregates email metadata Summarize: AI agent produces structured report: Key updates Critical issues Actionable tasks Send: Emails digest to team 🎯 Ideal For Project managers needing stakeholder briefings Customer success teams tracking communications Leadership staying informed efficiently Anyone streamlining email review 📦 Included Resources Complete n8n workflow JSON (6 nodes) Prompt template for structured summarization (Summary, Issues, Actions, Follow‑Ups). -Styled HTML email template, with responsive design and branding sections. 🆓 Get Started Import JSON into n8n Configure: Gmail / OpenRouter credentials Recipient lists Custom prompts Modify freely for your use case ⭐ Benefits ⏳ Time saved: Skip inbox digging 🔍 Focus: See only what matters ✨ Clarity: Structured tasks & deadlines 🔄 Consistency: Daily alignment for teams
by Wyeth
How it Works: You can now use the Slack node to create conditional breakpoints! This example shows the loop stop on 4 (of 10) and then you can hit "continue" in Slack when you are ready. You could extend this with a Form node and inject values you type in on the fly. Slack was already the best place to debug log n8n, and now it just got better!
by Evoort Solutions
🎥 YouTube Video Summarizer for Social Media Turn any YouTube video into a short, structured summary using AI — perfect for content creators, marketers, or social media managers. 🔧 What We Built We created a no-code automation in n8n that: Accepts a YouTube Video ID via a form Fetches the video transcript using an external API Summarizes the transcript using AI (Google Gemini) Automatically saves the summary to Google Docs for team use 🧩 Flow Overview | Step | Description | |------|-------------| | ✅ Form Trigger | User submits a YouTube video ID using an n8n form | | 🔁 Set Node | Maps the YouTube video ID for use in the API request | | 🌐 HTTP Request (External API) | Calls the YouTube Transcriptor AI API via RapidAPI to fetch transcript | | 🧹 Formatter (Code Node) | Joins transcript lines into a readable text block | | 🧠 AI Agent + Google Gemini (via Langchain) | Summarizes the full transcript into bullet points and tone | | 🧽 Optimizer (Code Node) | Extracts just the summary from the AI response | | 📝 Google Docs Node | Appends the clean summary to a shared Google Doc | 🌍 Real-World Problem Solved ❌ The Challenge Creators and marketers waste hours watching full videos just to extract the key points. Manual summarization is inconsistent, repetitive, and delays content planning. ✅ Our Solution ⏱️ Reduces time spent watching videos 🧠 AI-powered summaries keep tone consistent and structured 📄 Auto-sync with Google Docs makes summaries instantly available for teams 🔥 Bonus: This uses the YouTube Transcriptor AI API, so no need to manually scrape captions or use browser extensions. 🚀 Ideal Use Cases Repurpose YouTube content into Instagram Reels, LinkedIn posts, or blog content Build a video summary library for your editorial team Quickly extract talking points from podcast episodes 🛠️ Tech Stack n8n** – workflow automation engine YouTube Transcriptor AI API** – via RapidAPI Google Gemini (via Langchain)** – AI summarization Google Docs** – stores the final summary JavaScript nodes** – custom text parsing & formatting 💡 Want to customize it? Add Slack, Airtable, Notion, or Tweet auto-posting to expand the flow. Create your free n8n account and set up the workflow in just a few minutes using the link below: 👉 Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!