by Jaruphat J.
β οΈ Note: This template requires a community node and works only on self-hosted n8n installations. It uses the Typhoon OCR Python package and custom command execution. Make sure to install required dependencies locally. Who is this for? This template is for developers, operations teams, and automation builders in Thailand (or any Thai-speaking environment) who regularly process PDFs or scanned documents in Thai and want to extract structured text into a Google Sheet. It is ideal for: Local government document processing Thai-language enterprise paperwork AI automation pipelines requiring Thai OCR What problem does this solve? Typhoon OCR is one of the most accurate OCR tools for Thai text. However, integrating it into an end-to-end workflow usually requires manual scripting and data wrangling. This template solves that by: Running Typhoon OCR on PDF files Using AI to extract structured data fields Automatically storing results in Google Sheets What this workflow does Trigger: Run manually or from any automation source Read Files: Load local PDF files from a doc/ folder Execute Command: Run Typhoon OCR on each file using a Python command LLM Extraction: Send the OCR markdown to an AI model (e.g., GPT-4 or OpenRouter) to extract fields Code Node: Parse the LLM output as JSON Google Sheets: Append structured data into a spreadsheet Setup 1. Install Requirements Python 3.10+ typhoon-ocr: pip install typhoon-ocr Install Poppler and add to system PATH (needed for pdftoppm, pdfinfo) 2. Create folders Create a folder called doc in the same directory where n8n runs (or mount it via Docker) 3. Google Sheet Create a Google Sheet with the following column headers: | book\_id | date | subject | detail | signed\_by | signed\_by2 | contact | download\_url | | -------- | ---- | ------- | ------ | ---------- | ----------- | ------- | ------------- | You can use this example Google Sheet as a reference. 4. API Key Export your TYPHOON_OCR_API_KEY and OPENAI_API_KEY in your environment (or set inside the command string in Execute Command node). How to customize this workflow Replace the LLM provider in the Basic LLM Chain node (currently supports OpenRouter) Change output fields to match your data structure (adjust the prompt and Google Sheet headers) Add trigger nodes (e.g., Dropbox Upload, Webhook) to automate input About Typhoon OCR Typhoon is a multilingual LLM and toolkit optimized for Thai NLP. It includes typhoon-ocr, a Python OCR library designed for Thai-centric documents. It is open-source, highly accurate, and works well in automation pipelines. Perfect for government paperwork, PDF reports, and multilingual documents in Southeast Asia.
by ist00dent
This n8n template lets you instantly serve batches of inspirational quotes via a webhook using the free ZenQuotes API. Itβs perfect for developers, content creators, community managers, or educators who want to add dynamic, uplifting content to websites, chatbots, or internal toolsβwithout writing custom backend code. π§ How it works A Webhook node listens for incoming HTTP requests on your chosen path. Get Random Quote from ZenQuotes sends an HTTP Request to https://zenquotes.io/api/random?count=5 and retrieves five random quotes. Format data uses a Set node to combine each quote (q) and author (a) into a single string: "βquoteβ β author". Send response returns a JSON array of objects { quote, author } back to the caller. π€ Who is it for? This workflow is ideal for: Developers building motivational Slack or Discord bots. Website owners adding on-demand quote widgets. Educators or trainers sharing daily inspiration via webhooks. Anyone learning webhook handling and API integration in n8n. ποΈ Response Structure Your webhook response will be a JSON array, for example: [ { "quote": "Life is what happens when you're busy making other plans.", "author": "John Lennon" }, { "quote": "Be yourself; everyone else is already taken.", "author": "Oscar Wilde" } ] βοΈ Setup Instructions Import the workflow JSON into your n8n instance. In the Webhook node, set your desired path (e.g., /inspire). (Optional) Change the count parameter in the HTTP Request node to fetch more or fewer quotes. Activate the workflow. Test by sending an HTTP GET or POST to https://<your-n8n-domain>/webhook/<path>.
by Lucas Peyrin
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. How it works This workflow demonstrates how to create a resilient AI Agent that automatically falls back to a different language model if the primary one fails. This is useful for handling API errors, rate limits, or model outages without interrupting your process. State Initialization: The Agent Variables node initializes a fail_count to 0. This counter tracks how many models have been attempted. Dynamic Model Selection: The Fallback Models (a LangChain Code node) acts as a router. It receives a list of all connected AI models and, based on the current fail_count, selects which one to use for this attempt (0 for the first model, 1 for the second, etc.). Agent Execution: The AI Agent node attempts to run your prompt using the model selected by the router. The Fallback Loop: On Success: The workflow completes successfully. On Error: If the AI Agent node fails, its "On Error" output is triggered. This path loops back to the Agent Variables node, which increments the fail_count by 1. The process then repeats, causing the Fallback Models router to select the next model in the list. Final Failure: If all connected models are tried and fail, the workflow will stop with an error. Set up steps Setup time: ~3-5 minutes Configure Credentials: Ensure you have the necessary credentials (e.g., for OpenAI, Google AI) configured in your n8n instance. Define Your Model Chain: Add the AI model nodes you want to use to the canvas (e.g., OpenAI, Google Gemini, Anthropic). Connect them to the Fallback Models node. Important: The order in which you connect the models determines the fallback order. The model nodes first created/connected will be tried first. Set Your Prompt: Open the AI Agent node and enter the prompt you want to execute. Test: Run the workflow. To test the fallback logic, you can temporarily disable the First Model node or configure it with invalid credentials to force an error.
by Mihai Farcas
This n8n workflow automates the process of saving web articles or links shared in a chat conversation directly into a Notion database, using Google's Gemini AI and Browserless for web scraping. Who is this AI automation template for? It's useful for anyone wanting to reduce manual copy-pasting and organize web findings seamlessly within Notion. A smarter web clipping tool! What this AI automation workflow does Starts when a message is received Uses a Google Gemini AI Agent node to understand the context and manage the subsequent steps. It identifies if a message contains a request to save an article/link. If a URL is detected, it utilizes a tool configured with the Browserless API (via the HTTP Request node) to scrape the content of the web page. Creates a new page in a specified Notion database, populating it with thea summary scraped content, in a specific format, never leaving out any important details. It also saves the original URL, smart tags, publication date, and other metadata extracted by the AI. Posts a confirmation message (e.g., to a Discord channel) indicating whether the article was saved successfully or if an error occurred. Setup Import Workflow: Import this template into your n8n instance. Configure Credentials & Notion Database: Notion Database: Create or designate a Notion database (like the example "Knowledge Database") where articles will be saved. Ensure this database has the following properties (fields): Name (Type: Text) - This will store the article title. URL (Type: URL) - This will store the original article link. Description (Type: Text) - This can store the AI-generated summary. Tags (Type: Multi-select) - Optional, for categorization. Publication Date (Type: Date) - *Optional, store the date the article was published. Ensure the n8n integration has access to this specific database. If you require a different format to the Notion Database, not that you will have to update the Notion tool configuration in this n8n workflow accordingly. Notion Credential: Obtain your Notion API key and add it as a Notion credential in n8n. Select this credential in the save_to_notion tool node. Configure save_to_notion Tool: In the save_to_notion tool node within the workflow, set the 'Database ID' field to the ID of the Notion database you prepared above. Map the workflow data (URL, AI summary, etc.) to the corresponding database properties (URL, Description, etc.). In the blocks section of the notion tool, you can define a custom format for the research page, allowing the AI to fill in the exact details you want extracted from any web page! Google Gemini AI: Obtain your API key from Google AI Studio or Google Cloud Console (if using Vertex AI) and add it as a credential. Select this credential in the "Tools Agent" node. Discord (or other notification service): If using Discord notifications, create a Webhook URL (instructions) or set up a Bot Token. Add the credential in n8n and select it in the discord_notification tool node. Configure the target Channel ID. Browserless/HTTP Request: Cloud: Obtain your API key from Browserless and configure the website_scraper HTTP Request tool node with the correct API endpoint and authentication header. Self-Hosted: Ensure your Browserless Docker container is running and accessible by n8n. Configure the website_scraper HTTP Request tool node with your self-hosted Browserless instance URL. Activate Workflow: Save test and activate the workflow. How to customize this workflow to your needs Change AI Model:** Experiment with different AI models supported by n8n (like OpenAI GPT models or Anthropic Claude) in the Agent node if Gemini 2.5 Pro doesn't fit your needs or budget, keeping in mind potential differences in context window size and processing capabilities for large content. Modify Notion Saving:** Adjust the save_to_notion tool node to map different data fields (e.g., change the summary style by modifying the AI prompt, add specific tags, or alter the page content structure) to your Notion database properties. Adjust Scraping:** Modify the prompt/instructions for the website_scraper tool or change the parameters sent to the Browserless API if you need different data extracted from the web pages. You could also swap Browserless for another scraping service/API accessible via the HTTP Request node.
by Hueston
Who is this for? Content strategists analyzing web page semantic content SEO professionals conducting entity-based analysis Data analysts extracting structured data from web pages Marketers researching competitor content strategies Researchers organizing and categorizing web content Anyone needing to automatically extract entities from web pages What problem is this workflow solving? Manually identifying and categorizing entities (people, organizations, locations, etc.) on web pages is time-consuming and error-prone. This workflow solves this challenge by: Automating the extraction of named entities from any web page Leveraging Google's powerful Natural Language API for accurate entity recognition Processing web pages through a simple webhook interface Providing structured entity data that can be used for analysis or further processing Eliminating hours of manual content analysis and categorization What this workflow does This workflow creates an automated pipeline between a webhook and Google's Natural Language API to: Receive a URL through a webhook endpoint Fetch the HTML content from the specified URL Clean and prepare the HTML for processing Submit the HTML to Google's Natural Language API for entity analysis Return the structured entity data through the webhook response Extract entities including people, organizations, locations, and more with their salience scores Setup Prerequisites: An n8n instance (cloud or self-hosted) Google Cloud Platform account with Natural Language API enabled Google API key with access to the Natural Language API Google Cloud Setup: Create a project in Google Cloud Platform Enable the Natural Language API for your project Create an API key with access to the Natural Language API Copy your API key for use in the workflow n8n Setup: Import the workflow JSON into your n8n instance Replace "YOUR-GOOGLE-API-KEY" in the "Google Entities" node with your actual API key Activate the workflow to enable the webhook endpoint Copy the webhook URL from the "Webhook" node for later use Testing: Use a tool like Postman or cURL to send a POST request to your webhook URL Include a JSON body with the URL you want to analyze: {"url": "https://example.com"} Verify that you receive a response containing the entity analysis data How to customize this workflow to your needs Analyzing Specific Entity Modify the "Google Entities" node parameters to include entityType filters Add a "Function" node after "Google Entities" to filter specific entity types Create conditions to extract only entities of interest (people, organizations, etc.) Processing Multiple URLs in Batch: Replace the webhook with a different trigger (HTTP Request, Google Sheets, etc.) Add a "Split In Batches" node to process multiple URLs Use a "Merge" node to combine results before sending the response Enhancing Entity Data: Add additional API calls to enrich extracted entities with more information Implement sentiment analysis alongside entity extraction Create a data transformation node to format entities by type or relevance Additional Notes This workflow respects Google's API rate limits by processing one URL at a time The Natural Language API may not identify all entities on a page, particularly for highly technical content HTML content is trimmed to 100,000 characters if longer to avoid API limitations Consider legal and privacy implications when analyzing and storing entity data from web pages You may want to adjust the HTML cleaning process for specific website structures β€οΈ Hueston SEO Team
by Halfbit π
Daily YouTrack In-Progress Tasks Summary to Discord by Assignee Keep your team in sync with a daily summary of tasks currently In Progress in YouTrack β automatically posted to your Discord channel. This workflow queries issues, filters them by status, groups them by assignee and priority, and sends a formatted message to Discord. It's perfect for teams that need a lightweight, automated stand-up report. > π This workflow uses Discord as an example. You can easily replace the messaging integration with Slack, Mattermost, MS Teams, or any other platform that supports incoming webhooks. Use Case Remote development teams using YouTrack + Discord Replacing daily stand-up meetings with async updates Project managers needing quick visibility into active tasks Features Scheduled** daily execution (default: weekdays at 09:00) Status filter**: only issues marked as In Progress Grouping** by assignee and priority Custom mapping** for user mentions (YouTrack β Discord) Clean Markdown output** for Discord, with direct task links Setup Instructions YouTrack Configuration Get a permanent token: Go to your YouTrack profile β Account Security β Authentication Create a new permanent token with "Read Issue" permissions Copy the token value Set the base API URL: Format: https://yourdomain.youtrack.cloud/api/issues Replace yourdomain with your actual YouTrack instance Identify custom field IDs: Method 1: Go to YouTrack β Administration β Custom Fields β find your "Status" field and note its ID Method 2: Use API call GET /api/admin/customFieldSettings/customFields to list all field IDs Method 3: Inspect a task's API response and look for field IDs in the customFields array Example Status field ID: 105-0 or 142-1 Discord Configuration Create a webhook URL in your Discord server: Server Settings β Integrations β Webhooks β New Webhook Choose target channel and copy the webhook URL Extract webhook ID from URL (numbers after /webhooks/) Environment Variables & Placeholders | Placeholder | Description | |-------------|-------------| | {{API_URL}} | Your YouTrack API base URL | | {{TOKEN}} | YouTrack permanent token | | {{FIELD_ID}} | ID of the "Status" custom field | | {{QUERY_FIELDS}} | Fields to fetch (e.g., summary, id) | | {{PROJECT_LINK}} | Link to your YouTrack project | | {{USER_X}} | YouTrack usernames | | {{DISCORD_ID_X}} | Discord mentions or usernames | | {{NAME_X}} | Display names | | {{WEBHOOK_ID}} | Discord webhook ID | | {{DISCORD_CHANNEL}} | Discord channel name | | {{CREDENTIAL_ID}} | Your credential ID in n8n | Testing the Workflow Test YouTrack connection: Execute the "HTTP Request YT" node individually Verify that issues are returned from your YouTrack instance Check if the Status field ID is correctly filtering tasks Verify filtering: Run the "Filter fields" node Confirm only "In Progress" tasks pass through Check message formatting: Execute the "Discord message" node Review the generated message content and formatting Test Discord delivery: Run the complete workflow manually Verify the message appears in your Discord channel Schedule verification: Enable the workflow Test weekend skip functionality by temporarily changing dates Customization Tips Language**: All labels/messages are in English β customize if needed User mapping**: Adjust assignee β Discord mention logic in the message builder Priorities**: Update the priorityMap to reflect your own naming structure Schedule**: Modify the trigger time in the Schedule Trigger node Alternative platforms**: Swap out the Discord webhook for another messaging service if preferred
by Leonard
Who is this for? This workflow is designed for SEO specialists, content creators, marketers, and website developers who want to ensure their web content is easily accessible, understandable, and indexable by Large Language Models (LLMs) like ChatGPT, Perplexity, and Google AI Overviews. If you're looking to optimize your site for the evolving AI-driven search landscape, this template is for you. What problem is this workflow solving? / Use case Modern AI tools often crawl websites without executing JavaScript. This can lead to them "seeing" a very different version of your page than a human user or traditional search engine bot might. This workflow helps you: Identify how much of your content is visible without JavaScript. Check for crucial on-page SEO elements that AI relies on (headings, meta descriptions, structured data). Detect if your site presents JavaScript-blocking warnings. Get an AI-generated readability score and actionable recommendations to improve AI-friendliness. What this workflow does Receives a URL via a chat interface. Sanitizes the input URL to ensure it's correctly formatted. Fetches the website's HTML content, simulating a non-JavaScript crawler (like Googlebot). Extracts key HTML features: visible text length, presence of H1/H2/H3 tags, meta description, Open Graph data, structured data (JSON-LD), and <noscript> tags. It also checks for common JavaScript-blocking messages. Performs an AI SEO Analysis using an LLM (via OpenAI) based on the extracted features. Provides a report including an AI Readability Score (0-10), a summary, actionable recommendations, and a reminder to check the robots.txt file for AI bot access. Setup Estimated setup time:** 2-5 minutes. Import this workflow into your n8n instance. Ensure you have an OpenAI account and API key. Configure the "OpenAI Chat Model" node with your OpenAI API credentials. If you don't have credentials set up yet, create new ones in n8n. Activate the workflow. Interact with the chat interface provided by the "When chat message received" trigger node (you can access this via its webhook URL). How to customize this workflow to your needs Change LLM Model:** In the "OpenAI Chat Model" node, you can select a different model that suits your needs or budget. Adjust AI Prompt:** Modify the prompt in the "AI SEO Analysis" node (Chain Llm) to change the focus of the analysis or the format of the report. For example, you could ask for more technical details or a different scoring system. User-Agent:** The "Get HTML from Website" node uses a Googlebot User-Agent. You can change this to simulate other bots if needed. JS Block Indicators:** The "Extract HTML Features" node contains a list of common JavaScript-blocking phrases. You can expand this list with other languages or specific messages relevant to your checks.
by Ria
This workflow demonstrates how to use the workflowStaticData() function to set any type of variable that will persist within workflow executions. https://docs.n8n.io/code/cookbook/builtin/get-workflow-static-data/ This can be useful for example when working with access tokens that expire after a certain time period. Using staticData we can keep a record of that access token and the expiry time and build our workflow logic around it. Important Static Data only persists across production executions, i.e. triggered by Webhooks or Schedule Triggers (not manual executions!) For this the workflow will have to be activated. Setup configure HTTP Request node to fetch access token from your API (optional) activate workflow test the workflow with the webhook production link you can check the population of the static data in the single executions Feedback If you found this useful or want to report some missing information - I'd be happy to hear from you at ria@n8n.io
by Marcelo Abreu
Who is this workflow for? If you're using Meta Ads to generate new leads to your sales pipeline, this workflow is for you! ππ» What this workflow does Triggers every time you have a new calendar event on a chosen Google Acount Filter only events with the same name of your "Schedule a demo" event Formats and send event to Meta Conversion API What events can I send? Any event you'd like! It's preconfigured with the "Schedule" event, but you can change to "Purchase", "InitiateCheckout", "Lead" and custom events. Setup Guide Connect Google OAuth2 to n8n Get your PIXEL ID and Access Token from Meta Set your configuration node with Pixel ID, Access Token, source_url and event_name Requirements Meta Access Token + Pixel ID (via Meta Conversion API): Documentation Google Access (via OAuth2): Documentation This free template was created by pdforge. Feel free to contact us via the founder Linkedin, if you have any questions! ππ»
by JaredCo
Real-time Weather Forecasts with MCP Tools This n8n workflow demonstrates how to integrate real-time weather intelligence into any automation using the Model Context Protocol (MCP). Get current conditions and 5-day forecasts with natural language queries like "What's the weather like in Miami?" or "Will it rain next Tuesday in Seattle?" - all powered by live weather data and AI. Good to know No API keys required - uses hosted MCP weather server with built-in WorldWeatherOnline integration Provides current conditions and detailed 5-day forecasts Natural language queries work for any location worldwide Powered by WorldWeatherOnline - the world's most accurate weather system Fully preconfigured and ready to run out-of-the-box Enterprise-ready with error handling and rate limiting How it works Natural Language Input**: Receives weather queries via webhook, chat, email, or voice AI Agent Processing**: n8n Agent node interprets requests and determines: Location extraction from natural language Weather data type needed (current or 5-day forecast) Response formatting preferences MCP Weather Tool**: Live hosted server provides: Real-time current conditions (temperature, humidity, wind, conditions) 5-day detailed forecasts with daily highs/lows Weather descriptions and condition codes Powered by WorldWeatherOnline's premium data Intelligent Responses**: AI formats weather data into: Conversational natural language responses Structured data for downstream automation Action-triggering data for workflows How to use Import the workflow into n8n from the template Add your preferred AI model API key to the Agent node Customize the system prompt for your specific use case Connect to your preferred input/output channels Run and start querying weather with natural language Use Cases Smart Home Automation**: "Turn on sprinklers if no rain forecast for 3 days" Travel Planning**: "Check weather for my Paris trip next week" Event Management**: "Will outdoor wedding conditions be good Saturday?" Agriculture/Farming**: "Check 5-day forecast for planting schedule" Logistics**: "Delay shipping if severe weather forecast in delivery zone" Personal Assistant**: "Should I wear a jacket today in Chicago?" Sports/Recreation**: "Surf conditions and wind forecast for weekend" Construction**: "Safe working conditions for outdoor project this week" Requirements n8n instance (cloud or self-hosted) AI model provider account (OpenAI, Anthropic, Google, etc.) Internet connection for MCP weather server access Optional: Webhook endpoints for external integrations Customizing this workflow Location Intelligence**: Add geocoding for address-to-coordinates conversion Data Storage**: Save weather history to databases for trend analysis Dashboard Integration**: Connect to Grafana, Tableau, or custom visualizations Voice Integration**: Add speech-to-text for voice weather queries Scheduling**: Set up automated daily/weekly weather briefings Conditional Logic**: Trigger different actions based on weather conditions Sample Input/Output Natural Language Queries: "What's the weather like in Miami?" "Will it rain next Tuesday in Seattle?" "5-day forecast for London" "Temperature in Tokyo tomorrow" "Weather conditions for outdoor event Saturday" Rich Responses: { "location": "Miami, FL", "current": { "temperature": "78Β°F", "condition": "Partly Cloudy", "humidity": "65%", "wind": "10 mph SE" }, "forecast": { "today": "High 82Β°F, Low 71Β°F, 20% rain", "tomorrow": "High 85Β°F, Low 73Β°F, Sunny" }, "ai_summary": "Perfect beach weather in Miami today! Partly cloudy with comfortable temperatures and light winds." } Why This Workflow is Unique Zero Setup Weather Data**: No API key management - MCP server handles everything World-Class Accuracy**: Powered by WorldWeatherOnline's premium weather data AI-Powered Intelligence**: Natural language understanding of complex weather queries Enterprise Ready**: Built-in error handling, rate limiting, and reliability Global Coverage**: Worldwide weather data with location intelligence Action-Oriented**: Designed for automation decisions, not just information display Transform your automations with intelligent weather awareness powered by the world's most accurate weather system! π§ͺ Setup Steps β The Agent node is already configured: The system prompt is included The tool endpoint is pre-set All you need to do is: Add your AI model API key to the existing Agent credential Hit run and you're done β π Full project link: Github: weathertrax-mcp-agent-demo
by Alex Dunlop
Who is this for? Professionals and individuals who receive high volumes of emails, those who want to automatically organize their Gmail inbox using AI classification. What problem is this workflow solving? Manual email sorting is time-consuming and inconsistent. This workflow automatically categorizes incoming emails into 8 predefined labels (To respond, FYI, Comment, Notification, Meeting update, Awaiting reply, Actioned, Marketing) to help maintain inbox zero and prioritize responses. What this workflow does Monitors Gmail for new incoming emails Uses AI to analyze email content and classify into appropriate categories Automatically applies the corresponding Gmail label Runs on a schedule to process emails consistently Setup Prerequisites n8n instance (cloud or self-hosted) Gmail account with API access enabled Access to an LLM provider (OpenAI, Anthropic Claude, or similar) Step-by-Step Configure Gmail Credentials Create Gmail Labels Configure LLM Chain Set Email Polling Schedule Test the Workflow Create Gmail Labels Before running the workflow, create these 8 labels in your Gmail account: To respond FYI Comment Notification Meeting update Awaiting reply Actioned Marketing How to customize this workflow to your needs Modify Classification Categories To change the email categories, update two places: In the AI prompt (Basic LLM Chain node): Your new category - Description of what emails fit here Another category - Description [... continue with your categories] In Gmail labels: Create corresponding labels in your Gmail account with the exact same names and numbering. Adjust Classification Rules The AI prompt contains specific rules for each category. To modify: Edit the "Key classification rules" section in the LLM prompt Add examples of emails that should go into each category Specify edge cases and how they should be handled Change Email Sources Currently monitors all incoming emails. To filter specific emails: In the Gmail Trigger node, add filters such as: from:specific-sender@domain.com subject:contains-keyword -label:already-processed You can also change this use Outlook Modify Polling Frequency More frequent**: Add multiple poll times (e.g., 9 AM, 12 PM, 6 PM) Less frequent**: Change to once daily or weekly Real-time**: Switch to webhook-based triggering (requires Gmail API setup) I choose daily for cost.
by Yaron Been
π° AI News Digest Agent: Auto News Summarizer & Email Newsletter Create an intelligent news curation system that automatically fetches breaking headlines, generates AI-powered summaries, and delivers personalized news digests to your subscriber list. Perfect for newsletter creators, team leaders, and content curators who want to keep their audience informed without the manual effort of news monitoring and summarization. π How It Works This streamlined 5-step automation delivers fresh news insights around the clock: Step 1: Automated News Collection The workflow runs on a configurable schedule (default: every 10 minutes) to fetch the latest headlines from NewsAPI, ensuring your content stays current with breaking developments. Step 2: Intelligent Content Curation The system pulls top headlines from reliable news sources, filtering by country, category, and relevance to deliver the most important stories of the day. Step 3: AI-Powered Summarization GPT-4 processes the collected headlines and creates: Concise 5-bullet point summaries Key insights and implications Easy-to-digest news overviews Professional formatting for email distribution Step 4: Subscriber Management The workflow accesses your Google Sheets subscriber list, retrieving names and email addresses for personalized delivery. Step 5: Automated Email Distribution Personalized news digests are automatically sent to each subscriber via Gmail, with custom greetings and professionally formatted content. βοΈ Setup Steps Prerequisites NewsAPI account (free tier available) OpenAI API access for content summarization Google Sheets for subscriber management Gmail account for email distribution n8n instance (cloud or self-hosted) Required Google Sheets Structure Create a simple subscriber database: | Name | Email | |---------------|--------------------------| | John Smith | john@example.com | | Sarah Johnson | sarah@company.com | | Mike Chen | mike.chen@startup.co | Configuration Steps Credential Setup NewsAPI Key: Sign up at newsapi.org for free headline access OpenAI API Key: Required for AI-powered news summarization Google Sheets OAuth2: Access your subscriber spreadsheet Gmail OAuth2: Enable automated email sending News Source Configuration Country Selection: Choose target region (US, UK, CA, AU, etc.) Category Filters: Focus on specific topics (technology, business, health) Source Selection: Prefer certain news outlets or avoid others Language Settings: Configure for international audiences AI Summarization Customization Default prompt creates 5-bullet summaries, but can be tailored for: Industry Focus: Technology, finance, healthcare, politics Audience Type: General public, professionals, executives Content Depth: Brief overviews vs detailed analysis Tone & Style: Formal, conversational, or technical Email Template Personalization Subject Line Formatting: Include date, breaking news indicators Greeting Customization: Use subscriber names for personal touch Content Layout: Professional formatting with clear sections Branding Elements: Add your organization's signature or logo Delivery Schedule Optimization Frequency Settings: Every 10 minutes, hourly, or daily Time Zone Considerations: Optimize for subscriber locations Breaking News Alerts: Immediate delivery for urgent stories Digest Compilation: Collect multiple stories for periodic summaries π Use Cases Newsletter Publishers Content Automation: Generate newsletter content without manual curation Consistent Publishing: Maintain regular delivery schedules automatically Audience Growth: Provide value that encourages subscriptions and shares Time Savings: Eliminate hours of daily news monitoring and writing Corporate Communications Employee Updates: Keep teams informed about industry developments Executive Briefings: Deliver curated news summaries to leadership Client Communications: Share relevant industry insights with customers Stakeholder Relations: Maintain informed investor and partner networks Educational Institutions Student Resources: Provide current events for academic discussions Faculty Updates: Keep educators informed about relevant developments Research Support: Deliver news related to specific academic fields Parent Communications: Share educational policy and school-related news Professional Services Client Value Addition: Provide industry-specific news as a service benefit Thought Leadership: Position your firm as an informed industry expert Business Development: Share insights that demonstrate market knowledge Team Knowledge Sharing: Keep entire organization current on industry trends Community Organizations Member Engagement: Keep community members informed and engaged Local News Focus: Customize for regional or local news coverage Event Planning: Stay informed about developments affecting your community Advocacy Support: Monitor news relevant to your organization's mission π§ Advanced Customization Options Multi-Source News Aggregation Expand beyond NewsAPI with additional sources: RSS Feed Integration: Add specialized industry publications Social Media Monitoring: Include trending topics from Twitter/LinkedIn Government Sources: Official announcements and policy updates International Coverage: Global perspectives on major stories Intelligent Content Filtering Implement smart curation features: Sentiment Analysis: Filter positive, negative, or neutral news Relevance Scoring: Prioritize stories based on subscriber interests Duplicate Detection: Avoid sending repetitive story coverage Quality Assessment: Ensure content meets editorial standards Subscriber Segmentation Create targeted news experiences: Interest Categories: Technology, business, sports, entertainment Geographic Preferences: Local, national, or international focus Delivery Preferences: Frequency and format customization Engagement Tracking: Monitor opens, clicks, and subscriber behavior Enhanced Email Features Professional newsletter capabilities: HTML Templates: Rich formatting with images and links Call-to-Action Buttons: Drive engagement with your content or services Social Sharing: Enable easy sharing of newsletter content Analytics Integration: Track email performance and subscriber engagement π Content Generation Examples Sample Email Output: Subject: π° Your Daily News Digest - March 15, 2024 Hi John, Please find today's top news headlines summarized below: π BUSINESS & TECHNOLOGY Federal Reserve signals potential rate cuts following inflation data Major tech companies announce AI partnership for healthcare applications Renewable energy sector sees record investment levels in Q1 2024 Cryptocurrency markets stabilize after regulatory clarity announcement Supply chain disruptions ease as global shipping routes normalize π‘ These developments suggest growing economic optimism and continued technology sector innovation. The healthcare AI partnership particularly signals significant advances in medical technology accessibility. Stay informed and have a great day! Powered by AI News Digest Agent Unsubscribe | Update Preferences Breaking News Alert Format: Subject: π¨ Breaking News Alert - Major Development Hi Sarah, BREAKING: [Headline] Key Details: [Critical point 1] [Critical point 2] [Impact analysis] Full coverage in your next scheduled digest. AI News Digest Agent π οΈ Troubleshooting & Best Practices Common Issues & Solutions API Rate Limiting Monitor NewsAPI quota usage and upgrade plan if needed Implement intelligent caching to reduce redundant requests Stagger requests during high-traffic periods Set up alerts for approaching rate limits Email Delivery Challenges Monitor Gmail sending limits and implement delays if needed Use professional email authentication (SPF, DKIM) Maintain clean subscriber lists to avoid spam flags Implement unsubscribe functionality for compliance Content Quality Control Review AI summaries periodically for accuracy and bias Implement feedback loops for continuous prompt improvement Create editorial guidelines for consistent tone and style Monitor subscriber feedback and engagement metrics Optimization Strategies Performance Enhancement Use parallel processing for multiple news sources Implement intelligent caching for repeated content Optimize AI prompts for faster processing and better results Monitor workflow execution time and resource usage Subscriber Growth Create compelling value propositions for newsletter signups Implement referral systems for organic growth Share sample newsletters on social media and websites Collect feedback to continuously improve content quality Content Strategy A/B test different summary formats and lengths Analyze which news categories generate most engagement Experiment with sending times for optimal open rates Create themed newsletters for special events or topics π Success Metrics Engagement Indicators Open Rates: Percentage of subscribers reading newsletters Click-Through Rates: Engagement with linked news sources Subscriber Growth: New signups and retention rates Forward/Share Rates: Viral coefficient of your content Content Quality Measurements Relevance Scores: Subscriber feedback on content usefulness Timeliness: How quickly breaking news reaches subscribers Accuracy: Verification of AI-summarized content Completeness: Coverage of important stories in your focus areas π Questions & Support Need assistance with your AI News Digest Agent setup or optimization? π§ Technical Support Email: Yaron@nofluff.online Response Time: Within 24 hours on business days Specialization: NewsAPI integration, AI content optimization, email deliverability π₯ Educational Resources YouTube Channel: https://www.youtube.com/@YaronBeen/videos Complete setup and configuration tutorials Advanced customization techniques for different industries Email marketing best practices for automated newsletters Troubleshooting common integration issues Scaling strategies for growing subscriber lists π€ Professional Community LinkedIn: https://www.linkedin.com/in/yaronbeen/ Connect for ongoing newsletter automation support Share your news curation success stories Access exclusive templates and workflow variations Join discussions about content automation trends π¬ Support Request Best Practices Include in your support message: Your target audience and newsletter focus Current subscriber count and growth goals Specific news categories or geographic regions of interest Any technical errors or integration challenges Current content creation workflow and pain points