by ömerDrn
Automated Cryptocurrency Analysis & Reporting with Google Gemini and CoinGecko This powerful template is an n8n workflow that automates cryptocurrency market data analysis and delivers reports directly to your inbox. It fetches real-time data from CoinGecko API, generates AI-powered analysis, and sends the report via email. Features Scheduled Execution**: Runs automatically at a set time daily (default: 10:00 AM). Customizable Analysis**: Personalize analysis content/language via "AI Prompt" nodes. Easy Scalability**: Duplicate node groups to add more cryptocurrencies. Flexible AI Integration**: Defaults to Google Gemini, but supports ChatGPT/Ollama. Setup Instructions n8n Installation: Install n8n (self-hosted or Cloud version). Email Account Setup: Add email service credentials in n8n (e.g., Microsoft Outlook OAuth2). AI Model Credentials (Google Gemini): Obtain API key from Google AI Studio and add to n8n "Credentials". Import Template: Copy the JSON code into n8n as a new workflow. Customization Change Cryptocurrencies**: Update ids= parameter in HTTP Request nodes (e.g., ids=bitcoin). Edit AI Prompt**: Modify text in "AI Prompt" nodes. Use Different AI Model**: Replace Google Gemini with supported alternatives. Update Email Address**: Change recipient in "Send Mail" nodes. `
by gotoHuman
Generate AI video clips to promote products, services or events on social media. Use gotoHuman as an interface to control and supervise each step of the workflow to create content that's actually worth posting. How it works gotoHuman will show the workflow steps that need approval or input in its' inbox and notify you via email or Slack. We choose from different topics for our post suggested by AI We select the image style, a product to show, and review an AI-generated tag line We use Fal.ai to generate an image that serves as a reference image for our video clip. And we use Cloudinary to add an overlay for the tag line. We review the image in gotoHuman and can iterate on it by retrying or even changing the prompt. We review the video clip that's generated with Fal.ai based on the approved image and can, again, retry or reprompt. How to set up Most importantly, install the gotoHuman node before importing this template! (Just add the node to a blank canvas before importing) Follow the instructions shown along the workflow and in the incl. video guide. You mainly need to set up your credentials for gotoHuman, OpenAI, Fal.ai and Cloudinary import the review templates with these IDs in gotoHuman: Z7V1jyImY1pho9eY039R,0GBaOCWd27tqV562kkCL,E2wlCVPWmk2UnLHVt4uu,DitPdbIapS4rBxBTIYGt,Z2T7nFwkXVFQlD6z50eV select these templates in the gotoHuman nodes do a quick setup for Cloudinary Requirements You need accounts for gotoHuman (human supervision) OpenAI (ideation) Fal.ai (image/video generation) Cloudinary (text overlay) How to customize Adjust/Replace the workflow triggers as needed Change the prompt in the topics generation node Replace the product image URLs used in the "gotoHuman - Content" node Adjust the available styles for image generation in the gotoHuman review template and the prompts they link to in the "Set Initial Image Prompt" node Adjust the prompt used for video generation in the "Set Initial Video Prompt" node If you want to use a different service/model for image and video generation, replace the nodes related to Fal.ai. Also, if you do not need a text overlay, remove the Cloudinary nodes.
by Intuz
Disclaimer: Community nodes are used, and template can only be used on self-hosted n8n instances. This n8n template from Intuz provides a complete solution to automate your entire B2B lead generation pipeline, from discovering recently funded companies to drafting hyper-personalized outreach emails with AI. Who's this workflow for? Sales Development Representatives (SDRs) Business Development Teams Growth Hackers Startup Founders Marketing Agencies How it works 1. Scrape Funded Companies: The workflow begins by using Apify to scrape a target list of recently funded companies directly from a Crunchbase search. 2. Enrich with Apollo.io: It takes each company and uses the Apollo.io API to find key decision-makers (like VPs, Directors) and enrich their contact information, including finding their email addresses. 3. Populate Google Sheets: All the gathered lead data—company name, contact name, title, email, LinkedIn URL, etc.—is neatly organized and added to a Google Sheet. 4. AI-Personalized Email Crafting: The workflow sends the lead's information to OpenAI (GPT-4) with a highly specialized prompt, instructing it to write a concise, impactful, and hyper-personalized "first touch" cold email. 5. Update Lead List with Email Content: Finally, the unique, AI-generated email is saved back into the Google Sheet alongside the corresponding lead's information, making it ready for you to send. Pre-conditions and Requirements Before you can successfully execute this workflow, you must have the following accounts, credentials, and assets in place. 1. n8n Instance: You need an active n8n instance (self-hosted). 2. Apify Account & Crunchbase Access: Apify Account: A registered account on Apify. Crunchbase Account: An active, logged-in Crunchbase account (a paid subscription is recommended for accessing detailed search filters). 3. Apollo.io API: You need an Apollo.io plan that includes API access. You can generate the API from settings. 4. Google Sheet: Create a new Google Sheet to store your leads. The workflow is configured for two tabs: one for raw data ("HealthCare" in the template) and one for email generation ("Company sheet"). 5. OpenAI Account: An account with OpenAI with API access and billing set up. Setup Instructions 1. Apify Connection: Connect your Apify account in the Run an Actor node. You'll need an apify scrapper, here's the link In the Custom Body field, update the search.url with your target Crunchbase discovery URL and provide a valid cookie for authentication. 2. Apollo.io Connection: Connect your Apollo.io account using HTTP Header Authentication in the three Apollo nodes. You will need to provide your API key. 3. Google Sheets Connection: Connect your Google Sheets account. Create a spreadsheet and update the Document ID and Sheet Name in the three Google Sheets nodes to match yours. Ensure your sheet columns are set up to receive the data. 4. OpenAI Connection: Connect your OpenAI account in the Message a model node. The prompt is pre-engineered for high-quality output, but you can tailor it to better fit your specific value proposition. 5. Activate Workflow: Click "Execute workflow" to run the automation manually and watch your AI-powered lead list build itself. Customization Guide This workflow is a powerful template. To adapt it to your specific business needs, you should review and modify the following nodes. 1. Changing Your Target Companies (The Source) Node: Run an Actor What to change: The search.url parameter inside the customBody. How to do it: Go to Crunchbase and perform a search for your ideal companies (e.g., filter by different funding rounds, industry, location, keywords, etc.). Copy the URL from your browser's address bar after the search results have loaded. Paste this new URL as the value for "search.url" in the node. You can also adjust "count": 10 to pull more or fewer companies per run. Be mindful of Apify and Apollo credit usage. 2. Defining Your Ideal Contact Persona Node: Apollo - Get User What to change: The person_seniorities and person_titles arrays in the jsonBody. How to do it: 1. Seniority: Modify the person_seniorities list to match who you sell to. Examples: ["c_level", "founder"] or ["manager", "contributor"]. 2. Job Titles: This is crucial. Replace the existing list of titles ("engineering", "technology", etc.) with keywords relevant to your target buyer. For example, if you sell to marketing teams, you might use: ["marketing", "demand generation", "growth", "content", "brand"]. 3. Configuring Your Google Sheet Destination Nodes: Append or update row in sheet and Update row in sheet What to change: The documentId and sheetName. How to do it: Open your Google Sheet. The documentId is the long string of characters in the URL between /d/ and /edit. Copy and paste it into the "Document ID" field in both nodes. The sheetName (or Sheet ID/gid) needs to be set for your specific tabs. Make sure the sheet names/IDs in the nodes match the tabs in your document. Column Mapping: If you change the column names in your Google Sheet, you must update the column mapping inside these nodes to ensure the data is written to the correct place. 4. Tailoring the AI Email Generation Node: Message a model (OpenAI) What to change: The prompt, the model, and the input variables. How to do it: The Prompt: This is the heart of your outreach. Read the entire prompt carefully and edit it to reflect your company's value proposition, tone of voice, and specific call-to-action. Value Proposition: Change the line "We help them cut that specific infrastructure spend..." to match what your product does. Use a powerful, single data point if you have one. Call-to-Action (CTA): Modify the final question ("Curious if infra efficiency is on your roadmap...") to something that fits your sales process. Tone: Adjust the initial instructions (e.g., "Your tone is that of a peer...") if you want a different style. The Model: The workflow uses gpt-4.1. You can switch to a different model like gpt-4o (potentially better/faster) or gpt-3.5-turbo (much cheaper, but lower quality) depending on your budget and needs. Input Variables: The prompt uses {{ $json['Company Name'] }}, {{ $json['Person Designation'] }}, and {{ $json.Industry }}. If you want to add more personalization (e.g., based on a company's funding amount), you would first need to ensure that data is passed to this node, then add the new variable (e.g., {{ $json['Funding Amount'] }}) into the prompt. Connect with us Website: https://www.intuz.com/services Email: getstarted@intuz.com LinkedIn: https://www.linkedin.com/company/intuz Get Started: https://n8n.partnerlinks.io/intuz For Custom Workflow Automation Click here- Get Started
by Don Jayamaha Jr
Instantly fetch real-time Bitget spot market data directly in Telegram! This workflow integrates the Bitget REST v2 API with Telegram (plus optional AI-powered formatting) to deliver the latest crypto price, order book, candles, and recent trades. Perfect for crypto traders, analysts, and investors who need reliable market data at their fingertips—no API key required.  Sign-up for Bitget for 6,200 USDT in rewards to trade: Collect Now How It Works A Telegram bot listens for user requests (e.g., BTCUSDT). The workflow connects to Bitget public endpoints to fetch: Ticker (latest price & 24h stats) Order book depth (top bids/asks) Recent trades (price, side, volume, timestamp) Candlestick data (1m, 15m, 1h, 4h, 1d) Historical candles (optional, for backfill before endTime) A Calculator node derives useful metrics like mid-price and spread. A Think node reshapes raw JSON into Telegram-ready text. A splitter ensures reports over 4000 characters are chunked safely. The final market insights are delivered instantly back to Telegram. What You Can Do with This Agent ✅ Track live prices & 24h stats for any Bitget spot pair. ✅ Monitor order book liquidity and spreads in real-time. ✅ Analyze candlesticks across multiple timeframes. ✅ Review recent trades to see execution flow. ✅ Fetch historical candles for extended market context. ✅ Receive clean, structured reports with optional AI-enhanced formatting. Set Up Steps Create a Telegram Bot Use @BotFather to generate a bot token. Configure in n8n Import Bitget AI Agent v1.02.json into your n8n instance. Add your Telegram credentials (bot token + your Telegram ID in the User Authentication node). Add an OpenAI key if you want AI-powered formatting. (Optional) Add an *Bitget api key** . Deploy and Test Send BTCUSDT to your bot. Get live Bitget spot data instantly in Telegram! 🚀 Unlock powerful, real-time Bitget insights in Telegram—zero setup, zero API keys required! 📺 Setup Video Tutorial Watch the full setup guide on YouTube: 🧾 Licensing & Attribution © 2025 Treasurium Capital Limited Company Architecture, prompts, and trade report structure are IP-protected. No unauthorized rebranding permitted. 🔗 For support: Don Jayamaha – LinkedIn
by Robert Breen
Chat to write or reword a blog post. The workflow stores each result in Google Sheets and uses a sub-workflow “Google tool” to count rows per session (your running context). If a session exceeds a row threshold, the flow can branch (e.g., stop or notify). ⚙️ Setup Instructions 1️⃣ Set Up OpenAI Connection Go to OpenAI Platform Navigate to OpenAI Billing Add funds to your billing account Copy your API key into the OpenAI credentials in n8n 2️⃣ Prepare Your Google Sheet Connect your Data in Google Sheets Use this format: Sample Sheet Row 1 = column names (e.g., session, Rows, output) Data in rows 2–100 (or more if you prefer) In n8n, use Google Sheets OAuth2 → pick your Spreadsheet and Worksheet (Optional) You can adapt this to Airtable, Notion, or a Database 🧠 How It Works Chat Trigger**: Provide a topic (write) or paste existing text (reword). Code Node (“Choose to Write or Edit Blog”)**: Builds a system_prompt + user_prompt Instructs the agent to call the Google tool (sub-workflow) with only the sessionid to count existing rows. Tool Workflow (“google”)**: Fetches rows from the sheet → filters by session → summarizes row count. Agent (“Blog Writer & Editor”)**: Returns structured JSON (items/rows, session, blog body). Store (Google Sheets)**: Appends { session, Rows, output } to the sheet. If Node**: Example rule: Rows > 3 → branch/limit/notify as needed. 💬 Example Prompts “Write a 600-word blog about n8n agents with 3 bullet takeaways. Session: abc123.” “Reword this post into a concise LinkedIn article. Session: launchQ3:\n<your text here>” “Draft a blog intro and 5 SEO headlines on marketing automation. Session: mkt-01.” 📬 Contact Need help tailoring this to Airtable/Notion/DB, or adding auto-publishing? 📧 rbreen@ynteractive.com 🔗 Robert Breen 🌐 ynteractive.com
by JJ Tham
Generate AI Voiceovers from Scripts and Upload to Google Drive This is the final piece of the AI content factory. This workflow takes your text-based video scripts and automatically generates high-quality audio voiceovers for each one, turning your text into ready-to-use audio assets for your video ads. Go from a spreadsheet of text to a folder of audio files, completely on autopilot. ⚠️ CRITICAL REQUIREMENTS (Read First!) This is an advanced, self-hosted workflow that requires specific local setup: Self-Hosted n8n Only:** This workflow uses the Execute Command and Read/Write Files nodes, which requires you to run your own instance of n8n. It will not work on n8n Cloud. FFmpeg Installation:** You must have FFmpeg installed on the same machine where your n8n instance is running. This is used to convert the audio files to a standard format. What it does This is Part 3 of the AI marketing series. It connects to the Google Sheet where you generated your video scripts (in Part 2). For each script that hasn't been processed, it: Uses the Google Gemini Text-to-Speech (TTS) API to generate a voiceover. Saves the audio file to your local computer. Uses FFmpeg to convert the raw audio into a standard .wav file. Uploads the final .wav file to your Google Drive. Updates the original Google Sheet with a link to the audio file in Drive and marks the script as complete. How to set up IMPORTANT: This workflow is Part 3 of a series and requires the output from Part 2 ("Generate AI Video Ad Scripts"). If you need Part 1 or Part 2 of this workflow series, you can find them for free on my n8n Creator Profile. Connect to Your Scripts Sheet: In the "Getting Video Scripts" node, connect your Google Sheets account and provide the URL to the sheet containing your generated video scripts from Part 2. Configure AI Voice Generation (HTTP Request): In the "HTTP Request To Generate Voice" node, go to the Query Parameters and replace INSERT YOUR API KEY HERE with your Google Gemini API key. In the JSON Body, you can customize the voice prompt (e.g., change <INSERT YOUR DESIRED ACCENT HERE>). Set Your Local File Path: In the first "Read/Write Files from Disk" node, update the File Name field to a valid directory on your local machine where n8n has permission to write files. Replace /Users/INSERT_YOUR_LOCAL_STORAGE_HERE/. Connect Google Drive: In the "Uploading Wav File" node, connect your Google Drive account and choose the folder where your audio files will be saved. Update Your Tracking Sheet: In the final "Uploading Google Drive Link..." node, ensure it's connected to the same Google Sheet from Step 1. This node will update your sheet with the results. Name and Description for Submission Form Here are the name and description, updated with the new information, ready for you to copy and paste. Name: Generate AI Voiceovers from Scripts and Upload to Google Drive Description: Welcome to the final piece of the AI content factory! 🔊 This advanced workflow takes the video ad scripts you've generated and automatically creates high-quality audio voiceovers for each one, completing your journey from strategy to ready-to-use media assets. ⚠️ This is an advanced workflow for self-hosted n8n instances only and requires FFmpeg to be installed locally. ⚙️ How it works This workflow is Part 3 of a series. It reads your video scripts from a Google Sheet, then for each script it: Generates a voiceover using the Google Gemini TTS API. Saves the audio file to your local machine. Converts the file to a standard .wav format using FFmpeg. Uploads the final audio file to Google Drive. Updates your Google Sheet with a link to the new audio file. 👥 Who’s it for? Video Creators & Marketers: Mass-produce voiceovers for video ads, tutorials, or social media content without hiring voice actors. Automation Power Users: A powerful example of how n8n can bridge cloud APIs with local machine commands. Agencies: Drastically speed up the production of audio assets for client campaigns. 🛠️ How to set up This workflow requires specific local setup due to its advanced nature. IMPORTANT: This is Part 3 of a series. To find Part 1 ("Generate a Strategic Plan") and Part 2 ("Generate Video Scripts"), please visit my n8n Creator Profile where they are available for free. Setup involves connecting to your scripts sheet, configuring the AI voice API, setting a local file path for n8n to write to, and connecting your Google Drive.
by Ehsan
Analyze food ingredients from Telegram photos using Gemini and Airtable 🛡️ Personal Ingredient Bodyguard Turn your Telegram bot into an intelligent food safety scanner. This workflow analyzes photos of ingredient labels sent via Telegram, extracts the text using AI, and cross-references it against your personal database of "Good" and "Bad" ingredients in Airtable. It solves the problem of manually reading tiny, complex labels for allergies or dietary restrictions. Whether you are Vegan, Halal, allergic to nuts, or just avoiding specific additives, this workflow acts as a strict, personalized bodyguard for your diet. It even features a customizable "Persona" (like a Sarcastic Bodyguard) to make safety checks fun. 🎯 Who is it for? People with specific dietary restrictions (Vegan, Gluten-free, Keto). Individuals with food allergies (Nuts, Dairy, Shellfish). Special dietary observers (Halal, Kosher). Health-conscious shoppers avoiding specific additives (e.g., E120, Aspartame). 🚀 How it works Trigger: You send a photo of a product label to your Telegram Bot. Fetch Rules: The workflow retrieves your active "Watchlist" (Ingredients to avoid/prefer) and "Persona" settings from Airtable. Vision & Logic: It uses an AI Vision model to extract text from the image (OCR) and Google Gemini to analyze the text against your strict veto rules (e.g., "Safe" only if ZERO bad items are found). Response: The bot replies instantly on Telegram with a Safe/Unsafe verdict, highlighting detected ingredients using HTML formatting. Log: The result is saved back to Airtable for your records. ⚙️ How to set up This workflow relies on a specific Airtable structure to function as the "Brain." Set up Airtable Sign up for Airtable: Click here Copy the required Base: Click here to copy the "Ingredients Brain" base Connect Airtable to n8n (5-min guide): Watch Tutorial Set up Telegram Message @BotFather on Telegram to create a new bot and get your API Token. Add your Telegram credentials in n8n. Configure AI Add your Google Gemini API credentials. Note on OCR: This template is configured to use a local LLM for OCR to save costs (via the OpenAI-compatible node). If you do not have a local model running, simply swap the "OpenAI Chat Model" node for a standard GPT-4o or Gemini Vision node. 📋 Requirements n8n** (Cloud or Self-hosted) Airtable** account (Free tier works) Telegram** account Google Gemini** API Key Local LLM* (Optional, for free OCR) OR *OpenAI/Gemini** Key (for standard Cloud Vision) 🎨 How to customize Change the Persona:** Go to the "Preferences" table in Airtable to change the bot's personality (e.g., "Helpful Nutritionist") and output language. Update Ingredients:** Add or remove items in the "Watchlist" table. Mark them as "Good Stuff" or "Bad Stuff" and set Status to "Active". Adjust Sensitivity:** The AI prompt in the "AI Agent" node is set to strict "Veto" mode (Bad overrides Good). You can modify the system prompt to change this logic. ⚠️ Disclaimer This tool is for informational purposes only. Not Medical Advice: Do not rely on this for life-threatening allergies. AI Limitations: OCR can misread text, and AI can hallucinate. Verify: Always double-check the physical product label. Use at your own risk.
by Simeon Penev
Who’s it for Content/SEO teams who want a fast, consistent, research-driven brief for a copywriters from a single keyword—without manual review and analysis of the SERP (Google results). How it works / What it does Form Trigger collects the keyword/topic and redirects to Google Drive Folder after the final node. FireCrawl Search & Scrape pulls the top 5 pages for the chosen keyword. AI Agent (with Think + OpenAI Chat Model) analyzes sources and generates an original Markdown brief. Markdown to JSON converts the Markdown into Google Docs batchUpdate requests (H1/H2/H3, lists, links, spacing). Then this is used in Update a document for updating the empty doc. Create a document + Update a document write a Google Doc titled “SEO Brief for ” and update the Google Doc in your target Drive folder. How to set up Add credentials: Firecrawl (Authorization header), OpenAI (Chat), Google Docs OAuth2. Replace placeholders: {{APIKEY}}, {{googledrivefolderid}}, {{googledrivefolderurl}}. Publish and open the Form URL to test. Requirements Firecrawl API key • OpenAI API key • Google account with access to the target Drive folder. Resources Google OAuth2 Credentials Setup - https://docs.n8n.io/integrations/builtin/credentials/google/oauth-generic/ Firecrawl API key - https://take.ms/lGcUp OpenAI API key - https://docs.n8n.io/integrations/builtin/credentials/openai/
by David Roberts
This workflow allows you to ask questions about a PDF document. The answers are provided by an AI model of your choice, and the answer includes a citation pointing to the information it used. You can use n8n’s built-in chat interface to ask the questions, or you could customise this workflow to use another one (e.g. Slack, Teams, etc.) Example The workflow is set up with the Bitcoin whitepaper. So you could ask things like: Question: “Which email provider does the creator of Bitcoin use?“ Answer: “GMX [Bitcoin whitepaper.pdf, lines 1-35]” Requirements A Pinecone account (they have a free tier at the time of writing that is easily enough for this workflow) Access to a large language model (e.g. an OpenAI account) Customizing this workflow The workflow only reads in one document, but you could customise it to read in all the documents in a folder (or more). The workflow is set up to use GPT 3.5, but you could swap that out for any other model (including self-hosted ones).
by Yaron Been
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. This workflow automatically analyzes customer lifetime value (CLV) metrics to optimize customer acquisition and retention strategies. It saves you time by eliminating the need to manually calculate CLV and provides data-driven insights for maximizing customer profitability and improving business growth. Overview This workflow automatically scrapes customer data, purchase history, and engagement metrics to calculate and analyze customer lifetime value patterns. It uses Bright Data to access customer analytics platforms and AI to intelligently segment customers, predict CLV, and identify high-value customer characteristics. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping customer analytics and CRM platforms without being blocked OpenAI**: AI agent for intelligent CLV analysis and customer segmentation Google Sheets**: For storing CLV calculations and customer analysis data How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your CLV analysis spreadsheet Customize: Define customer data sources and CLV calculation parameters Use Cases Customer Success**: Focus retention efforts on high-value customers Marketing Strategy**: Optimize customer acquisition costs based on projected CLV Sales Teams**: Prioritize prospects with higher lifetime value potential Business Strategy**: Make data-driven decisions about customer investments Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #customerlifetimevalue #clv #customeranalytics #brightdata #webscraping #customerdata #n8nworkflow #workflow #nocode #customersegmentation #valueanalysis #customerinsights #revenueoptimization #customervalue #clvanalysis #customermetrics #customerprofitability #businessintelligence #customerretention #valueprediction #customeroptimization #revenueanalysis #customerstrategy #lifetimevalue #customerroi #valuedriven #customerworth #profitability
by Paul-François GORIAUX
This workflow acts as your personal AI-powered analyst for Meta Ads. It's pretty straightforward: First, it grabs a list of Facebook Ad Library URLs you want to check out from a Google Sheet. Then, it automatically scrapes the active ads from those pages. Here's the cool part: it sends each ad's image and text to Google Gemini, which analyzes it like an expert marketer would. Finally, Gemini's full analysis—we're talking strengths, weaknesses, actionable suggestions, and a performance score—gets dropped neatly into another Google Sheet for you. Set up steps You should be ready to roll in about 5 minutes. There are no complex configurations, you just need to: Connect your accounts: The workflow has placeholders waiting for your credentials for Google (for Sheets and the Gemini API) and ScrapingFlash. Link your Google Sheets: Just point the first Google Sheets node to the sheet with your URLs, and tell the last node where you want to save the results. All the nitty-gritty details and expressions are explained in the sticky notes inside the workflow itself!
by Ari Nakos
This n8n workflow automates lead generation by searching Reddit for relevant posts based on keywords, filtering them, using OpenRouter AI to analyze and summarize content, and logging the findings (link, summary, etc.) to Google Sheets. Watch the full setup tutorial on how I setup this ETL pipeline using n8n: https://youtu.be/F3-fbU3UmYQ Required Authentication: To run this workflow, you need to set up credentials in n8n for: Reddit: Uses OAuth 2.0. Requires creating an app on Reddit to get a Client ID & Secret. (YT Tutorial for Reddit App Creation: https://youtu.be/zlGXtW4LAK8) OpenRouter: Uses an API Key. Generate this key directly from your OpenRouter account settings. (YT Tutorial : https://youtu.be/Cq5Y3zpEhlc) Google Sheets: Uses OAuth 2.0. Requires setup in Google Cloud Console (enable Sheets API, create OAuth Client ID with n8n redirect URI) to get a Client ID & Secret. Ensure these credentials are created and selected in the respective n8n nodes (Get Posts, OpenRouter Chat Model nodes, Output The Results).