by Jay Hartley
What this workflow does Downloads the daily top podcasts of a selected genre Summarizes the content of each podcast in a few paragraphs Sends the summaries and the direct link to each podcast in a formatted email Setup Create a free API key on Taddy here: https://taddy.org/signup/developers Input your user number and API key into the TaddyTopDaily node in the header parameters X-USER-ID and X-API-KEY respectively. Create access credentials for your Gmail as described here: https://developers.google.com/workspace/guides/create-credentials. Use the credentials from your client_secret.json in the Gmail node. In the Genre node, set the genre of podcasts you want a summary for. Valid values are: TECHNOLOGY, NEWS, ARTS, COMEDY, SPORTS, FICTION, etc. Look at api.taddy.org for the full list (they will be displayed in the help docs as PODCASTSERIES_TECHNOLOGY, PODCASTSERIES_NEWS, etc.) Enter your email address in the Gmail node. Change the schedule time for sending email from Schedule to whichever time you want to receive the email. Test: Hit Test Workflow. Check your email for the results. That's it! It should take less than 5 minutes total.
by Polina Medvedieva
This n8n workflow template lets you easily generate comprehensive FAQ (Frequently Asked Questions) content for multiple services (or any items or pages you need to add the FAQs to). Simply provide the Google Sheets document containing the items to scrape, and the workflow automatically creates detailed, AI-enhanced FAQ documents. How it works The workflow reads data from a Google Sheets document containing information about different services and categories (again, in your case - whatever objects you need). For each service and category, it generates a set of standard questions and answers covering setup, permissions, integrations, use cases, and pricing benefits. An AI model (OpenAI's GPT) is used to enhance or complete some of the answers, making the content more comprehensive and natural-sounding. The workflow formats the Q&A pairs, combining AI-generated content with predefined answers where applicable. It creates a text file (JSON) for each service or category, containing the formatted Q&A pairs. The generated files are saved to specific folders in Google Drive, organized by the type of integration (native, credential-only, non-native) or category. After processing each service or category, it updates the status in the original Google Sheets document to mark it as completed. Ideal for: Marketing teams: Rapidly create comprehensive FAQ documents for multiple products or services. Customer support: Generate consistent and detailed answers for common customer queries. Product managers: Easily maintain up-to-date documentation as products evolve. Content creators: Streamline the process of creating informative content about various offerings. Accounts required Google account (for Google Sheets and Google Drive) OpenAI API account (for AI-enhanced content generation) n8n.io account (for workflow execution) Set up instructions Set up the required credentials for Google Sheets, Google Drive, and OpenAI when you first open the workflow. Prepare your Google Sheets document with the service/category information. Here's an example of Google Sheet. Fill the "Define Sheets" node with your sheets Adjust the folder IDs in the "Prepare Job" node to match your Google Drive structure. Configure the OpenAI model settings in the "OpenAI Chat Model" node if needed. Test the workflow with a small subset of data before running it on your entire dataset. Adjust the questions asked in the "Create your Q&A templates" section After testing, activate your workflow for automated FAQ generation. 🙏 Big, big kudos to Jim Le for his ideas, input and support when building this workflow. Your approach to AI workflows is always super helpful!
by Keith Rumjahn
Case Study I'm too lazy to record every transaction for my expense tracking. Since all my expenses are digital, I just extract the transactions from bank PDF statements and screenshots into CSV to import into my budgeting software. Read more -> How I used A.I. to track all my expenses What this workflow does Upload your PDF or screenshots into Google Drive It then passes the PDF/image to Vertex Gemini to do some A.I. image recognition It then sends the transactions as CSV and stores it into another Google Drive folder Setup Set up 2 google drive folders. 1 for uploading and 1 for the output. Input your Google Drive crendtials Input your Vertex Gemini credentials How to adjust it to your needs You can upload other types of documents for information extraction. You can extract any text data from any image or PDF You can adjust the A.I. prompt to do different things
by John Pranay Kumar Reddy
✨ Summary Efficiently monitor Kubernetes environments by sending only unique error logs from Grafana Loki to Slack. Reduces alert fatigue while keeping your team informed about critical log events. 🧑💻 Who’s it for DevOps or SRE engineers running EKS/GKE/AKS Anyone using Grafana Loki and Promtail for centralized logging Teams that want Slack alerts but hate alert spam 🔍 What it does This n8n workflow queries your Loki logs every 5 minutes, filters only the critical ones (error, timeout, exception, etc.), removes duplicate alerts within the batch, and sends clean alerts to a Slack channel with full metadata (pod, namespace, node, container, log, timestamp). 🧠 How it works 🕒 Schedule Trigger Every 5 minutes (customizable) 🌐 Loki HTTP Query Pulls logs from the last 10 minutes Keyword match: error, failed, oom, etc. 🧹 Log Parsing Extracts log fields (pod, container, etc.) Skips empty/malformed results 🧠 Deduplication Removes repeated error messages (within query window) 📤 Slack Notification Sends nicely formatted message to Slack ⚙️ Requirements Tool Notes Loki- Exposed internally or externally Slack App- With chat:write OAuth n8n- Cloud or self-hosted 🔧 How to Set It Up Import the JSON file into n8n Update: Loki API URL (e.g., http://loki-gateway.monitoring.svc.cluster.local) Slack Bearer Token (via credentials) Target Slack channel (e.g., #k8s-alerts) (Optional) Change keywords in the query regex Activate the workflow Ensure n8n pod/container is having access to your kubernetes cluster/pods/namespaces 🛠 How to Customize Want more or fewer keywords? Adjust the regex in the Query Loki for Error Logs node. Need to increase deduplication logic? Enhance the Remove Duplicate Alerts node. Want 5-log summaries every 5 min? Fork this and add a Batch + Slack group sender. Grafana Loki logs to Slack Output
by Rodrigue Gbadou
How it works Regulatory monitoring**: Continuously tracks changes in laws, regulations, and compliance requirements across multiple jurisdictions Contract analysis**: AI-powered review of existing contracts to identify compliance gaps and risks Automated alerts**: Real-time notifications when regulatory changes affect your contracts or business operations Compliance reporting**: Generates audit-ready reports and documentation for regulatory compliance Set up steps Legal databases**: Connect to legal research platforms (Westlaw, LexisNexis, EUR-Lex) Contract repository**: Integrate with your contract management system or document storage Regulatory feeds**: Configure government and regulatory body RSS feeds and APIs AI legal analysis**: Set up OpenAI or specialized legal AI for contract analysis Compliance calendar**: Integrate with calendar systems for deadline tracking Audit trail**: Configure logging and documentation systems for compliance records Key Features 🔍 Multi-jurisdiction monitoring**: Tracks regulatory changes across different countries and regions 📊 Risk assessment**: Automatically scores compliance risks and potential impact ⚡ Real-time alerts**: Instant notifications when regulations affecting your business change 📋 Gap analysis**: Identifies discrepancies between current contracts and new requirements 🤖 AI-powered analysis**: Uses natural language processing to understand legal text 📈 Compliance dashboard**: Visual overview of compliance status across all contracts 🔄 Automated remediation**: Suggests contract amendments and compliance actions 📱 Mobile notifications**: Critical compliance alerts on mobile devices Compliance areas monitored Data protection**: GDPR, CCPA, and other privacy regulations Financial services**: Banking regulations, securities law, anti-money laundering Healthcare**: HIPAA, medical device regulations, pharmaceutical compliance Employment law**: Labor regulations, workplace safety, discrimination laws Environmental**: ESG requirements, environmental protection regulations Industry-specific**: Sector-specific regulations and standards Contract types supported Vendor agreements**: Supplier contracts and service agreements Employment contracts**: Employee agreements and contractor terms Data processing agreements**: Privacy and data handling contracts Customer agreements**: Terms of service and customer contracts Partnership agreements**: Joint ventures and strategic partnerships Licensing agreements**: Software licenses and intellectual property Automated responses Low risk (0-30)**: Routine monitoring and documentation Medium risk (31-60)**: Enhanced review and stakeholder notification High risk (61-80)**: Immediate legal review and action planning Critical risk (81-100)**: Emergency legal intervention and compliance measures Integration capabilities Legal research**: Westlaw, LexisNexis, Bloomberg Law Document management**: SharePoint, Google Drive, Dropbox Contract systems**: DocuSign, PandaDoc, ContractWorks Communication tools**: Slack, Teams, email for legal team alerts Calendar systems**: Outlook, Google Calendar for compliance deadlines This workflow ensures continuous legal compliance by monitoring regulatory changes and automatically assessing their impact on your contracts and business operations.
by Jorge Martínez
Automate tweet engagement on X (formerly Twitter) Description Automate professional engagement on X (formerly Twitter) by searching for, filtering, liking, and replying to tweets that match your key topics. This workflow enables you to engage consistently and efficiently with relevant conversations, using your defined professional role and the power of GPT for filtering and replies. Save time and maintain high-quality interactions, while staying focused on your business or personal brand interests. How it Works Rotating Topic Selection The workflow selects one search term from your list on each run, using a rotating index based on the date. Search Tweets & Extract Essentials Searches X (formerly Twitter) for tweets matching the chosen topic, then extracts only the tweet id and text for further processing. GPT‑Based Filtering with Role Context Filters tweets based on your role and strict criteria, removing non-English tweets, memes, spam, Grok-generated content, political posts, internships, and more. Engagement Loop For every filtered tweet, the workflow likes the post, generates a professional, concise reply with GPT (matching language and context), and posts the reply. Wait nodes ensure compliance with Twitter’s API rate limits (can be adjusted for paid API tiers). Requirements X (Twitter) API credentials (for searching, liking, and replying to tweets) OpenAI API key (for GPT-based steps) Setup Steps Obtain your X (Twitter) API credentials. Obtain your OpenAI API key. Configure the schedule in the trigger node to your desired frequency (e.g., every 3 days or daily). Set your list of topics and professional role in the variables node. How to Customize the Workflow (Optional) Adjust prompts** in the GPT nodes to fine-tune filtering and reply style. Upgrade your Twitter API plan** to increase request limits and search for more tweets per run. Change tweet processing logic:** For high-volume engagement (e.g., analyzing 100+ tweets per run), consider switching to a per-tweet loop for advanced filtering and response handling. This workflow enables scalable, professional, and targeted engagement on X (formerly Twitter), fully customizable to your audience and objectives.
by Tony Paul
How it works ++Download the google sheet here++ and replace this with the googles sheet node: Google sheet , upload to google sheets and replace in the google sheets node. Scheduled trigger: Runs once a day at 8 AM (server time). Fetch product list: Reads your “master” sheet (product_url + last known price) from Google Sheets. Loop with delay: Iterates over each row (product) one at a time, inserting a short pause (20 s) between HTTP requests to avoid blocking. Scrape current price: Loads each product_url, extracts the current price via a simple CSS selector. Compare & normalize: Compares the newly scraped price against the “last_price” from your sheet, calculates percentage change, and tags items where price_changed == true. On price change: Send alert: Formats a Telegram message (“Price Drop” or “Price Hike”) and pushes it to your configured chat. Log history: Appends a new row to a separate “price_tracking” tab with timestamp, old price, new price, and % change. Update master sheet: After a 1 min pause, writes the updated current_price back to your “master” sheet so future runs use it as the new baseline. Set up step Google Sheets credentials (~5 min) Create a Google Sheets OAuth credential in n8n. Copy your sheet’s ID and ensure you have two tabs: product_data (columns: product_url, price) price_tracking (columns: timestamp, product_url, last_price, current_price, price_diff_pct, price_changed) Paste the sheet ID into both Google Sheets nodes (“Read” and “Append/Update”). Telegram credentials (~5 min) Create a Telegram Bot token via BotFather. Copy your chat_id (for your target group or personal chat). Add those credentials to n8n and drop them into the “Telegram” node. Workflow parameters (~5 min) Verify the schedule in the Schedule Trigger node is set to 08:00 (or adjust to your preferred run time). In the Loop Over Items node, confirm “Batch Size” is 1 (to process one URL at a time). Adjust the Delay to avoid Request Blocking node if your site requires a longer pause (default is 20 s). In the Parse Data From The HTML Page node, double-check the CSS selector matches how prices appear on your target site. Once credentials are in place and your sheet tabs match the expected column names, the flow should be ready to activate. Total setup time is under 15 minutes—detailed notes are embedded as sticky comments throughout the workflow to help you tweak selectors, change timeouts, or adjust sheet names without digging into code.
by Robert Breen
This n8n workflow scrapes recent Instagram posts by hashtag and generates new, relevant caption ideas using OpenAI. It avoids making up suggestions by analyzing real-world content and surfacing common patterns. ✅ Use Case Marketing teams, content creators, or social media managers can: Discover what’s trending for a specific topic Automatically generate Instagram captions based on real posts Understand common caption styles for a niche Save time brainstorming ideas while staying on-brand 🧠 How It Works 1️⃣ Manual Trigger 🧩 Node: When clicking ‘Execute workflow’ Manually starts the workflow for testing or single-run execution. 2️⃣ Define the Hashtag 🧩 Node: Create Search Term Sets the value of the hashtag you'd like to scan. Default is n8n, but you can modify it to anything. { "Search_Term": "yourCustomHashtag" } 3️⃣ Scrape Instagram Posts 🧩 Node: Find Recent Posts API**: Apify Instagram Hashtag Scraper Setup**: Visit Apify Console Create an API token In n8n, go to Credentials and add HTTP Query Auth Use ?token=yourTokenHere as the query string JSON Body: { "hashtags": ["{{ $json.Search_Term }}"], "resultsLimit": 20, "resultsType": "posts" } 4️⃣ Extract Captions 🧩 Node: Set bio and follower count Extracts just the caption from each post and stores it in a clean variable for the AI agent to use. 5️⃣ Aggregate Captions 🧩 Node: Aggregate Gathers all captions into one list before processing. Useful for passing a large text block into the AI. 6️⃣ Convert to Single Text Block 🧩 Node: Convert table names and columns into single text for agent Uses a Code node to combine all captions into a single string for OpenAI to read: return [ { json: { text: items .map(item => - ${JSON.stringify(item.json)}) .join('\n\n'), }, }, ]; 7️⃣ Generate Caption Ideas with AI 🧩 Node: AI Agent Takes the combined post text and sends it to GPT-4o-mini. Includes this system message: I'm looking for ideas for posts about {{ $('Create Search Term').item.json.Search_Term }}. Here’s the last 5 posts on Instagram about the topic. Use those to help me generate a list of relevant captions. Do not make up ideas that are not like the others in the list. Output like this: { "Post Idea": ["Idea1", "Idea2"], "Most Common Post": ["common post 1", "common post 2"] } 8️⃣ Choose Language Model 🧩 Node: OpenAI Chat Model Model**: gpt-4o-mini Credential**: Use your OpenAI API key. Get it from: OpenAI API Keys Add it in n8n under OpenAI credentials. 9️⃣ Parse the AI Output 🧩 Node: Structured Output Parser Parses the GPT response into structured JSON: { "Post Idea": ["Idea1", "Idea2"], "Most Common Post": ["common post 1", "common post 2"] } 🔟 Split the Outputs 🧩 Nodes: Split Out, Split Out1 Separates the Post Idea list and Most Common Post list into individual items. 🔁 Merge for Final Output 🧩 Node: Merge Combines the two split lists into one output stream. 👤 Need More Help? Robert Breen Automation Consultant | AI Workflow Designer | n8n Expert 📧 robert@ynteractive.com 🔗 LinkedIn
by Lucas Correia
AI Automated "Viral Style" Carousels Generator for Instagram, TikTok, LinkedIn, or X 🚀 Overview Automate your social media content creation with this powerful n8n workflow! Generate engaging, viral-style carousels for Instagram, TikTok, LinkedIn, or X (Twitter) in minutes. This template leverages AI (xAI Grok) to craft compelling, high-retention text and uses n8n's Edit Image node to automatically design your slides with your custom branding. Output examples: ✨ Features AI-Powered Content:** Utilizes xAI Grok to generate witty, substantive, 7-slide carousel content based on a theme and CTA. Multi-Platform Ready:* Perfect for *Instagram carousels, **TikTok carousels, LinkedIn carousels, and X (Twitter) threads. Automated Design:** Overlays AI-generated text onto your chosen background image, creating visually consistent slides. Easy Customization:** Adapt the AI persona, font styles, colors, and background images to match your brand. Google Drive Integration:** Seamlessly downloads your background template and uploads finished carousel slides. No Code Automation:** Set up once and generate endless content with minimal effort. 💡 How it Works Input Trigger: Provide a theme and call to action (CTA) via a webhook or manual trigger. Content Generation: The AI (acting as "The Carousel Cynic") writes 7 distinct slides, each with a provocative title and a detailed description, formatted for maximum engagement. Image Assembly: Downloads a base background image from Google Drive. Loops through each of the 7 AI-generated slides. Uses the Edit Image node to dynamically add the slide's title and description to the background. Outputs sequentially numbered .png files (e.g., 1.png, 2.png). Output & Storage: Uploads all final carousel images to a specified folder in your Google Drive, ready for publishing. 🛠️ Setup Steps xAI Credentials: Add your xAI API Key to the xAI Grok Chat Model node. Google Drive Integration: Connect your Google Drive OAuth2 credentials. In the Download file node, update the File ID to point to your desired blank background image. In the Upload file node, select the Google Drive folder where you want to save the generated carousels. Customization (Optional): Adjust AI persona in the AI Agent node's "System Prompt." Modify fontSize, fontColor, positionX, and positionY in the Params Style Config node to perfectly align text on your background images. 🎁Bonus Added in workflow a Canva link to editable background style I use in my carousels. 🔑 Keywords AI, Automation, Social Media, Carousel, Instagram, TikTok, LinkedIn, X, Twitter, Content Creation, Viral Content, Marketing, Grok, xAI, Image Generation, No-code, Workflow, Productivity, Creator Economy, Digital Marketing, Engagement, Visual Content, Dynamic Image, Automated Marketing.
by n8n Team
This workflow syncs Zendesk tickets to Pipedrive contact owners. This workflow is triggered every day at 09:00 with Zendesk collecting all the tickets updated after the last execution timestamp and updating them according to Pipedrive contacts. It also adds Zendesk comments to the tickets as notes in Pipedrive. Prerequisites Pipedrive account and Pipedrive credentials Zendesk account and Zendesk credentials Note: The Pipedrive and the Zendesk accounts need to be created by the same person / with the same email. How it works Cron node triggers the workflow every day at 09:00. Zendesk node collects all the tickets updated after the last execution timestamp. If node checks if the channel in the ticket is an email, and if so, it continues the workflow. The Item Lists node removes duplicates to make search efficient. Pipedrive node searches persons by email. Set node renames and keeps only needed fields (email & person id) Merge by key node adds the Pipedrive contact id to Zendesk tickets. The HTTP Request node gets Zendesk comments for tickets and the Merge node adds them to tickets. Split node adds nodes in batches with each iteration. Item list node splits comments into separate items. Pipedrive node adds comment as notes. If node checks if the data processing is done and if not, goes back to the Split node. The Function Item node sets the new last execution timestamp.
by Ricardo Espinozaas
Use Case When having a call with a new potential customer, one of the keys to getting the most out of the call is to find out as much information as you can about them before the call. Normally this involves a lot of manual research before every call. This workflow automates this tedious work for you. What this workflow does The workflow runs every time a new call is booked via your Calendly. It then filters out personal emails, before enriching the email. If the email is attached to a company it enriches the company and upserts it in your Hubspot CRM. Setup Add Clearbit, Hubspot, and Calendly credentials. Click on Test workflow. Book a meeting on Calendly so the event starts the workflow. Be aware that you can adapt this workflow to work with your enrichment tool, CRM, and booking tool of choice.
by Autonomous Work
This workflow exports every table in a base as its own CSV, saves the files in a time-stamped folder in Amazon S3, pings you on Slack, and optionally prunes older copies. You get an automated weekly backup that is easy to inspect or re-import as needed. You can easily swap the S3 node for the storage provider of your choice. ++How it works++ Weekly Backup Schedule trigger fires weekly Sets and formats the week ex. [2025-W12] Create a folder in S3 bucket with the week Loops through all tables in Airtable base creating CSVs and uploading to the new path Slack message is sent on completion Monthly Prune Schedule trigger fires weekly Sets a cut-off date 4 weeks in the past Lists folders in S3 Deletes all folders > 4 weeks old ++Setup Steps++ Clone workflow Swap credentials for Airtable, AWS, and Slack Ensure AWS credential has appropriate IAM policy to manage bucket & objects Set workflow to "Active"