by Jay Hartley
What this workflow does Downloads the daily top podcasts of a selected genre Summarizes the content of each podcast in a few paragraphs Sends the summaries and the direct link to each podcast in a formatted email Setup Create a free API key on Taddy here: https://taddy.org/signup/developers Input your user number and API key into the TaddyTopDaily node in the header parameters X-USER-ID and X-API-KEY respectively. Create access credentials for your Gmail as described here: https://developers.google.com/workspace/guides/create-credentials. Use the credentials from your client_secret.json in the Gmail node. In the Genre node, set the genre of podcasts you want a summary for. Valid values are: TECHNOLOGY, NEWS, ARTS, COMEDY, SPORTS, FICTION, etc. Look at api.taddy.org for the full list (they will be displayed in the help docs as PODCASTSERIES_TECHNOLOGY, PODCASTSERIES_NEWS, etc.) Enter your email address in the Gmail node. Change the schedule time for sending email from Schedule to whichever time you want to receive the email. Test: Hit Test Workflow. Check your email for the results. That's it! It should take less than 5 minutes total.
by Robert Breen
This n8n workflow scrapes recent Instagram posts by hashtag and generates new, relevant caption ideas using OpenAI. It avoids making up suggestions by analyzing real-world content and surfacing common patterns. ✅ Use Case Marketing teams, content creators, or social media managers can: Discover what’s trending for a specific topic Automatically generate Instagram captions based on real posts Understand common caption styles for a niche Save time brainstorming ideas while staying on-brand 🧠 How It Works 1️⃣ Manual Trigger 🧩 Node: When clicking ‘Execute workflow’ Manually starts the workflow for testing or single-run execution. 2️⃣ Define the Hashtag 🧩 Node: Create Search Term Sets the value of the hashtag you'd like to scan. Default is n8n, but you can modify it to anything. { "Search_Term": "yourCustomHashtag" } 3️⃣ Scrape Instagram Posts 🧩 Node: Find Recent Posts API**: Apify Instagram Hashtag Scraper Setup**: Visit Apify Console Create an API token In n8n, go to Credentials and add HTTP Query Auth Use ?token=yourTokenHere as the query string JSON Body: { "hashtags": ["{{ $json.Search_Term }}"], "resultsLimit": 20, "resultsType": "posts" } 4️⃣ Extract Captions 🧩 Node: Set bio and follower count Extracts just the caption from each post and stores it in a clean variable for the AI agent to use. 5️⃣ Aggregate Captions 🧩 Node: Aggregate Gathers all captions into one list before processing. Useful for passing a large text block into the AI. 6️⃣ Convert to Single Text Block 🧩 Node: Convert table names and columns into single text for agent Uses a Code node to combine all captions into a single string for OpenAI to read: return [ { json: { text: items .map(item => - ${JSON.stringify(item.json)}) .join('\n\n'), }, }, ]; 7️⃣ Generate Caption Ideas with AI 🧩 Node: AI Agent Takes the combined post text and sends it to GPT-4o-mini. Includes this system message: I'm looking for ideas for posts about {{ $('Create Search Term').item.json.Search_Term }}. Here’s the last 5 posts on Instagram about the topic. Use those to help me generate a list of relevant captions. Do not make up ideas that are not like the others in the list. Output like this: { "Post Idea": ["Idea1", "Idea2"], "Most Common Post": ["common post 1", "common post 2"] } 8️⃣ Choose Language Model 🧩 Node: OpenAI Chat Model Model**: gpt-4o-mini Credential**: Use your OpenAI API key. Get it from: OpenAI API Keys Add it in n8n under OpenAI credentials. 9️⃣ Parse the AI Output 🧩 Node: Structured Output Parser Parses the GPT response into structured JSON: { "Post Idea": ["Idea1", "Idea2"], "Most Common Post": ["common post 1", "common post 2"] } 🔟 Split the Outputs 🧩 Nodes: Split Out, Split Out1 Separates the Post Idea list and Most Common Post list into individual items. 🔁 Merge for Final Output 🧩 Node: Merge Combines the two split lists into one output stream. 👤 Need More Help? Robert Breen Automation Consultant | AI Workflow Designer | n8n Expert 📧 robert@ynteractive.com 🔗 LinkedIn
by Lucas Correia
AI Automated "Viral Style" Carousels Generator for Instagram, TikTok, LinkedIn, or X 🚀 Overview Automate your social media content creation with this powerful n8n workflow! Generate engaging, viral-style carousels for Instagram, TikTok, LinkedIn, or X (Twitter) in minutes. This template leverages AI (xAI Grok) to craft compelling, high-retention text and uses n8n's Edit Image node to automatically design your slides with your custom branding. Output examples: ✨ Features AI-Powered Content:** Utilizes xAI Grok to generate witty, substantive, 7-slide carousel content based on a theme and CTA. Multi-Platform Ready:* Perfect for *Instagram carousels, **TikTok carousels, LinkedIn carousels, and X (Twitter) threads. Automated Design:** Overlays AI-generated text onto your chosen background image, creating visually consistent slides. Easy Customization:** Adapt the AI persona, font styles, colors, and background images to match your brand. Google Drive Integration:** Seamlessly downloads your background template and uploads finished carousel slides. No Code Automation:** Set up once and generate endless content with minimal effort. 💡 How it Works Input Trigger: Provide a theme and call to action (CTA) via a webhook or manual trigger. Content Generation: The AI (acting as "The Carousel Cynic") writes 7 distinct slides, each with a provocative title and a detailed description, formatted for maximum engagement. Image Assembly: Downloads a base background image from Google Drive. Loops through each of the 7 AI-generated slides. Uses the Edit Image node to dynamically add the slide's title and description to the background. Outputs sequentially numbered .png files (e.g., 1.png, 2.png). Output & Storage: Uploads all final carousel images to a specified folder in your Google Drive, ready for publishing. 🛠️ Setup Steps xAI Credentials: Add your xAI API Key to the xAI Grok Chat Model node. Google Drive Integration: Connect your Google Drive OAuth2 credentials. In the Download file node, update the File ID to point to your desired blank background image. In the Upload file node, select the Google Drive folder where you want to save the generated carousels. Customization (Optional): Adjust AI persona in the AI Agent node's "System Prompt." Modify fontSize, fontColor, positionX, and positionY in the Params Style Config node to perfectly align text on your background images. 🎁Bonus Added in workflow a Canva link to editable background style I use in my carousels. 🔑 Keywords AI, Automation, Social Media, Carousel, Instagram, TikTok, LinkedIn, X, Twitter, Content Creation, Viral Content, Marketing, Grok, xAI, Image Generation, No-code, Workflow, Productivity, Creator Economy, Digital Marketing, Engagement, Visual Content, Dynamic Image, Automated Marketing.
by n8n Team
This workflow syncs Zendesk tickets to Pipedrive contact owners. This workflow is triggered every day at 09:00 with Zendesk collecting all the tickets updated after the last execution timestamp and updating them according to Pipedrive contacts. It also adds Zendesk comments to the tickets as notes in Pipedrive. Prerequisites Pipedrive account and Pipedrive credentials Zendesk account and Zendesk credentials Note: The Pipedrive and the Zendesk accounts need to be created by the same person / with the same email. How it works Cron node triggers the workflow every day at 09:00. Zendesk node collects all the tickets updated after the last execution timestamp. If node checks if the channel in the ticket is an email, and if so, it continues the workflow. The Item Lists node removes duplicates to make search efficient. Pipedrive node searches persons by email. Set node renames and keeps only needed fields (email & person id) Merge by key node adds the Pipedrive contact id to Zendesk tickets. The HTTP Request node gets Zendesk comments for tickets and the Merge node adds them to tickets. Split node adds nodes in batches with each iteration. Item list node splits comments into separate items. Pipedrive node adds comment as notes. If node checks if the data processing is done and if not, goes back to the Split node. The Function Item node sets the new last execution timestamp.
by Manuel
Effortlessly optimize your workflow by automatically save all files you are receiving on Telegram to a Google Drive Folder. How it works Retrieve a message sent to your Telegram Bot containing a file Upload the file to your Google Drive Folder Set up Steps Create a Telegram Account and a Telegram Bot and connect your Telegram Bot to n8n by following the official n8n instructions Create a Google Drive Folder Connect your Google Drive with n8n following the official n8n instructions Set the right folder in the Google Drive node Use case examples Backup and Recovery Cross-Platform Access File Organization and Management File Collaboration and Sharing Storage Space Management
by Karol
How it works This workflow automates publishing content from any RSS feed directly to Facebook and Instagram. It reads new RSS entries, extracts the article content, generates a short social-media-friendly summary using an AI model, and then creates an AI-generated image based on the topic. The post is uploaded to Facebook and Instagram (via Graph API) and logged in Google Sheets for reference. Finally, a Telegram bot sends you a notification with links to the published posts. Set up steps Insert your RSS feed URL in the RSS Feed Trigger node. Configure Google Sheets credentials and replace the example sheet with your own. In Supabase Config, insert your Supabase URL and bucket name. In Facebook/Instagram nodes, replace [INSERT_YOUR_SITE_ID] with your own page or account ID. Connect your Facebook Graph API credentials (remove hardcoded tokens). Connect your OpenAI / Anthropic / Gemini credentials for text and image generation. Set up your Telegram Bot credentials if you want to receive notifications. Notes • Sticky notes inside the workflow explain each section (RSS trigger, filtering, content generation, posting, logging, notifications). • No credentials are saved in the template – you must connect your own before running. • All generated content (text + images) is fully automated but can be customized (e.g. change AI prompts for your preferred style).
by Trung Tran
🎧 IT Voice Support Automation Bot – Telegram Voice Message to JIRA ticket with OpenAI Whisper > Automatically process IT support requests submitted via Telegram voice messages by transcribing, extracting structured data, creating a JIRA ticket, and notifying relevant parties. 🧑💼 Who’s it for Internal teams that handle IT support but want to streamline voice-based requests. Employees who prefer using mobile/voice to report incidents or ask for support. Organizations aiming to integrate conversational AI into existing support workflows. ⚙️ How it works / What it does A user sends a voice message to a Telegram bot. The system checks whether it’s an audio message. If valid, the audio is: Downloaded Transcribed via OpenAI Whisper Backed up to Google Drive The transcription and file metadata are merged. The merged content is processed through an AI Agent (GPT) to extract structured request info. A JIRA ticket is created using the extracted data. The IT team is notified via Slack (or other channels). The requester receives a Telegram confirmation message with the JIRA ticket link. If the input is not audio, a polite rejection message is sent. 📌 Key Features Supports voice-based ticket creation Accurate transcription using Whisper Context-aware request parsing using GPT-4.1 mini Fully automated ticket creation in JIRA Notifies both IT and the original requester Cloud backup of original voice messages (Google Drive) 🛠️ Setup Instructions Prerequisites | Component | Required | |----------|----------| | Telegram Bot & API Key | ✅ | | OpenAI Whisper / Transcription Model | ✅ | | Google Drive Credentials (OAuth2) | ✅ | | Google Sheets or other storage (optional) | ⬜ | | JIRA Cloud API Access | ✅ | | Slack Bot or Webhook | ✅ | Workflow Steps Telegram Voice Message Trigger: Starts the flow when a user sends a voice message. Is Audio Message?: If false → reply "only voice is supported" Download Audio: Download .oga file from Telegram. Transcribe Audio: Use OpenAI Whisper to get text transcript. Backup to Google Drive: Upload original voice file with metadata. Merge Results: Combine transcript and metadata. Pre-process Output: Clean formatting before AI extraction. Transcript Processing Agent: GPT-based agent extracts: Requester name, department Request title & description Priority & request type Submit JIRA Request Ticket: Create ticket from AI-extracted data. Setup Slack / Email / Manual Steps: Optional internal routing or approvals. Inform Reporter via Telegram: Sends confirmation message with JIRA ticket link. 🔧 How to Customize Replace JIRA with Zendesk, GitHub Issues, or other ticketing tools. Change Slack to Microsoft Teams or Email. Add Notion/Airtable logging. Enhance agent to extract department from user ID or metadata. 📦 Requirements | Integration | Notes | |-------------|-------| | Telegram Bot | Used for input/output | | Google Drive | Audio backup | | OpenAI GPT + Whisper | Transcript & Extraction | | JIRA | Ticketing platform | | Slack | Team notification | Built with ❤️ using n8n
by Rodrigue Gbadou
How it works Continuous monitoring**: Real-time surveillance of supplier performance, financial health, and operational status Risk scoring**: AI-powered assessment of supplier risks across multiple dimensions (financial, operational, geopolitical) Automated alerts**: Instant notifications when supplier risk levels exceed predefined thresholds Contingency activation**: Automatic triggering of backup suppliers and alternative sourcing plans Set up steps Supplier database**: Connect your ERP/procurement system with complete supplier information Financial data sources**: Integrate with credit monitoring services (Dun & Bradstreet, Experian) News monitoring**: Configure news APIs for real-time supplier-related news tracking Performance metrics**: Set up KPIs tracking (delivery times, quality scores, compliance) Alert systems**: Configure Slack, Teams, or email notifications for risk alerts Backup protocols**: Define alternative supplier activation procedures Key Features 🔍 360° supplier visibility**: Complete view of supplier ecosystem health and performance ⚡ Real-time risk detection**: Immediate identification of potential supply chain disruptions 📊 Predictive analytics**: Forecasting potential supplier issues before they impact operations 🚨 Automated escalation**: Risk-based alert system with appropriate stakeholder notifications 📈 Performance benchmarking**: Continuous comparison against industry standards and peers 🔄 Contingency management**: Automated backup supplier activation and procurement rerouting 🌍 Geopolitical monitoring**: Tracking of regulatory changes and political risks by region 💰 Cost impact analysis**: Financial impact assessment of supplier disruptions Risk categories monitored Financial stability**: Credit scores, payment delays, bankruptcy indicators Operational performance**: Delivery reliability, quality metrics, capacity utilization Compliance status**: Regulatory adherence, certifications, audit results Geopolitical risks**: Political instability, trade restrictions, regulatory changes Environmental factors**: Natural disasters, climate risks, sustainability metrics Cyber security**: Security breaches, data protection compliance Automated responses Low risk (0-30)**: Routine monitoring and performance tracking Medium risk (31-60)**: Enhanced monitoring with supplier engagement High risk (61-80)**: Immediate supplier contact and mitigation planning Critical risk (81-100)**: Emergency protocols and backup supplier activation Integration capabilities ERP systems**: SAP, Oracle, Microsoft Dynamics for procurement data Risk platforms**: Resilinc, Riskmethods, Prewave for specialized risk intelligence Financial services**: Credit monitoring and financial health assessment News APIs**: Real-time news monitoring and sentiment analysis Communication tools**: Slack, Teams, email for stakeholder notifications This workflow provides comprehensive supply chain visibility and proactive risk management, enabling companies to maintain operational continuity while minimizing disruption costs.
by Mikal Hayden-Gates
Overview Automates your complete social media content pipeline: sources articles from Wallabag RSS, generates platform-specific posts with AI, creates contextual images, and publishes via GetLate API. Built with 63 nodes across two workflows to handle LinkedIn, Instagram, and Bluesky—with easy expansion to more platforms. Ideal for: Content marketers, solo creators, agencies, and community managers maintaining a consistent multi-platform presence with minimal manual effort. How It Works Two-Workflow Architecture: Content Aggregation Workflow Monitors Wallabag RSS feeds for tagged articles (#to-share-linkedin, #to-share-instagram, etc.) Extracts and converts content from HTML to Markdown Stores structured data in Airtable with platform assignment AI Generation & Publishing Workflow Scheduled trigger queries Airtable for unpublished content Routes to platform-specific sub-workflows (LinkedIn, Instagram, Bluesky) LLM generates optimized post text and image prompts based on custom brand parameters Optionally generates AI images and hosts them on Imgbb CDN Publishes via GetLate API (immediate or draft mode) Updates Airtable with publication status and metadata Key Features: Tag-based content routing using Wallabag's native system Swappable AI providers (Groq, OpenAI, Anthropic) Platform-specific optimization (tone, length, hashtags, CTAs) Modular design—duplicate sub-workflows to add new platforms in \~30 minutes Centralized Airtable tracking with 17 data points per post Set Up Steps Setup time: \~45-60 minutes for initial configuration Create accounts and get API keys (\~15 min) Wallabag (with RSS feeds enabled) GetLate (social media publishing) Airtable (create base with provided schema—see sticky notes) LLM provider (Groq, OpenAI, or Anthropic) Image service (Hugging Face, Fal.ai, or Stability AI) Imgbb (image hosting) Configure n8n credentials (\~10 min) Add all API keys in n8n's credential manager Detailed credential setup instructions in workflow sticky notes Set up Airtable database (\~10 min) Create "RSS Feed - Content Store" base Add 19 required fields (schema provided in workflow sticky notes) Get Airtable base ID and API key Customize brand prompts (\~15 min) Edit "Set Custom SMCG Prompt" node for each platform Define brand voice, tone, goals, audience, and image preferences Platform-specific examples provided in sticky notes Configure platform settings (\~10 min) Set GetLate account IDs for each platform Enable/disable image generation per platform Choose immediate publish vs. draft mode Adjust schedule trigger frequency Test and deploy Tag test articles in Wallabag Monitor the first few executions in draft mode Activate workflows when satisfied with the output Important: This is a proof-of-concept template. Test thoroughly with draft mode before production use. Detailed setup instructions, troubleshooting tips, and customization guidance are in the workflow's sticky notes. Technical Details 63 nodes**: 9 Airtable operations, 8 HTTP requests, 7 code nodes, 3 LangChain LLM chains, 3 RSS triggers, 3 GetLate publishers Supports**: Multiple LLM providers, multiple image generation services, unlimited platforms via modular architecture Tracking**: 17 metadata fields per post, including publish status, applied parameters, character counts, hashtags, image URLs Prerequisites n8n instance (self-hosted or cloud) Accounts: Wallabag, GetLate, Airtable, LLM provider, image generation service, Imgbb Basic understanding of n8n workflows and credential configuration Time to customize prompts for your brand voice Detailed documentation, Airtable schema, prompt examples, and troubleshooting guides are in the workflow's sticky notes. Category Tags #social-media-automation, #ai-content-generation, #rss-to-social, #multi-platform-posting, #getlate-api, #airtable-database, #langchain, #workflow-automation, #content-marketing
by Joseph LePage
Generate SEO-Optimized WordPress Content with Perplexity Research Who is This For? This workflow is ideal for content creators, marketers, and businesses looking to streamline the creation of SEO-optimized blog posts for WordPress. It is particularly suited for professionals in the AI consulting and workflow automation industries. What Problem Does This Workflow Solve? Creating high-quality, SEO-friendly blog posts can be time-consuming and challenging, especially when trying to balance research, formatting, and publishing. This workflow automates the process by integrating research capabilities, AI-driven content creation, and seamless WordPress publishing. It reduces manual effort while ensuring professional-grade output. What This Workflow Does Research: Gathers detailed insights from Perplexity AI based on user-provided queries. Content Generation: Uses OpenAI models to create structured blog posts, including titles, slugs, meta descriptions, and HTML content optimized for WordPress. Image Handling: Automatically fetches and uploads featured images to WordPress posts. Publishing: Drafts the blog post directly in WordPress with all necessary formatting and metadata. Notification: Sends a success message via Telegram upon completion. Setup Guide Prerequisites: A WordPress account with API access. OpenAI API credentials. Perplexity AI API credentials. Telegram bot credentials for notifications. Steps: Import the workflow into your n8n instance. Configure API credentials for WordPress, OpenAI, Perplexity AI, and Telegram. Customize the form trigger to define your research query. Test the workflow using sample queries to ensure smooth execution. How to Customize This Workflow to Your Needs Modify the research query prompt in the "Form Trigger" node to suit your industry or niche. Adjust content generation guidelines in the "Copywriter AI Agent" node for specific formatting preferences. Replace the image URL in the "Set Image URL" node with your own source or dynamic image selection logic.
by Yang
📄 What this workflow does This workflow captures a full-page screenshot of any website added to a Google Sheet and automatically uploads the screenshot to a designated Google Drive folder. It uses Dumpling AI’s screenshot API to generate the image and manages file storage through Google Drive. 👤 Who is this for This is ideal for: Marketers and outreach teams capturing snapshots of client or lead websites Lead generation specialists tracking landing page visuals Researchers or analysts who need to archive website visuals from URLs Anyone looking to automate website screenshot collection at scale ✅ Requirements A Google Sheet with a column labeled Website where URLs will be added Dumpling AI** API access for screenshot capture A connected Google Drive account with an accessible folder to store screenshots ⚙️ How to set up Replace the Google Sheet and folder IDs in the workflow with your own. Connect your Dumpling AI and Google credentials in n8n. Make sure your sheet contains a Website column with valid URLs. Activate the workflow to begin watching for new entries. 🔁 How it works (Workflow Steps) Watch New Row in Google Sheets: Triggers when a new row is added to the sheet. Request Screenshot from Dumpling AI: Sends the website URL to Dumpling AI and gets a screenshot URL. Download Screenshot: Fetches the image file from the returned URL. Upload Screenshot to Google Drive: Uploads the file to a selected folder in Google Drive. 🛠️ Customization Ideas Add timestamped filenames using the current date or domain name Append the Google Drive URL back to the same row in the sheet for easy access Extend the workflow to send Slack or email notifications when screenshots are saved Add filters to validate URLs before sending them to Dumpling AI
by Immanuel
Automated Research Report Generation with OpenAI, Wikipedia, Google Search, Gmail/Telegram and PDF Output Description What Problem Does This Solve? 🛠️ This workflow automates the process of generating professional research reports for researchers, students, and professionals. It eliminates manual research and report formatting by aggregating data, generating content with AI, and delivering the report as a PDF via Gmail or Telegram. Target audience: Researchers, students, educators, and professionals needing quick, formatted research reports. What Does It Do? 🌟 Aggregates research data from Wikipedia, Google Search, and SerpApi. Refines user queries and generates structured content using OpenAI. Converts the content into a professional HTML report, then to PDF. Sends the PDF report via Gmail or Telegram. Key Features 📋 Real-time data aggregation from multiple sources. AI-driven content generation with OpenAI. Automated HTML-to-PDF conversion for professional reports. Flexible delivery via Gmail or Telegram. Error handling for robust execution. Setup Instructions Prerequisites ⚙️ n8n Instance**: Self-hosted or cloud n8n instance. API Credentials**: OpenAI API: API key with GPT model access, stored in n8n credentials. SerpApi (Google Search): API key from SerpApi, stored in n8n credentials (do not hardcode in nodes). Gmail API: Credentials from Google Cloud Console with Gmail scope. Telegram API: Bot token from BotFather on Telegram. Installation Steps 📦 Import the Workflow: Copy the workflow JSON from the "Template Code" section below. Import it into n8n via "Import from File" or "Import from URL". Configure Credentials: Add API credentials in n8n’s Credentials section for OpenAI, SerpApi, Gmail, and Telegram. Assign credentials to respective nodes. For example: In the SerpApi Google Search node, use n8n credentials for SerpApi: api_key={{ $credentials.SerpApiKey }}. In the Send Research PDF on Gmail node, use Gmail credentials. In the Send PDF to Telegram node, use Telegram bot credentials. Set Up Nodes: OpenAI Nodes (Research AI Agent, OpenAI Chat Model, OpenAI Chat Middle Memory): Update the model (e.g., gpt-4o) and prompt as needed. Input Validation (Input Validation node): Ensure your input query format matches the expected structure (e.g., topic: "AI ethics"). Delivery Options (Send Research PDF on Gmail, Send PDF to Telegram): Configure recipient email or Telegram chat ID. Test the Workflow: Run the workflow by clicking the "Test Workflow" node. Verify that the research report PDF is generated and sent via Gmail or Telegram. How It Works High-Level Steps 🔍 Query Refinement**: Refines the input query for better research. Aggregate Data**: Fetches data from Wikipedia, Google Search, and SerpApi. Generate Report**: Uses OpenAI to create a structured report. Convert to PDF**: Converts the report to HTML, then PDF. Deliver Report**: Sends the PDF via Gmail or Telegram. Detailed descriptions are available in the sticky notes within the workflow screenshot above. Node Names and Actions Research and Report Generation Test Workflow: Triggers the workflow for testing. Input Validation: Validates the input query. Query Refiner: Refines the query for better results. Research AI Agent: Coordinates research using OpenAI. OpenAI Chat Model: Generates content for the report. Structured Output Parser: Parses OpenAI output into structured data. OpenAI Chat Middle Memory: Retains context during research. Wikipedia Google Search: Fetches data from Wikipedia. SerpApi Google Search: Fetches data via SerpApi. Merge Split Items: Merges data from multiple sources. Aggregate: Aggregates all research data. Generate PDF HTML: Creates an HTML report. Convert HTML to PDF: Converts HTML to PDF. Download PDF: Downloads the PDF file. Send PDF to Telegram: Sends the PDF via Telegram. Send Research PDF on Gmail: Sends the PDF via Gmail. Customization Tips Expand Data Sources** 📡: Add more sources (e.g., academic databases) by adding nodes to Merge Split Items. Change Report Style** ✍️: Update the Generate PDF HTML node to modify the HTML template (e.g., adjust styling or sections). Alternative Delivery** 📧: Add nodes to send the PDF via other platforms (e.g., Slack). Adjust AI Model** 🧠: Modify the OpenAI Chat Model node to use a different model (e.g., gpt-3.5-turbo).