by Yaron Been
Prunaai Flux Schnell Image Generator Description This is a 3x faster FLUX.1 [schnell] model from Black Forest Labs, optimised with pruna with minimal quality loss. Contact us for more at pruna.ai Overview This n8n workflow integrates with the Replicate API to use the prunaai/flux-schnell model. This powerful AI model can generate high-quality image content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Prompt for generated image Optional Parameters seed** (integer, default: None): Random seed. Set for reproducible generation megapixels** (string, default: 1): Approximate number of megapixels for generated image speed_mode** (string, default: Juiced 🔥 (default)): Run faster predictions with model optimized for speed num_outputs** (integer, default: 1): Number of outputs to generate aspect_ratio** (string, default: 1:1): Aspect ratio of the output image output_format** (string, default: jpg): Format of the output images output_quality** (integer, default: 80): Quality when saving the output images, from 0 to 100. 100 is best quality, 0 is lowest quality. Not relevant for .png outputs num_inference_steps** (integer, default: 4): Number of denoising steps. 4 is recommended, and lower number of steps produce lower quality outputs, faster. How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate image content Access the generated output from the final node API Reference Model: prunaai/flux-schnell API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of image generation parameters
by Shahrukh
AI-Powered Workflow for Auto-Responding to Positive Cold Email Replies This workflow is designed for agencies, freelancers, and sales teams who want to turn positive cold email replies into booked meetings automatically—without hiring VAs or spending hours on manual responses. ❓ The Problem Most teams waste time replying manually or pay for virtual assistants, leading to delays and missed opportunities. This template eliminates that bottleneck. ✅ What the Workflow Does Detects positive replies from Instantly.ai campaigns Uses AI to analyze intent and craft natural, human-like responses Adds personalization to keep replies authentic Includes Calendly links, product docs, or FAQs based on the lead’s intent Sends responses instantly—so you never miss a hot lead again No robotic AI text. Just smooth, human-style emails that get booked calls faster. 👥 Who is This For? Agencies** running Instantly.ai or similar outbound tools Founders** handling their own cold email outreach Sales teams** looking to automate follow-up and booking Anyone who gets 5–20 positive replies a week and wants to 2x–4x conversions ✅ Requirements n8n** (Cloud or self-hosted) Instantly.ai account** with API access OpenAI API key** (stored securely in n8n credentials) (Optional) Calendly or booking link, Notion or Google Docs for resources ⚙️ How to Set Up Import the workflow into n8n Add your Instantly.ai API credentials and OpenAI key using n8n’s credential manager Customize the AI prompt for your tone, CTA, and offer Insert your Calendly or booking link in the response template Test with one positive reply to confirm filtering and response quality Activate the workflow to auto-reply in real time 🔧 How to Customize Adjust the filtering logic for different keywords or intent signals Add branching for multiple booking links (e.g., based on region or service type) Push responses to a CRM for tracking Include extra resources like case studies or pricing docs
by Evoort Solutions
🎬 YouTube Video to Blog – Multilingual Blog Generator Convert YouTube videos into SEO-friendly blog posts in just seconds using this fully automated n8n workflow. Perfect for content creators, marketers, educators, and bloggers looking to repurpose video content without manual transcription or formatting. 🔧 What It Does 📥 Accepts a YouTube video URL and preferred language via a simple form 🧠 Uses a third-party API to convert the video into a blog-style article 📄 Automatically inserts the generated content into a Google Docs document 🌍 Supported Languages Supports all major languages, including but not limited to: English Hindi French German Gujarati 🎯 The workflow is flexible and can generate blog content in any language supported by the API. Just select your language when submitting the form. 🚀 Benefits ⏱️ Time-Saving: Eliminate manual video transcription and formatting 🌐 Multilingual: Easily generate blogs in multiple languages 📚 Centralized Storage: Store all generated blogs in a single Google Docs file 🔧 Customizable: Extend the flow to auto-publish, email, or analyze content 🧠 Use Cases Repurpose YouTube content into keyword-rich blog posts Generate multilingual content for global reach Convert educational videos into study guides or summaries Create email newsletters or social media posts from video content 🛠️ Requirements ✅ An n8n instance (self-hosted or cloud) 🔑 RapidAPI key for youtube-to-blog.p.rapidapi.com 🧾 A Google Docs account with API access 🚨 Note: Be sure to update the API key and Google Docs URL with your own credentials before activating the workflow. Create your free n8n account and set up the workflow in just a few minutes using the link below: 👉 Start Automating with n8n Save time, stay consistent, and grow your LinkedIn presence effortlessly!
by Davide
This workflow builds a conversational AI chatbot agent using Claude 3.7 Sonnet model with the new . It enhances standard LLM capabilities with Anthropic’s features: Web Search and Think: Real-time web search**, to answer up-to-date factual queries. A “Think” function, to support internal reasoning and memory-like behavior by Anthropic. A memory buffer, allowing the agent to maintain conversation history. A system prompt defining clear ethical, functional, and formatting rules for interaction. When a user sends a message (trigger), the chatbot evaluates the query, optionally performs a web search if needed, processes the result using Claude, and responds accordingly. ✅ Advantages 🧠 Enhanced Reasoning Abilities** The Think tool allows the agent to simulate deep thought processes or contextual memory storage, improving conversational intelligence. 🌐 Real-Time Knowledge via Web Search** The integrated web_search tool enables the agent to fetch the latest information from the internet, making it ideal for dynamic or news-driven use cases. 🧾 Contextual Responses with Memory Buffer** The inclusion of a memory buffer allows the agent to maintain state across messages, improving dialogue flow and continuity. 🛡️ Built-in Ethical Guidelines** The system prompt enforces privacy, factual integrity, neutrality, and ethical response generation, making the agent safe for public or enterprise use. How It Works Chat Trigger: The workflow begins when a chat message is received via a webhook. This triggers the AI Agent to process the user's query. AI Agent Processing: The AI Agent analyzes the query to determine if it requires information from the website or external sources. It follows a structured approach: For website-related queries, it uses the provided context. For external information, it employs the web_search tool to fetch up-to-date data from the internet. The Think tool is used for internal reasoning or caching thoughts without altering data. Language Model: The Anthropic Chat Model (Claude 3.7 Sonnet) generates responses based on the analyzed query, incorporating website context or web search results. Memory: A simple memory buffer retains context from previous interactions to maintain continuity in conversations. Output: The final response is delivered to the user, excluding internal processes like web searches or reasoning steps. Set Up Steps Configure Nodes: Chat Trigger: Set up the webhook to receive user messages. AI Agent: Define the system message and rules for handling queries. Anthropic Chat Model: Select the Claude 3.7 Sonnet model and configure parameters like maxTokensToSample. Memory: Initialize the memory buffer to store conversation context. Tools: web_search: Configure the HTTP request to the Anthropic API for web searches, including headers and authentication. Think: Set up the tool for internal reasoning. Connect Nodes: Link the Chat Trigger to the AI Agent. Connect the Anthropic Chat Model, Memory, and Tools (web_search and Think) to the AI Agent. Credentials: Ensure the Anthropic API credentials are correctly configured for both the chat model and the web_search tool. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Samir Saci
Tags: Sustainability, CSRD, Reporting, ESG, Compliance, Automation Context Hey! I'm Samir, a Supply Chain Engineer and Data Scientist from Paris, founder of LogiGreen Consulting We help companies automate sustainability workflows using AI, Data Analytics, and No-Code tools like N8N. > Sustainability Reporting meets Automation with n8n! 📬 For business inquiries, you can add me on Here What is a CSRD XHTML Report? Under the Corporate Sustainability Reporting Directive (CSRD), companies must publish their ESG disclosures in a machine-readable XHTML format, embedding XBRL tags that make the report structured and standardized. These files must follow strict formatting and tagging rules to ensure compliance, traceability, and accessibility for both regulators and analysts. Who is this template for? This workflow is designed for sustainability teams, ESG consultants, or developers who want to automatically check the structure and format of CSRD reports submitted in XHTML. How does it work? This N8N workflow automates the audit process: 📤 Input Node → Uploads or fetches the XHTML file via URL or Webhook. 🧪 Validates Structure → Uses a custom code node to parse HTML and identify required tags (e.g., <ix:nonNumeric>, namespaces). 📋 Outputs a Report → Returns a summary report of errors, warnings, and key metadata (like entity name, reporting period). 📤 Export Option → Save the results in Google Sheets or send via email. Prerequisite A sample XHTML file that you can find in my GitHub Repository Google Sheets API* and *OpenAI API** credentials Next Steps Follow the sticky notes inside each node to adjust parsing rules or extend validation to specific XBRL tags relevant to your sector (e.g., GHG emissions, water usage). *📺 Check my complete tutorial to understand how to use it: * 🎥 Check My Tutorial 🚀 Interested in combining CSRD compliance with automation and analytics? Let’s connect on LinkedIn Notes This workflow includes an example XHTML file to test the validator. You can plug this into your internal systems or even extend it with AI to auto-summarize the sustainability report. This workflow has been created with N8N 1.82.1 Submitted: April 3rd, 2025
by Dataki
This workflow enriches new Pipedrive organization's data by adding a note to the organization object in Pipedrive. It assumes there is a custom "website" field in your Pipedrive setup, as data will be scraped from this website to generate a note using OpenAI. Then, a notification is sent in Slack. ⚠️ Disclaimer This workflow uses a scraping API. Before using it, ensure you comply with the regulations regarding web scraping in your country or state. Important Notes The OpenAI model used is GPT-4o, chosen for its large input token capacity. However, it is not the cheapest model if cost is very important to you. The system prompt in the OpenAI Node generates output with relevant information, but feel free to improve or modify it according to your needs. How It Works Node 1: Pipedrive Trigger - An Organization is Created This is the trigger of the workflow. When an organization object is created in Pipedrive, this node is triggered and retrieves the data. Make sure you have a "website" custom field in Pipedrive (the name of the field in the n8n node will appear as a random ID and not with the Pipedrive custom field name). Node 2: ScrapingBee - Get Organization's Website's Homepage Content This node scrapes the content from the URL of the website associated with the Pipedrive Organization created in Node 1. The workflow uses the ScrapingBee API, but you can use any preferred API or simply the HTTP request node in n8n. Node 3: OpenAI - Message GPT-4o with Scraped Data This node sends HTML-scraped data from the previous node to the OpenAI GPT-4o model. The system prompt instructs the model to extract company data, such as products or services offered and competitors (if known by the model), and format it as HTML for optimal use in a Pipedrive Note. Node 4: Pipedrive - Create a Note with OpenAI Output This node adds a Note to the Organization created in Pipedrive using the OpenAI node output. The Note will include the company description, target market, selling products, and competitors (if GPT-4o was able to determine them). Node 5 & 6: HTML To Markdown & Code - Markdown to Slack Markdown These two nodes format the HTML output to Slack Markdown. The Note created in Pipedrive is in HTML format, as specified by the System Prompt of the OpenAI Node. To send it to Slack, it needs to be converted to Markdown and then to Slack Markdown. Node 7: Slack - Notify This node sends a message in Slack containing the Pipedrive Organization Note created with this workflow.
by AmirHossein MnasouriZade
📦 Send Telegram Notifications for New WooCommerce Orders This workflow automatically sends a Telegram notification when an order status in WooCommerce changes to "Processing." Perfect for online store owners who want instant updates on order fulfillment. ⚙️ Set Up Telegram Alerts for WooCommerce Orders Configure WooCommerce Webhook to trigger on order updates. Create a Telegram Bot and obtain the API token. Set Up Telegram Credentials in n8n. Configure the Telegram Node with your chat ID. Activate and Test the workflow by placing a new order. ##💡 Notes You can customize the message format in the 🖋️ Design Message Template node to include additional order details. Contact me on [Telegram]: https://t.me/amir676080 Message structure includes the following details 🆔 Order Number: 11234 👦🏻 Customer Name: John Doe 💵 Amount: 299.99 USD 📅 Order Date: ➖ 25th November 2024 at 14:42 🏙 City: New York 📞 Phone: +1 555-1234 ✍🏻 Order Note: Fast delivery requested 📦 Ordered Products: 🔹 Wireless Earbuds (2 items) 📝 Type: Premium Sound Edition Contact me on [Telegram]: https://t.me/amir676080
by Samir Saci
Tags*: Sustainability, Supply Chain, AI Agent, CO2 Emissions, Carbon Interface API, Logistics, Automation Context Hi! I’m Samir — a Supply Chain Engineer and Data Scientist based in Paris, and founder of LogiGreen Consulting. I help logistics teams reduce their environmental footprint by combining AI automation and carbon estimation APIs. This workflow is part of our green logistics initiative, allowing businesses to track the CO₂ emissions of last-mile or regional shipments. > Automate carbon tracking for shipping operations with n8n! 📬 For business inquiries, feel free to connect with me on LinkedIn Who is this template for? This workflow is designed for logistics coordinators, transportation planners, or sustainability officers who want to estimate and record emissions for B2B shipments. Let’s imagine your carrier sends a shipment confirmation email after a pickup is scheduled: An AI Agent reads the email and extracts structured data: addresses, distance, cargo weight, and delivery time. The Carbon Interface API is then called to calculate CO₂ emissions based on weight and distance, and the results are stored in a Google Sheet. How does it work? This workflow automates the process of tracking CO₂ emissions for scheduled shipments: 📨 Gmail Trigger captures shipment confirmation emails 🧠 AI Agent parses the shipment info (pickup, delivery, weight, distance) 🚚 Carbon Interface API estimates CO₂ emissions 📊 Google Sheets is used to store shipment metadata and carbon results Steps: 💌 Trigger on new shipment confirmation email 🧠 Extract structured shipment info with AI Agent 📋 Store metadata in Google Sheets ⚙️ Call Carbon Interface API with weight and distance 📥 Append estimated CO₂ emissions to the shipment row What do I need to get started? You’ll need: A Gmail account to receive shipment confirmation emails A Google Sheet to track shipment data and CO₂ A free Carbon Interface API key OpenAI access for using the AI Agent parser A few sample emails from your logistics provider to test Next Steps 🗒️ Use the sticky notes in the n8n canvas to: Add your Gmail and Carbon Interface credentials Try with a sample shipment confirmation email Check your Google Sheet to verify emissions and timestamps This template was built using n8n v1.93.0 Submitted: June 7, 2025
by Adnan
This workflow allows users to generate beautifully stylized 3D-rendered food emoji icons based on a simple text prompt. It combines user input, structured visual design generation, and image rendering using OpenAI’s GPT models. ✨ What It Does Collects user input via a form: e.g. "green apple" Generates a structured JSON specification describing the emoji’s form, lighting, texture, and color scheme Uses AI to render an image based on that spec—styled like a high-quality emoji icon with a transparent background 🧠 Use Case This template is ideal for: Designers or creators needing icon ideas or drafts for food items Developers building emoji packs or digital stickers Inspiration for AI-assisted product illustration or branding 💡 Why It's Useful Instead of prompting a model directly with vague terms, this flow creates a structured visual spec tailored to food items. The final emoji-style icon is polished, modern, and downloadable. ✅ Requirements To get started with this workflow, follow these steps: 🔑 Configure Credentials: Set up your API credentials for OpenAI and Google Drive 💳 Add OpoenAI Credit: Make sure to add credit to your OpenAI account, verify your organization (required for generating images) 📊 Connect Google Drive: Authenticate your Google Drive account ⚙️ (Optional) Customize Prompts: Adjust the prompts within the workflow to better suit your specific needs Note: Each image generation will cost you about $0.17
by mariskarthick
Reduce human delays between malware detection and remediation in MSSP/SOC environments. This workflow automates full endpoint antivirus scanning immediately after high-severity endpoint infection wazuh alerts, closing the gap between alerting and action. Why Use This Workflow? Malware alerts are only effective if acted upon swiftly. Manual follow-ups are slow or often missed, letting threats persist. Automates detection, triage, scan initiation, and notification—all within one minute of alerting. Ensures consistent, auditable actions across endpoints running Linux or Windows. 🔑 Key Features Listens for high-severity Wazuh AV infection alerts (e.g., rule 52502). Uses GPT-4 for AI-powered alert summaries to speed triage and decision making. Extracts exact infected file paths using AI and regex for targeted scanning. Runs ClamAV/defender scans directly on endpoints via SSH with least-privilege credentials. Sends real-time scan results and remediation updates through Telegram, Slack, or email. Runs locally with limited permissions—no need for elevated Wazuh manager access. 🎯 Impact Eliminates manual lag—scans start automatically and immediately. Standardizes response playbooks for reliable, repeatable remediation. Reduces threat dwell time, minimizing risk exposure. Provides full event-to-remediation visibility via logs and notifications. 🚀 Get Started Configure Wazuh Manager to forward AV alerts to this n8n webhook. Import this workflow JSON into your n8n instance. Set up required credentials: OpenAI API, SSH access for ClamAV scanning, notification channels (Telegram/Slack/email). Activate the workflow and monitor alerts triggering automated scans and reports. 📂 Enjoy customizing Swap ClamAV with your preferred antivirus commands (e.g., Defender) as needed. Integrate with your existing communication or ticketing systems. Extend or adapt for multi-endpoint orchestration or other alert rules. Created by Mariskarthick M Senior Security Analyst | Detection Engineer | Threat Hunter | Open-Source Enthusiast
by Omer Fayyaz
An intelligent AI-powered agent that automatically browses publication websites, analyzes page content with natural language understanding, and identifies the latest downloadable reports, research papers, and data files across multiple sources using advanced structured output parsing. What Makes This Different: AI-Powered Content Analysis** - Uses advanced language models (GPT-4/GPT-5.1) to understand page context and identify downloadable reports, even when links aren't explicitly labeled, handling complex page layouts and dynamic content Structured Output Parsing** - Enforces JSON schema validation ensuring consistent data extraction with required fields (title, link, file_type, description), eliminating parsing errors and data inconsistencies HTML to Markdown Conversion** - Converts raw HTML to clean Markdown before AI processing, removing noise and improving AI comprehension of page structure and content hierarchy Intelligent Link Detection** - AI agent identifies direct download URLs, converts relative links to absolute URLs, and prioritizes the most recent reports based on publication dates and page positioning Comprehensive Validation** - Multi-layer validation checks link format, file type detection, and report relevance before saving, ensuring only valid, downloadable reports enter your library Flexible Source Management** - Reads publication sources from Google Sheets, enabling easy addition/removal of sources without workflow modification, with support for categories and custom metadata Key Benefits of AI-Powered Report Discovery: Automated Discovery** - Eliminates manual browsing and searching across multiple publication sites, saving hours of research time while ensuring you never miss new reports Context-Aware Extraction** - AI understands page context, distinguishing between actual reports and navigation links, category pages, or promotional content Prioritized Results** - Automatically selects the most recent and relevant report from each source, focusing on quality over quantity Structured Data Output** - All discovered reports are saved with consistent metadata (title, link, file type, description, source), making them easy to search, filter, and integrate with other systems Error Resilience** - Handles missing reports gracefully, logging when no reports are found without failing the entire workflow, ensuring continuous operation Integration Ready** - Can be called by other workflows (e.g., PDF downloader), enabling end-to-end automation from discovery to storage Who's it for This template is designed for researchers, market analysts, competitive intelligence teams, academic institutions, industry monitoring services, and anyone who needs to systematically discover and track downloadable reports from multiple publication sources. It's perfect for organizations that need to monitor industry publications, track competitor research, discover new market reports, build research libraries, or stay updated on latest publications without manually visiting dozens of websites daily. How it works / What it does This workflow creates an AI-powered report discovery system that reads publication source URLs from Google Sheets, fetches their pages, uses AI to analyze content, and extracts information about downloadable reports. The system: Reads Active Sources - Fetches publication URLs and metadata from Google Sheets "Report Sources" sheet, processing each source in sequence Loops Through Sources - Processes sources one at a time using Split in Batches, ensuring proper error isolation and preventing batch failures Fetches Publication Pages - Downloads HTML content from each source URL with proper browser headers (User-Agent, Accept, Accept-Language) to avoid blocking Converts HTML to Markdown - Transforms raw HTML into clean Markdown format, removing styling, scripts, and navigation elements to improve AI comprehension AI Analysis - LangChain agent analyzes the Markdown content using GPT-4/GPT-5.1, identifying downloadable reports based on context, link patterns, and content structure Structured Output Parsing - Enforces JSON schema validation, ensuring the AI returns data in the exact format: source, title, link, file_type, description Validates & Normalizes Output - Validates extracted links are absolute URLs, checks file type indicators, determines report validity, and normalizes all fields Routes by Validity - IF node routes valid reports to save operation, invalid/missing reports to logging Saves Discovered Reports - Appends valid reports to Google Sheets "Discovered Reports" sheet with metadata, source URL, category, and discovery timestamp Logs No Report Found - Records sources where no valid reports were found in "Discovery Log" sheet for monitoring and troubleshooting Tracks Completion - Generates completion summary with number of sources checked and processing timestamp Key Innovation: AI-Powered Context Understanding - Unlike traditional web scrapers that rely on fixed CSS selectors or regex patterns, this workflow uses AI to understand page context and semantics. The AI can identify reports even when they're embedded in complex layouts, use non-standard naming, or require understanding of surrounding text to determine relevance. This makes it adaptable to any website structure without manual configuration. How to set up 1. Prepare Google Sheets Create a Google Sheet with three tabs: "Report Sources", "Discovered Reports", and "Discovery Log" In "Report Sources" sheet, create columns: Source_Name, Source_URL, Category (optional) Add publication URLs in the Source_URL column (e.g., "https://example.com/research" or "https://publisher.com/reports") Add descriptive names in Source_Name column for easy identification Optionally add Category values (e.g., "Market Research", "Industry Reports", "Academic Papers") The "Discovered Reports" sheet will be automatically populated with columns: source, title, link, fileType, description, sourceUrl, category, discoveredAt, status, isValid The "Discovery Log" sheet will record sources where no reports were found Verify your Google Sheets credentials are set up in n8n (OAuth2 recommended) 2. Configure Google Sheets Nodes Open the "Read Active Sources" node and select your spreadsheet from the document dropdown Set sheet name to "Report Sources" Configure the "Save Discovered Report" node: select same spreadsheet, set sheet name to "Discovered Reports", operation should be "Append or Update" Configure the "Log No Report Found" node: same spreadsheet, "Discovery Log" sheet, operation "Append or Update" Test connection by running the "Read Active Sources" node manually to verify it can access your sheet 3. Set Up OpenAI Credentials Open the "OpenAI GPT-5.1" node (or configure the model you want to use) Connect your OpenAI API credentials (API key required) The workflow uses GPT-5.1 by default, but you can change to GPT-4, GPT-4 Turbo, or other models Temperature is set to 0.1 for consistent, deterministic output Verify API key has sufficient credits and access to the selected model For cost optimization, GPT-4 Turbo is recommended for similar results at lower cost 4. Configure AI Agent & Output Parser The "AI Report Discovery Agent" node contains a detailed system prompt that instructs the AI on what to look for The prompt is pre-configured but can be customized for your specific needs (e.g., prioritize certain file types, look for specific keywords) The "Structured Output Parser" enforces the JSON schema - verify the schema matches your needs: { "source": "Publisher Name", "title": "Report Title", "link": "https://example.com/report.pdf", "file_type": "pdf", "description": "Brief description" } The parser ensures the AI always returns valid JSON with all required fields Test the AI agent by manually running with a sample source URL to verify it correctly identifies reports 5. Customize Discovery Rules (Optional) The AI agent's system prompt can be modified in the "AI Report Discovery Agent" node Current rules prioritize: downloadable files (PDF, Excel, Word, PowerPoint), most recent publications, direct download URLs To customize: Edit the system message to add specific keywords, file types, or discovery patterns Example customization: Add industry-specific terms or prioritize reports with certain keywords in titles The validation code in "Validate & Normalize Output" can be adjusted to change what's considered "valid" Test with your specific sources to ensure discovery rules work as expected 6. Set Up Scheduling & Test The workflow includes Manual Trigger (for testing), Schedule Trigger (runs daily), and Execute Workflow Trigger (for calling from other workflows) To customize schedule: Open "Schedule (Daily)" node and adjust interval (e.g., twice daily, weekly) For initial testing: Use Manual Trigger, add 2-3 test publication URLs to your "Report Sources" sheet Verify execution: Check that pages are fetched, AI analysis completes, and reports are saved to "Discovered Reports" Monitor execution logs: Check for API errors, timeout issues, or parsing failures Review Discovery Log: Verify sources with no reports are properly logged Common issues: OpenAI API rate limits (add delays if processing many sources), invalid URLs (check source URLs), timeout errors (increase timeout for slow-loading pages), AI not finding reports (may need to adjust system prompt for specific site structures) Requirements OpenAI API Key** - Active OpenAI account with API access and sufficient credits for GPT-4/GPT-5.1 model usage (API key configured in n8n credentials) Google Sheets Account** - Active Google account with OAuth2 credentials configured in n8n for reading and writing spreadsheet data Source Spreadsheet** - Google Sheet with "Report Sources", "Discovered Reports", and "Discovery Log" tabs, properly formatted with required columns Valid Publication URLs** - Direct links to publication pages that contain downloadable reports (not direct PDF links - the workflow discovers those) n8n Instance** - Self-hosted or cloud n8n instance with access to external websites (HTTP Request node needs internet connectivity) and LangChain nodes enabled
by PiAPI
What this workflow does? This workflow primarily uses the GPT-4o API from PiAPI and automatically creates front/side/top views of 3D models from commands. Who is this for? 3D Designers: Quickly generate standardized orthographic views for design review E-commerce Operators: Create multi-angle product display images 3D Modeling Beginners: Instantly produce basic reference views Step-by-step Instruction Fill in X-API-Key of your PiAPI account and the image prompt based on your inspiration. Click Test workflow. Get the image url in the final node. OutPut