by vinci-king-01
Multi-Source RAG System with GPT-4 Turbo, News & Academic Papers Integration This workflow provides an enterprise-grade RAG (Retrieval-Augmented Generation) system that intelligently searches multiple sources and generates AI-powered responses using GPT-4 Turbo. How it works This workflow provides an enterprise-grade RAG (Retrieval-Augmented Generation) system that intelligently searches multiple sources and generates AI-powered responses using GPT-4 Turbo. Key Steps Form Input - Collects user queries with customizable search scope, response style, and language preferences Intelligent Search - Routes queries to appropriate sources (web, academic papers, news, internal documents) Data Aggregation - Unifies and processes information from multiple sources with quality scoring AI Processing - Uses GPT-4 Turbo to generate context-aware, source-grounded responses Response Enhancement - Formats outputs in various styles (comprehensive, concise, technical, etc.) Multi-Channel Delivery - Delivers results via webhook, email, Slack, and optional PDF generation Data Sources & AI Models Search Sources Web Search**: Google, Bing, DuckDuckGo integration Academic Papers**: arXiv, PubMed, Google Scholar News Articles**: News API, RSS feeds, real-time news Technical Documentation**: GitHub, Stack Overflow, documentation sites Internal Knowledge**: Google Drive, Confluence, Notion integration AI Models GPT-4 Turbo**: Primary language model for response generation Embedding Models**: For semantic search and similarity matching Custom Prompts**: Specialized prompts for different response styles Set up steps Setup time: 15-20 minutes Configure API credentials - Set up OpenAI API, ScrapeGraphAI, Google Drive, and other service credentials Set up search sources - Configure academic databases, news APIs, and internal knowledge sources Connect analytics - Link Google Sheets for usage tracking and performance monitoring Configure notifications - Set up Slack channels and email templates for automated alerts Test the workflow - Run sample queries to verify all components are working correctly Keep detailed configuration notes in sticky notes inside your workflow
by Alexandru Burca
Daily News Digest Video Generator for YouTube Shorts Instalations Instructions Youtube Instalation Instructions Overview This workflow automatically creates and publishes daily news digest videos from WordPress articles to YouTube. It runs every evening at 7 PM, compiling the day's top stories from a news portal into a professionally formatted vertical video (1080x1920px) optimized for social media platforms like YouTube Shorts. What It Does 1. 🕐 Scheduled Trigger Runs automatically every day at 19:00 (7 PM) 2. 📰 Fetches Today's Articles Retrieves all published WordPress posts from the current day 3. ✅ Validates Content Ensures there are at least 3 articles before proceeding 4. 🎬 Video Detection Scans article content HTML for embedded videos Extracts MP4 URLs from WordPress video players Parses wp-playlist-script JSON data Falls back to ` and <source>` tag detection 5. 🧹 Data Processing Extracts** article titles, links, and featured media IDs Decodes HTML entities**: Converts – to –, " to ", etc. Fetches featured images** from WordPress Media API Assigns default images** for articles without featured media Calculates reading time** per article (3-7 seconds based on word count) Cleans text**: Removes HTML tags and normalizes whitespace 6. 🎥 Video Generation (via Shotstack API) Intro Slide (3 seconds) Black background Large logo (centered) Title on center Current date in DD-MM-YYYY format News Slides (3-7 seconds each) Each article is displayed with: Background**: Video (if available) or featured image, cropped to fit Dark overlay**: 40% opacity black layer for text readability Article headline**: Large white text at top Small logo**: Top-right corner Pagination counter**: Bottom-right white badge (e.g., "1 / 22") CTA button**: Centered CTA Background music**: Subtle looped audio track Transitions**: Smooth fade in/out between slides Outro Slide (3 seconds) Identical to intro slide Provides clean ending to the video 7. ⏳ Processing Wait Waits 30 seconds for Shotstack to render the video Polls Shotstack API to verify video completion 8. 📥 Download Video Retrieves the finished MP4 file from Shotstack Downloads video data for YouTube upload 9. 📤 YouTube Upload Automatically uploads to YouTube with: Title**: "Daily Digest - [Day] [Weekday], [Year]" Description**: Same as title Category**: News & Politics Made for kids**: Yes Tags**: dailydigest ✨ Key Features Intelligent Content Handling ✅ Automatic video/image detection and intelligent media selection ✅ Dynamic reading time calculation for optimal viewer engagement ✅ HTML entity cleaning for proper text display (WordPress compatibility) ✅ Fallback default images for articles without media ✅ Video background support with automatic muting Professional Video Production ✅ Vertical format optimized for mobile viewing (1080x1920px) ✅ Professional branding with logos and consistent styling ✅ Smooth fade transitions between slides ✅ Background music with looping support ✅ Dynamic pagination counters ✅ Call-to-action buttons for engagement Customization ✅ Centralized variables for easy branding updates ✅ Configurable logos, colors, and text ✅ Adjustable reading time calculation ✅ Flexible date formatting ✅ Customizable audio track 🎯 Use Cases Perfect for: 📰 News websites wanting to repurpose daily articles 📱 Media outlets creating social media content 🎥 Content creators automating video production 🔄 Publishers maximizing content distribution 📊 Marketing teams driving traffic from social platforms 🔧 Customization Options Easy Changes Update logos by changing logo_big and logo_small URLs Modify branding colors via button_bg_color variable Adjust button text with button_text variable Change video title with daily_digest_text variable Update background music by replacing audio URL Advanced Customization Adjust reading time formula in calculateReadingTime() function Modify date format in getRomanianDate() function Change video dimensions (currently 1080x1920) Update font family and sizes Adjust overlay opacity and colors Modify transition effects 📋 Prerequisites Required Credentials WordPress API - Access to stire.md WordPress site Shotstack API - API key for video rendering (Stage environment) YouTube OAuth2 - Authenticated YouTube account for uploads
by vinci-king-01
Enterprise Knowledge Search with GPT-4 Turbo, Google Drive & Academic APIs This workflow provides an enterprise-grade RAG (Retrieval-Augmented Generation) system that intelligently searches multiple sources and generates AI-powered responses using GPT-4 Turbo. How it works This workflow provides an enterprise-grade RAG (Retrieval-Augmented Generation) system that intelligently searches multiple sources and generates AI-powered responses using GPT-4 Turbo. Key Steps Form Input - Collects user queries with customizable search scope, response style, and language preferences Intelligent Search - Routes queries to appropriate sources (web, academic papers, news, internal documents) Data Aggregation - Unifies and processes information from multiple sources with quality scoring AI Processing - Uses GPT-4 Turbo to generate context-aware, source-grounded responses Response Enhancement - Formats outputs in various styles (comprehensive, concise, technical, etc.) Multi-Channel Delivery - Delivers results via webhook, email, Slack, and optional PDF generation Data Sources & AI Models Search Sources Web Search**: Google, Bing, DuckDuckGo integration Academic Papers**: arXiv, PubMed, Google Scholar via Crossref API News Articles**: News API, RSS feeds, real-time news Technical Documentation**: GitHub, Stack Overflow, documentation sites Internal Knowledge**: Google Drive, Confluence, Notion integration AI Models GPT-4 Turbo**: Primary language model for response generation Embedding Models**: For semantic search and similarity matching Custom Prompts**: Specialized prompts for different response styles Set up steps Setup time: 15-20 minutes Configure API credentials - Set up OpenAI API, News API, Google Drive, and other service credentials Set up search sources - Configure academic databases, news APIs, and internal knowledge sources Connect analytics - Link Google Sheets for usage tracking and performance monitoring Configure notifications - Set up Slack channels and email templates for automated alerts Test the workflow - Run sample queries to verify all components are working correctly Keep detailed configuration notes in sticky notes inside your workflow
by Jordan Hoyle
Description Automate the discovery and analysis of PDF files across a deeply nested OneDrive folder structure. This workflow recursively searches folders, filters for new or updated PDFs, extracts text, and uses a Mistral AI agent to generate a concise Executive Summary, Key Findings, and Structured Metadata (Date, Location, etc.), storing all insights into a n8n Data Table for easy access and further automation. Key Features & How It Works Scheduled Trigger & Recursive Folder Search: The workflow runs automatically (scheduled for 8 PM in this template) to monitor a specified main folder on OneDrive. It performs a deep, multi-level search (up to 8 layers) across subfolders to ensure no documents are missed. Smart Deduplication & Filtering: It checks new files against an internal n8n Data Table using the Compare Datasets node, ensuring only new or unique PDF files are processed, saving AI credits and processing time. A size check is also included, preventing attempts to process excessively large files. AI-Powered Document Intelligence (Mistral LLM): For each new PDF, the workflow extracts the text and passes it to a Mistral AI model for dual-stream analysis: Overview Agent: Generates an impartial, professional Executive Summary, a list of Key Findings & Data Points, and the document's Scope/Context. Document Information Agent: Extracts crucial metadata, including the single most relevant date, location (City/State/Country), and professional information (Name, Title, Organization). Structured Output and Archiving: AI outputs are meticulously validated and reformatted into a clean JSON object using Structured Output Parsers. The complete analysis, along with the original file name and path, is then logged as a new row in an n8n Data Table. Setup Notes OneDrive Folder: You must specify the exact name of your main folder in the 'Search for Main Folder' node. Data Table: Ensure your n8n Data Table exists with the required columns: Summary, Key_Findings, Scope, Date, Location, File_Name, and Path. Deep Folder Structure: The current configuration supports up to 8 levels of subfolders. If your files go deeper, you may need to add more "Get items in a folder" and "If" nodes. AI Customization: Review the AI agent prompts and the structured output schemas to customize the fields you want to extract or the summary style you require. Extend This Workflow The final output is organized data. You can easily extend this workflow to: Send daily/weekly digest emails with new summaries. Sync the extracted data to a Google Sheet, Airtable, or other database. Add a secondary AI agent to perform follow-up actions based on the "Key Findings."
by Mauricio Perera
📁 Analyze uploaded images, videos, audio, and documents with specialized tools — powered by a lightweight language-only agent. 🧭 What It Does This workflow enables multimodal file analysis using Google Gemini tools connected to a text-only LLM agent. Users can upload images, videos, audio files, or documents via a chat interface. The workflow will: Upload each file to Google Gemini and obtain an accessible URL. Dynamically generate contextual prompts based on the file(s) and user message. Allow the agent to invoke Gemini tools for specific media types as needed. Return a concise, helpful response based on the analysis. 🚀 Use Cases Customer support**: Let users upload screenshots, documents, or recordings and get helpful insights or summaries. Multimedia QA**: Review visual, audio, or video content for correctness or compliance. Educational agents**: Interpret content from PDFs, diagrams, or audio recordings on the fly. Low-cost multimodal assistants: Achieve multimodal functionality **without relying on large vision-language models. 🎯 Why This Architecture Matters Unlike end-to-end multimodal LLMs (like Gemini 1.5 or GPT-4o), this template: Uses a text-only LLM (Qwen 32B via Groq) for reasoning. Delegates media analysis to specialized Gemini tools. ✅ Advantages | Feature | Benefit | | ----------------------- | --------------------------------------------------------------------- | | 🧩 Modular | LLM + Tools are decoupled; can update them independently | | 💸 Cost-Efficient | No need to pay for full multimodal models; only use tools when needed | | 🔧 Tool-based Reasoning | Agent invokes tools on demand, just like OpenAI’s Toolformer setup | | ⚡ Fast | Groq LLMs offer ultra-fast responses with low latency | | 📚 Memory | Includes context buffer for multi-turn chats (15 messages) | 🧪 How It Works 🔹 Input via Chat Users submit a message and (optionally) files via the chatTrigger. 🔹 File Handling If no files: prompt is passed directly to the agent. If files are included: Files are split, uploaded to Gemini (to get public URLs). Metadata (name, type, URL) is collected and embedded into the prompt. 🔹 Prompt Construction A new chatInput is dynamically generated: User message Media: [array of file data] 🔹 Agent Reasoning The Langchain Agent receives: The enriched prompt File URLs Memory context (15 turns) Access to 4 Gemini tools: IMG: analyze image VIDEO: analyze video AUDIO: analyze audio DOCUMENT: analyze document The agent autonomously decides whether and how to use tools, then responds with concise output. 🧱 Nodes & Services | Category | Node / Tool | Purpose | | --------------- | ---------------------------- | ------------------------------------- | | Chat Input | chatTrigger | User interface with file support | | File Processing | splitOut, splitInBatches | Process each uploaded file | | Upload | googleGemini | Uploads each file to Gemini, gets URL | | Metadata | set, aggregate | Builds structured file info | | AI Agent | Langchain Agent | Receives context + file data | | Tools | googleGeminiTool | Analyze media with Gemini | | LLM | lmChatGroq (Qwen 32B) | Text reasoning, high-speed | | Memory | memoryBufferWindow | Maintains session context | ⚙️ Setup Instructions 1. 🔑 Required Credentials Groq API key** (for Qwen 32B model) Google Gemini API key** (Palm / Gemini 1.5 tools) 2. 🧩 Nodes That Need Setup Replace existing credentials on: Upload a file Each GeminiTool (IMG, VIDEO, AUDIO, DOCUMENT) lmChatGroq 3. ⚠️ File Size & Format Considerations Some Gemini tools have file size or format restrictions. You may add validation nodes before uploading if needed. 🛠️ Optional Improvements Add logging and error handling (e.g., for upload failures). Add MIME-type filtering to choose the right tool explicitly. Extend to include OCR or transcription services pre-analysis. Integrate with Slack, Telegram, or WhatsApp for chat delivery. 🧪 Example Use Case > "Hola, ¿qué dice este PDF?" Uploads a document → Agent routes it to Gemini DOCUMENT tool → Receives extracted content → LLM summarizes it in Spanish. 🧰 Tags multimodal, agent, langchain, groq, gemini, image analysis, audio analysis, document parsing, video analysis, file uploader, chat assistant, LLM tools, memory, AI tools 📂 Files This template is ready to use as-is in n8n. No external webhooks or integrations required.
by PDF Vector
Overview Healthcare organizations face significant challenges in digitizing and processing medical records while maintaining strict HIPAA compliance. This workflow provides a secure, automated solution for extracting clinical data from various medical documents including discharge summaries, lab reports, clinical notes, prescription records, and scanned medical images (JPG, PNG). What You Can Do Extract clinical data from medical documents while maintaining HIPAA compliance Process handwritten notes and scanned medical images with OCR Automatically identify and protect PHI (Protected Health Information) Generate structured data from various medical document formats Maintain audit trails for regulatory compliance Who It's For Healthcare providers, medical billing companies, clinical research organizations, health information exchanges, and medical practice administrators who need to digitize and extract data from medical records while maintaining HIPAA compliance. The Problem It Solves Manual medical record processing is time-consuming, error-prone, and creates compliance risks. Healthcare organizations struggle to extract structured data from handwritten notes, scanned documents, and various medical forms while protecting PHI. This template automates the extraction process while maintaining the highest security standards for Protected Health Information. Setup Instructions: Configure Google Drive credentials with proper medical record access controls Install the PDF Vector community node from the n8n marketplace Configure PDF Vector API credentials with HIPAA-compliant settings Set up secure database storage with encryption at rest Define PHI handling rules and extraction parameters Configure audit logging for regulatory compliance Set up integration with your Electronic Health Record (EHR) system Key Features: Secure retrieval of medical documents from Google Drive HIPAA-compliant processing with automatic PHI masking OCR support for handwritten notes and scanned medical images Automatic extraction of diagnoses with ICD-10 code validation Medication list processing with dosage and frequency information Lab results extraction with reference ranges and flagging Vital signs capture and normalization Complete audit trail for regulatory compliance Integration-ready format for EHR systems Customization Options: Define institution-specific medical terminology and abbreviations Configure automated alerts for critical lab values or abnormal results Set up custom extraction fields for specialized medical forms Implement medication interaction warnings and contraindication checks Add support for multiple languages and international medical coding systems Configure integration with specific EHR platforms (Epic, Cerner, etc.) Set up automated quality assurance checks and validation rules Implementation Details: The workflow uses advanced AI with medical domain knowledge to understand clinical terminology and extract relevant information while automatically identifying and protecting PHI. It processes various document formats including handwritten prescriptions, lab reports, discharge summaries, and clinical notes. The system maintains strict security protocols with encryption at rest and in transit, ensuring full HIPAA compliance throughout the processing pipeline. Note: This workflow uses the PDF Vector community node. Make sure to install it from the n8n community nodes collection before using this template.
by Davide
This workflow automates the process of creating short videos from multiple image references (up to 7 images). It uses "Vidu Reference to Video" model, a video generation API to transform a user-provided prompt and image set into a consistent, AI-generated video. This workflow automates the process of generating AI-powered videos from a set of reference images and then uploading them to TikTok and Youtube. The process is initiated via a user-friendly web form. Advantages ✅ Consistent Video Creation: Uses multiple reference images to maintain subject consistency across frames. ✅ Easy Input: Just a simple form with prompt + image URLs. ✅ Automation: No manual waiting—workflow checks status until video is ready. ✅ SEO Optimization: Automatically generates a catchy, optimized YouTube title using AI. ✅ Multi-Platform Publishing: Uploads directly to Google Drive, YouTube, and TikTok in one flow. ✅ Time Saving: Removes repetitive tasks of video generation, download, and manual uploading. ✅ Scalable: Can run periodically or on-demand, perfect for content creators and marketing teams. ✅ UGC & Social Media Ready: Designed for creating viral short videos optimized for platforms like TikTok and YouTube Shorts. How It Works Form Trigger: A user submits a web form with two key pieces of information: a text Prompt describing the desired video and a list of Reference images (URLs separated by commas or new lines). Data Processing: The workflow processes the submitted image URLs, converting them from a text string into a proper array format for the AI API. AI Video Generation: The processed data (prompt and image array) is sent to the Fal.ai VIDU API endpoint (reference-to-video) to start the video generation job. This node returns a request_id. Status Polling: The workflow enters a loop where it periodically checks the status of the generation job using the request_id. It waits for 60 seconds and then checks if the status is "COMPLETED". If not, it waits and checks again. Result Retrieval: Once the video is ready, the workflow fetches the URL of the generated video file. Title Generation: Simultaneously, the original user prompt is sent to an AI model (GPT-4o-mini via OpenRouter) to generate an optimized, engaging title for the social media post. Upload & Distribution: The video file is downloaded from the generated URL. A copy is saved to a specified Google Drive folder for storage. The video, along with the AI-generated title, is automatically uploaded to YouTube and TikTok via the Upload-Post.com API service. Set Up Steps This workflow requires configuration and API keys from three external services to function correctly. Step 1: Configure Fal.ai for Video Generation Create an account and obtain your API key. In the "Create Video" HTTP node, edit the "Header Auth" credentials. Set the following values: Name: Authorization Value: Key YOUR_FAL_API_KEY (replace YOUR_FAL_API_KEY with your actual key) Step 2: Configure Upload-Post.com for Social Media Uploads Get an API key from your Upload-Post Manage Api Keys dashboard (10 free uploads per month). In both the "HTTP Request" (YouTube) and "Upload on TikTok" nodes, edit their "Header Auth" credentials. Set the following values: Name: Authorization Value: Apikey YOUR_UPLOAD_POST_API_KEY (replace YOUR_UPLOAD_POST_API_KEY with your actual key) Crucial: In the body parameters of both upload nodes, find the user field and replace YOUR_USERNAME with the exact name of the social media profile you configured on Upload-Post.com (e.g., my_youtube_channel). Step 3: Configure Google Drive (Optional Storage) The "Upload Video" node is pre-configured to save the video to a Google Drive folder named "Fal.run". Ensure your Google Drive credentials in n8n are valid and that you have access to this folder, or change the folderId parameter to your desired destination. Step 4: Configure AI for Title Generation The "Generate title" node uses OpenAI to access the gpt-5-mini model.. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Daniel
Harness OpenAI's Sora 2 for instant video creation from text or images using fal.ai's API—powered by GPT-5 for refined prompts that ensure cinematic quality. This template processes form submissions, intelligently routes to text-to-video (with mandatory prompt enhancement) or image-to-video modes, and polls for completion before redirecting to your generated clip. 📋 What This Template Does Users submit prompts, aspect ratios (9:16 or 16:9), models (sora-2 or pro), durations (4s, 8s, or 12s), and optional images via a web form. For text-to-video, GPT-5 automatically refines the prompt for optimal Sora 2 results; image mode uses the raw input. It calls one of four fal.ai endpoints (text-to-video, text-to-video/pro, image-to-video, image-to-video/pro), then loops every 60s to check status until the video is ready. Handles dual modes: Text (with GPT-5 enhancement) or image-seeded generation Supports pro upgrades for higher fidelity and longer clips Auto-uploads images to a temp host and polls asynchronously for hands-free results Redirects directly to the final video URL on completion 🔧 Prerequisites n8n instance with HTTP Request and LangChain nodes enabled fal.ai account for Sora 2 API access OpenAI account for GPT-5 prompt refinement 🔑 Required Credentials fal.ai API Setup Sign up at fal.ai and navigate to Dashboard → API Keys Generate a new key with "sora-2" permissions (full access recommended) In n8n, create "Header Auth" credential: Name it "fal.ai", set Header Name to "Authorization", Value to "Key [Your API Key]" OpenAI API Setup Log in at platform.openai.com → API Keys (top-right profile menu) Click "Create new secret key" and copy it (store securely) In n8n, add "OpenAI API" credential: Paste key, select GPT-5 model in the LLM node ⚙️ Configuration Steps Import the workflow JSON into your n8n instance via Settings → Import from File Assign fal.ai and OpenAI credentials to the relevant HTTP Request and LLM nodes Activate the workflow—the form URL auto-generates in the trigger node Test by submitting a sample prompt (e.g., "A cat chasing a laser"); monitor executions for video output Adjust polling wait (60s node) for longer generations if needed 🎯 Use Cases Social Media Teams**: Generate 9:16 vertical Reels from text ideas, like quick product animations enhanced by GPT-5 for professional polish Content Marketers**: Animate uploaded images into 8s promo clips, e.g., turning a static ad graphic into a dynamic story for email campaigns Educators and Trainers**: Create 4s explainer videos from outlines, such as historical reenactments, using pro mode for detailed visuals App Developers**: Embed as a backend service to process user prompts into Sora 2 videos on-demand for creative tools ⚠️ Troubleshooting API quota exceeded**: Check fal.ai dashboard for usage limits; upgrade to pro tier or extend polling waits Prompt refinement fails**: Ensure GPT-5 credential is set and output matches JSON schema—test LLM node independently Image upload errors**: Confirm file is JPG/PNG under 10MB; verify tmpfiles.org endpoint with a manual curl test Endless polling loop**: Add an IF node after 10 checks to timeout; increase wait to 120s for 12s pro generations
by Khairul Muhtadin
The Prompt converter workflow tackles the challenge of turning your natural language video ideas into perfectly formatted JSON prompts tailored for Veo 3 video generation. By leveraging Langchain AI nodes and Google Gemini, this workflow automates and refines your input to help you create high-quality videos faster and with more precision—think of it as your personal video prompt translator that speaks fluent cinematic! 💡 Why Use Prompt Converter? Save time: Automate converting complex video prompts into structured JSON, cutting manual formatting headaches and boosting productivity. Avoid guesswork: Eliminate unclear video prompt details by generating detailed, cinematic descriptions that align perfectly with Veo 3 specs. Improve output quality: Optimize every parameter for Veo 3's video generation model to get realistic and stunning results every time. Gain a creative edge: Turn vague ideas into vivid video concepts with AI-powered enhancement—your video project's secret weapon. ⚡ Perfect For Video creators: Content developers wanting quick, precise video prompt formatting without coding hassles. AI enthusiasts: Developers and hobbyists exploring Langchain and Google Gemini for media generation. Marketing teams: Professionals creating video ads or visuals who need consistent prompt structuring that saves time. 🔧 How It Works ⏱ Trigger: User submits a free text prompt via message or webhook. 📎 Process: The text goes through an AI model that understands and reworks it into detailed JSON parameters tailored for Veo 3. 🤖 Smart Logic: Langchain nodes parse and optimize the prompt with cinematic details, set reasonable defaults, and structure the data precisely. 💌 Output: The refined JSON prompt is sent to Google Gemini for video generation with optimized settings. 🔐 Quick Setup Import the JSON file to your n8n instances Add credentials: Azure OpenAI, Gemini API, OpenRouter API Customize: Adjust prompt templates or default parameters in the Prompt converter node Test: Run your workflow with sample text prompts to see videos come to life 🧩 You'll Need Active n8n instances Azure OpenAI API Gemini API Key OpenRouter API (alternative AI option) 🛠️ Level Up Ideas Add integration with video hosting platforms to auto-upload generated videos 🧠 Nodes Used Prompt Input** (Chat Trigger) OpenAI** (Azure OpenAI GPT model) Alternative** (OpenRouter API) Prompt converter** (Langchain chain LLM for JSON conversion) JSON parser** (structured output extraction) Generate a video** (Google Gemini video generation) Made by: Khaisa Studio Tags: video generation, AI, Langchain, automation, Google Gemini Category: Video Production Need custom work? Contact me
by furuidoreandoro
Automated TikTok Repurposing & Video Generation Workflow Who’s it for This workflow is designed for content creators, social media managers, and marketers—specifically those in the career, recruitment, or "job change" (転職/就職) niches. It is ideal for anyone looking to automate the process of finding trending short-form content concepts and converting them into fresh AI-generated videos. How it works / What it does This workflow automates the pipeline from content research to video creation: Scrape Data: It triggers an Apify actor (clockworks/tiktok-scraper) to search and scrape TikTok videos related to "Job Change" (転職) and "Employment" (就職). Store Raw Data: It saves the scraped TikTok metadata (text, stats, author info) into a Google Sheet. AI Analysis & Prompting: An AI Agent (via OpenRouter) analyzes the scraped video content and creates a detailed prompt for a new video (concept, visual cues, aspect ratio). Log Prompts: The generated prompt is saved to a separate tab in the Google Sheet. Video Generation: The prompt is sent to Fal AI (Veo3 model) to generate a new 8-second, vertical (9:16) video with audio. Wait & Retrieve: The workflow waits for the generation to complete, then retrieves the video file. Cloud Storage: Finally, it uploads the generated video file to a specific Google Drive folder. How to set up Credentials: Configure the following credentials in n8n: Apify API: (Currently passed via URL query params in the workflow, recommended to switch to Header Auth). Google Sheets OAuth2: Connect your Google account. OpenRouter API: For the AI Agent. Fal AI (Header Auth): For the video generation API. Google Drive OAuth2: For uploading the final video. Google Sheets: Create a spreadsheet. Note the documentId and update the Google Sheets nodes. Ensure you have the necessary Sheet names (e.g., "シート1" for raw data, "生成済み" for prompts) and columns mapped. Google Drive: Create a destination folder. Update the Upload file node with the correct folderId. Apify: Update the token in the HTTP Request and HTTP Request1 URLs with your own Apify API token. Requirements n8n Version:** 1.x or higher (Workflow uses version 4.3 nodes). Apify Account:** With access to clockworks/tiktok-scraper and sufficient credits. Fal.ai Account:** With credits for the fal-ai/veo3 model. OpenRouter Account:** With credits for the selected LLM. Google Workspace:** Access to Drive and Sheets. How to customize the workflow Change the Niche:* Update the searchQueries JSON body in the first *HTTP Request** node (e.g., change "転職" to "Cooking" or "Fitness"). Adjust AI Logic:* Modify the *AI Agent** system prompt to change the style, tone, or structure of the video prompts it generates. Video Settings:* In the *Fal Submit** node, adjust bodyParameters to change the duration (e.g., 5s), aspect ratio (e.g., 16:9), or disable audio. Scale:* Increase the amount in the *Limit** node to process more than one video per execution.
by gotoHuman
Collaborate with an AI Agent on a joint document, e.g. for creating your content marketing strategy, a sales plan, project status updates, or market analysis. The AI Agent generates markdown text that you can review and edit it in gotoHuman, and only then is the existing Google Doc updated. In this example we use AI to update our company's content strategy for the next quarter. How It Works The AI Agent has access to other documents that provide enough context to write the content strategy. We ask it to generate the text in markdown format. To ensure our strategy document is not changed without our approval, we request a human review using gotoHuman. There the markdown content can be edited and properly previewed. Our workflow resumes once the review is completed. We check if the content was approved and then write the (potentially edited) markdown to our Google Docs file via the Google Drive node. How to set up Most importantly, install the verified gotoHuman node before importing this template! (Just add the node to a blank canvas before importing. Works with n8n cloud and self-hosted) Set up your credentials for gotoHuman, OpenAI, and Google Docs/Drive In gotoHuman, select and create the pre-built review template "Strategy agent" or import the ID: F4sbcPEpyhNKBKbG9C1d Select this template in the gotoHuman node Requirements You need accounts for gotoHuman (human supervision) OpenAI (Doc writing) Google Docs/Drive How to customize Let the workflow run on a schedule, or create and connect a manual trigger in gotoHuman that lets you capture additional human input to feed your agent Provide the agent with more context to write the content strategy Use the gotoHuman response (or a Google Drive file change trigger) to run additional AI agents that can execute on the new strategy
by InfyOm Technologies
✅ What problem does this workflow solve? Sending a plain PDF resume doesn’t stand out anymore. This workflow allows candidates to convert their resume and photo into a personalized video resume. Recruiters get a more engaging first impression, while candidates showcase their profile in a modern, impactful way. ⚙️ What does this workflow do? Presents a form for uploading: 📄 Resume (PDF) 🖼 Photo (headshot) Extracts key details from the resume (education, experience, skills). Detects gender from the photo to choose a suitable voice/avatar. Generates a script (spoken resume summary) based on the extracted information. Uploads the photo to HeyGen to create an avatar. Requests video generation on HeyGen: Uses the avatar photo Uses gender-specific settings Uses the generated script as narration Monitors video generation status until completion. Stores the final video URL in a Google Sheet for easy access and tracking. 🔧 Setup Instructions Google Services Connect Google Sheets to n8n to store records with: Candidate name Resume link Video link HeyGen Setup Get an API key from HeyGen. Configure: Avatar upload endpoint (image upload) Video generation endpoint (image ID + script) Form Setup Use the n8n Form Trigger to allow candidates to upload: Resume (PDF) Photo (JPEG/PNG) 🧠 How it Works – Step-by-Step 1. Candidate Submission A candidate fills out a form and uploads: Resume (PDF) Photo 2. Extract Resume Data The resume PDF is processed using OCR/AI to extract: Name Experience Skills Education highlights 3. Gender Detection The uploaded photo is analyzed to detect gender (used for voice/avatar selection). 4. Script Generation Based on the extracted resume info, a concise, natural script is generated automatically. 5. Avatar Upload & Video Creation The photo is uploaded to HeyGen to create a custom avatar. A video generation request is made using: The script The avatar (image ID) A matching voice for the detected gender 6. Video Status Monitoring The workflow polls HeyGen’s API until the video is ready. 7. Save Final Video URL Once complete, the video link is added to a Google Sheet alongside the candidate’s details. 👤 Who can use this? This workflow is ideal for: 🧑🎓 Students and job seekers looking to stand out 🧑💼 Recruitment agencies offering modern resume services 🏢 HR teams wanting engaging candidate submissions 🎥 Portfolio builders for professionals 🚀 Impact Instead of a static PDF, you can now send a dynamic video resume that captures attention, adds personality, and makes a lasting impression.