by Davide
💬🗂️🤖 This workflow automates the translation of Google Slides presentations from any languages, while preserving the original formatting and slide structure. It leverages Google APIs, AI translation (Gemini/PaLM), and modular execution for high flexibility and accuracy. DISCLAIMER: texts are split by Google Slides APIs into small blocks, so the translation will not always be contextualized. Key Benefits ⚡ Time-Saving**: Automates a typically manual and error-prone task of translating slides. 🌍 AI-Powered Accuracy**: Uses Google Gemini to provide context-aware translations while respecting defined rules. 🔒 Safe & Non-Destructive**: The original presentation is never modified — a new copy is always created. 🎯 Precision**: Skips irrelevant text (e.g., emails, URLs, names) to avoid mistranslation. 🔁 Modular & Scalable**: Uses subworkflows and batching, ideal for presentations with many slides. 🎨 Layout Preservation**: Keeps the original design and formatting intact. How it Works Initialization: The workflow starts with a manual trigger ("When clicking ‘Execute workflow’"). Set the language to translate (IMPORTANT format ISO-639) It duplicates a specified Google Slides presentation ("Duplicate presentation") to create a new copy for translation, preserving the original. Slide Processing: The workflow retrieves slides from the copied presentation ("Get slides from a presentation") and processes them in batches ("Loop Over Items"). For each slide, text content is extracted ("Extract Text") using a custom JavaScript snippet, which identifies and collects text elements while retaining the slide's objectId. Translation: The extracted texts are passed to a LangChain agent ("Translation expert"), which translates the content from Italian to English. The agent follows strict guidelines (e.g., skipping URLs, brand names, etc.). The translated text is sent to the "Translate Google Slides" node, which replaces the original text in the presentation using the slide's objectId for targeting. Execution Flow: The workflow includes delays ("Wait 10 sec" and "Wait 3 sec") to manage API rate limits and ensure smooth execution. The process repeats for each batch of slides until all content is translated. Set Up Steps Prerequisites: Ensure access to the source Google Slides presentation (specified by fileId in "Duplicate presentation"). Set up Google OAuth2 credentials for Google Drive and Slides (nodes reference credentials like "Google Slides account"). Configure the Google Gemini (PaLM) API credentials for the translation agent. Configuration: Update the fileId in the "Duplicate presentation" node to point to your source presentation. Adjust the translation guidelines in the "Translation expert" node if needed (e.g., language pairs or exclusion rules). Modify batch sizes or wait times (e.g., "Wait 10 sec") based on API constraints. Execution: Run the workflow manually or trigger it via the "Execute Workflow" node from another workflow. Monitor progress in n8n’s execution log, as each slide is processed and translated sequentially. Output: The translated presentation is saved as a new file in Google Drive, with the filename including a timestamp (e.g., NAME_PRESENTATION_{lang}_{timestamp}). Note: The workflow is currently inactive ("active": false); enable it after configuration. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by GiovanniSegar
How to use the workflow This workflow takes a Loom link, extracts the video ID, uses the Loom API to download the video, then sends it to Gemini along with your question. Finally, it sends the output to Slack. To use it, you just need to add your own API key for Gemini and Slack connection. Click the link above to get your Gemini API key, then add a generic "Query auth" type credential in n8n. The name will be "key" and the value will be your API key. One way to customize this workflow would be to make the trigger any received email, extract the Loom link, and run an auto-prompt like "Describe this video in detail".
by n8n Team
This workflow automates a two way sync of customer data between Pipedrive and MySQL. It will create new records in one source if it only exists in the other. Where matching records have different data for name, phone number or email address, it will sync the most recently updated version.
by Ranjan Dailata
Description This workflow automates the process of scraping the latest discussions from HackerNews, transforming raw threads into human readable content using Google Gemini, and exporting the final content into a well-formatted Google Doc. Overview This n8n workflow is responsible for extracting trending posts from the HackerNews API. It loops through each item, performs HTTP data extraction, utilizes Google Gemini to generate human-readable insights, and then exports the enriched content into Google Docs for distribution, archiving, or content creation. Who this workflow is for Tech Newsletter Writers**: Automate the collection and summarization of trending HackerNews posts for inclusion in weekly or daily newsletters. Content Creators & Bloggers**: Quickly generate structured summaries and insights from HackerNews threads to use as inspiration or supporting content for blog posts, videos, or social media. Startup Founders & Product Builders**: Monitor HackerNews for discussions relevant to your niche or competitors, and keep a pulse on the community’s opinions. Investors & Analysts**: Surface early signals from the tech ecosystem by identifying what’s trending and how the community is reacting. Researchers & Students**: Analyze popular discussions and emerging trends in technology, programming, and startups—enriched with AI-generated insights. Digital Agencies & Consultants**: Offer HackerNews monitoring and insight reports as a value-added service to clients interested in the tech space. Tools Used n8n**: The core automation engine that manages the trigger, transformation, and export. HackerNews API**: Provides access to trending or new HN posts. Google Gemini**: Enriches HackerNews content with structured insights and human-like summaries. Google Docs**: Automatically creates and updates a document with the enriched content, ready for sharing or publishing. How to Install Import the Workflow**: Download the .json file and import it into your n8n instance. Set Up HackerNews Source**: Choose whether to use the HN API (via HTTP Request node) or RSS Feed node. Configure Gemini API**: Add your Google Gemini API key and design the prompt to extract pros/cons, key themes, or insights. Set Up Google Docs Integration**: Connect your Google account and configure the Google Docs node to create/update a document. Test and Deploy**: Run a test job to ensure data flows correctly and outputs are formatted as expected. Use Cases Tech Newsletter Authors**: Generate ready-to-use summaries of trending HackerNews threads. Startup Founders**: Stay informed on key discussions, product launches, and community feedback. Investors & Analysts**: Spot early trends, technical insights, and startup momentum directly from HN. Researchers**: Track community reactions to new technologies or frameworks. Content Creators**: Use the enriched data to spark blog posts, YouTube scripts, or LinkedIn updates. Connect with Me Email: ranjancse@gmail.com LinkedIn: https://www.linkedin.com/in/ranjan-dailata/ Get Bright Data: Bright Data (Supports free workflows with a small commission) #n8n #automation #hackernews #contentcuration #aiwriting #geminiapi #googlegemini #techtrends #newsletterautomation #googleworkspace #rssautomation #nocode #structureddata #webscraping #contentautomation #hninsights #aiworkflow #googleintegration #webmonitoring #hnnews #aiassistant #gdocs #automationtools #gptlike #geminiwriter
by Abbas Ali
This n8n workflow automatically finds apartments for rent in Germany, filters them by your city, rent budget, and number of rooms, and applies to them via email. Each application includes: A personalized German cover letter Schufa report (fetched dynamically from Google Drive) Recent salary slips (also fetched from Google Drive) The workflow runs daily at a scheduled time, emails landlords or agencies automatically, and logs every application into a Google Sheet for tracking. How It Works Scheduled Trigger – Runs every day at 9 AM (adjustable). Fetch Listings – Uses immobilienscout24 API (or similar) to pull rental listings for your selected city. Filter Listings – Keeps only listings matching your CITY, MAX_RENT, and ROOMS settings. Fetch Documents – Retrieves your Schufa report and salary slips from Google Drive (no need for local hosting). Generate Cover Letter – Creates a personalized German-language letter per apartment. Send Email Application – Sends the email to the landlord or agent with cover letter + documents attached. Log Applications – Saves each application (title, address, rent, date) in a Google Sheet. How to Use Import the workflow JSON into n8n. Set environment variables in n8n (for security): immobilienscout24_TOKEN: Your immobilienscout24 API token immobilienscout24_LISTING_ACTOR: Actor ID for your preferred rental listing scraper (or custom) MY_EMAIL: Your sender email address (SMTP configured in n8n) SCHUFA_FILE_ID: Google Drive File ID for your Schufa PDF SALARY_FILE_ID: Google Drive File ID for your Salary Slips PDF APPLICATION_SHEET_ID: Google Sheet ID to log applications Authenticate Google Drive and Google Sheets (OAuth2 in n8n). Customize search filters in the Set Config node: CITY (e.g., Berlin) MAX_RENT (e.g., 1200) ROOMS (e.g., 2) Activate the workflow – It will run daily at the configured time and send applications automatically. Check your Google Sheet – Every application will be logged for tracking. Requirements An immobilienscout24 account (or another apartment listing API, can be substituted). A Google account (for Drive and Sheets integration). A Schufa report (PDF) uploaded to Google Drive. Recent salary slips (PDF) uploaded to Google Drive. An SMTP-configured email account for sending applications. n8n instance (self-hosted or cloud) with: Google Drive and Google Sheets credentials configured Environment variables set for tokens and file IDs A working email SMTP setup
by dmr
This n8n workflow implements a version of the Adaptive Retrieval-Augmented Generation (RAG) framework. It recognizes that the best way to retrieve information often depends on the type of question asked. Instead of a one-size-fits-all approach, this workflow adapts its strategy based on the user's query intent. 🌟 How it Works Receive Query: Takes a user query as input (along with context like a chat session ID and Vector Store collection ID if used as sub-workflow). Classify Query: First, the workflow classifies the query into a predefined category. This template uses four examples: Factual: For specific facts. Analytical: For deeper explanations or comparisons. Opinion: For subjective viewpoints. Contextual: For questions relying on specific background. Select & Adapt Strategy: Based on the classification, it selects a corresponding strategy to prepare for information retrieval. The example strategies aim to: Factual: Refine the query for precision. Analytical: Break the query into sub-questions for broad coverage. Opinion: Identify different viewpoints to look for. Contextual: Incorporate implied or user-specific context. Retrieve Info: Uses the output of the selected strategy to search the specified knowledge base (Qdrant vector store - change as needed) for relevant documents. Generate Response: Constructs a response using the retrieved documents, guided by a prompt tailored to the original query type. By adapting the retrieval strategy, this workflow aims to provide more relevant results tailored to the user's intent. ⚙️ Usage & Flexibility Sub-Workflow:** Designed to be called from other n8n workflows, passing user_query, chat_memory_key, and vector_store_id as inputs. Chat Testing:** Can also be triggered directly via the n8n Chat interface for easy testing and interaction. Customizable Framework:** The query categories (Factual, Analytical, etc.) and the associated retrieval strategies are examples. You can modify or replace them entirely to fit your specific domain or requirements. 🛠️ Requirements Credentials:** You will need API credentials configured in your n8n instance for: Google Gemini (AI Models) Qdrant (Vector Store)
by Automate With Marc
🎬 Veo3 Instagram Reel Generator – AI-Powered Ad Creation in Minutes Description: This no-code workflow transforms your creative brief into an engaging Instagram Reel using OpenAI and Veo3 API (via Wavespeed) — fully automated in n8n. Just type a product, theme, or trend via chat, and get a short-form video plus caption delivered and logged, ready to post. Perfect for marketers, creators, and content teams looking to scale their ad content output without hiring editors or creative agencies. Watch step-by-step build video tutorial here: https://www.youtube.com/@Automatewithmarc ⚙️ How It Works: 💬 Chat Trigger Start by sending a message like “Create an ad for a minimalist perfume brand using the ‘quiet luxury’ trend.” 🧠 Prompt Engineer (ChatGPT) Generates a 5–8 second descriptive video prompt suitable for Veo3 based on your input — including visual tone, motion, and hook. 📡 API Call to Veo3 via Wavespeed Submits the prompt to create a short video (9:16 ratio, ~8 seconds), then polls for the final video URL. ✍️ Caption Generator (GPT) Creates an Instagram-friendly caption to pair with the video, using a playful, impactful writing style. 📄 Google Sheets Integration Logs each generated video prompt, final video URL, caption, and status into a Google Sheet for easy management and scheduling. 🔌 Tools & Integrations: OpenAI GPT (Prompt generation & caption copywriting) Veo3 via Wavespeed API (Video generation) Google Sheets (Content tracking and publishing queue) Telegram / Chat UI trigger (Optional – easily swappable) 💡 Use Cases: Instagram & TikTok ad generation Creative automation for digital agencies Short-form UGC testing at scale Trend-driven campaign ideation
by Automate With Marc
📱 Veo3 Instagram Agent – Create & Auto-Post Reels with Blotato Description: This no-code workflow automates the full pipeline of generating and publishing Instagram Reels using Veo3 (via Wavespeed API). From prompt to post, it handles content ideation, short-form video generation, caption writing, logging, and even automatic publishing to Instagram via Blotato. Perfect for creators, brands, and marketers who want to scale content creation without needing to shoot or edit videos manually. 🔗 Watch the full step-by-step tutorial on how to build this workflow: https://youtu.be/s-KzxeKWmIA?si=6x8WKMeiyWodZWVq Google Sheet Template: https://docs.google.com/spreadsheets/d/1bA-PQTrvekC1Rti-XumGANgjIwLjvcFCqoIxVCYsq2E/edit?usp=sharing 🚀 What This Workflow Does: Trigger via Chat or Telegram Start with a simple message like: "Make a reel for a luxury minimalist candle brand using calm aesthetics." AI Video Prompt Generation Uses OpenAI to craft a visually rich, platform-optimized video description prompt. 🎞️ Video Creation with Veo3 API Submits your prompt to Veo3 to create a short video (9:16 ratio, 8 seconds) with motion, tone, and trend styles. ✍️ Caption Writing An AI agent writes an engaging and playful caption based on the video content. 📄 Google Sheets Logging Stores prompt, video URL, caption, and status in a GSheet to keep track of all generated assets. 📤 Auto-Publish to Instagram Posts the video + caption directly to Instagram using Blotato’s social media publishing API. 🔌 Tools & Integrations Used: OpenAI for prompt & caption generation Wavespeed API (Veo3) for video generation Google Sheets for tracking Blotato for scheduling & publishing content n8n for orchestration and automation logic 💡 Use Cases: Content calendar automation for small teams Trend-based ad creation and testing UGC-style reel generation for e-commerce Rapid ideation & creative experimentation
by Catalina Kuo
Overview Do you often forget to record expenses? 你是不是會常常忘記紀錄花費? Let Spending Tracker Bot help you! 讓 Spending Tracker Bot 來幫你! This AI image/text Spending Tracker LINE Bot Workflow allows you to quickly create a customized spending tracker robot without writing a line of code. At any time, you can speak or send a photo, and the AI will parse it and automatically log the expense to your cloud ledger. 這套 AI 圖片文字記帳 LINE Bot Workflow ,讓你不用寫一行程式碼,就能快速打造一個量身訂製的記帳機器人。無論何時,只需要口述或發送一張照片,AI 就會幫你整理好自動計入雲端帳本 Preparation ① Enable the Google Sheets API in GCP and complete the OAuth setup ② Create the Google Sheet and populate the field names (Feel free to modify based on your own needs) ③ Configure the Webhook URL in the LINE Developers Console ④ OpenAI API Key ① 在 GCP 啟用 Google Sheets API,並完成 OAuth ② 建立並填好 Google Sheet 欄名 (按照自己的需求做更動) ③ 於 LINE Developers 控制台設定 Webhook URL ④ OpenAI API Key Node Configurations Webhook Purpose: The URL is used to receive incoming requests from LINE. Configuration: Paste this URL into the Webhook URL field in your LINE Developers Console. 用途: 要接收 Line 的 URL 設定: 將 URL 放到 Line Webhook URL Switch based on Expense Type & Set/Https Purpose: To distinguish whether the incoming message is text or an image. Configuration: Use a Switch node to route the flow accordingly. 用途: 區分 text 或 image 設定: switch 分流 AI Agent Purpose: To extract and organize the required fields. Configuration: Chat Model & Structured Output Parser. 用途: 整理出需要的欄位 設定: Chat Model & Structured Output Parser Create a deduplication field Purpose: To prevent duplicate entries by creating a unique "for_deduplication" field. Configuration: Join multiple field names using hyphens (-) as separators. 用途: 確保不會重複寫入,先創建一個"去重使用"欄位 設定: 用 - 連接多個欄位 Aggregrate & Merge_all Purpose: To prevent duplicate entries in the data table. Configuration: Read the Google sheet, extract the existing "for_deduplication" column into a dedupeList, and compare it against the newly generated "for_deduplication" value from the previous step. 用途: 防止重複寫入資料表 設定:讀取雲端表,將原本的"去重使用欄位"整理成dedupeList,與前一步整理好的"去重使用"欄位做比對 Response Switch Purpose: To route data and send appropriate replies based on the content. Configuration: Use the replyToken to respond after branching logic. Depending on the result, either write to the data table or return a message: ✅ Expense recorded successfully: <for_deduplication> Irrelevant details or images will not be logged. ⚠️ This entry has already been logged and will not be duplicated. 用途: 資料分流,回應訊息 設定:使用 replyToken ,資料分流後,寫入資料表或回應訊息 ✅ 記帳成功 : <去重使用欄位> 不相關明細或圖片,不會計入 ⚠️ 此筆資料已記錄過,不會重複記帳 Step by step teaching notes 【Auto Expense Tracker from LINE Messages with GPT-4 and Google Sheets】 【AI 圖片文字記帳 Line Bot,自動記帳寫入 Google Sheet】
by Oneclick AI Squad
This automated n8n workflow detects and manages fraudulent booking transactions through comprehensive AI-powered analysis and multi-layered security checks. The system processes incoming travel booking data, performs IP geolocation verification, enriches transaction details with AI insights, calculates dynamic risk scores, and executes automated responses based on threat levels. All transactions are logged and appropriate notifications are sent to relevant stakeholders. Good to Know The workflow combines multiple detection methods, including IP geolocation, AI analysis, and risk scoring algorithms Google Gemini Chat Model provides advanced natural language processing for transaction analysis Risk levels are dynamically calculated and categorized as CRITICAL, HIGH, or standard risk Automated blocking and flagging system protects against fraudulent transactions in real-time All transaction data is logged to Google Sheets for audit trails and pattern analysis The system respects API rate limits and includes proper error handling mechanisms How It Works 1. Initial Data Ingestion & Extraction Monitors and captures incoming booking transaction data from various sources Extracts key booking details including user information, payment data, booking location, and transaction metadata Performs initial data validation and formatting for downstream processing 2. IP Geolocation and AI Analysis IP Geolocation Check**: Validates booking IP addresses by checking geolocation details and comparing against expected user locations AI Agent Integration**: Utilizes Google Gemini Chat Model to analyze booking patterns, user behavior, and transaction anomalies Enhanced Data Processing**: Enriches transaction data with geographical context and AI-driven risk indicators 3. Risk Calculation and Decision Logic Enhanced Risk Calculator**: Combines AI-generated risk scores with geolocation-based factors, payment method analysis, and historical patterns Critical Risk Check**: Flags transactions with risk levels marked as CRITICAL for immediate action High Risk Check**: Identifies HIGH risk transactions requiring additional verification steps Dynamic Scoring**: Adjusts risk calculations based on real-time threat intelligence and pattern recognition 4. Action & Notification Block User Account**: Automatically blocks user accounts for CRITICAL risk transactions to prevent immediate fraud Flag for Review**: Marks HIGH risk transactions for manual review by fraud prevention teams Send Notifications**: Dispatches real-time alerts via email and messaging systems to security teams Automated Responses**: Sends appropriate messages to users based on transaction status and risk level 5. Logging & Response Log to Google Sheets**: Records all transaction details, risk scores, and actions taken for comprehensive audit trails Flag for Review**: Maintains detailed logs of flagged transactions for pattern analysis and machine learning improvements Response Tracking**: Monitors and logs all automated responses and manual interventions How to Use Import the workflow into your n8n instance Configure Google Gemini Chat Model API credentials for AI analysis Set up IP geolocation service API access for location verification Configure Google Sheets integration for transaction logging Establish Gmail/email credentials for notification delivery Define risk thresholds and scoring parameters based on your fraud tolerance levels Test the workflow with sample booking data to verify all components function correctly Monitor initial deployments closely to fine-tune risk scoring algorithms Establish manual review processes for flagged transactions Set up regular monitoring and maintenance schedules for optimal performance Requirements Google Gemini Chat Model API access IP Geolocation service API credentials Google Sheets API integration Gmail API or SMTP email service for notifications n8n instance with appropriate node modules installed Customizing This Workflow Risk Scoring Parameters**: Adjust risk calculation algorithms and thresholds based on your specific fraud patterns and business requirements AI Model Configuration**: Fine-tune Google Gemini prompts and analysis parameters for improved accuracy in your use case Notification Channels**: Add or modify notification methods including Slack, SMS, or webhook integrations Data Sources**: Extend input methods to accommodate additional booking platforms or payment processors Logging Destinations**: Configure alternative or additional logging systems such as databases or external SIEM platforms Geographic Rules**: Customize geolocation validation rules based on your service areas and customer base Automated Actions**: Modify or expand automated response actions based on your fraud prevention policies Review Workflows**: Integrate with existing fraud review systems or ticketing platforms for seamless manual review processes
by Evozard
This workflow functions by integrating Shopify customers into Odoo customers. Trigger: Shopify – New Customer Created The workflow starts when a new customer is added in Shopify. Action: Odoo – Search Contact by Email It checks in Odoo to see if a contact already exists with the same email address as the Shopify customer. Condition: Email Match Check If a contact with the same email is found, the workflow ends (no duplicate contact is created). If no match is found, the workflow proceeds to the next step. Action: Odoo – Create New Contact A new contact is created in Odoo using the customer's: Full name Email address Phone number Full Address (whichever is available)
by Oneclick AI Squad
This guide walks you through setting up an AI-driven workflow to automate flight and hotel reservation processes using a conversational travel booking system. The workflow accepts booking requests, processes them via APIs, and sends confirmations, enabling a seamless travel booking experience. What’s the Goal? Automatically accept and process booking requests for flights and hotels via HTTP POST. Use AI to understand natural language requests and route them to appropriate data processors. Search for flights and hotels using external APIs and process booking confirmations. Send confirmation emails and return structured booking data to users. Enable an automated system for efficient travel reservations. By the end, you’ll have a self-running system that handles travel bookings effortlessly. Why Does It Matter? Manual booking processes are time-consuming and prone to errors. This workflow offers: Zero Human Error**: AI ensures accurate request parsing and booking processing. Time-Saving Automation**: Automates the entire booking lifecycle, boosting efficiency. Seamless Confirmation**: Sends automated emails and responses without manual intervention. Enhanced User Experience**: Provides a conversational interface for bookings. Think of it as your reliable travel booking assistant that keeps the process smooth and efficient. How It Works Here’s the step-by-step flow of the automation: Step 1: Trigger the Workflow Webhook Trigger**: Accepts incoming booking requests via HTTP POST, initiating the workflow. Step 2: Parse the Request AI Request Parser**: Uses AI to understand natural language booking requests (e.g., flight or hotel) and extracts relevant details. Step 3: Route Booking Type Booking Type Router**: Determines whether the request is for a flight or hotel and routes it to the respective data processor. Step 4: Process Flight Data Flight Data Processor**: Handles flight-specific data and prepares it for the search API. Step 5: Search Flight API Flight Search API**: Searches for available flights based on parameters (e.g., https://api.aviationstack.com) and returns results. Step 6: Process Hotel Data Hotel Data Processor**: Handles hotel-specific data and prepares it for the search API. Step 7: Search Hotel API Hotel Search API**: Searches for available hotels based on parameters (e.g., https://api.booking.com) and returns results. Step 8: Process Flight Booking Flight Booking Processor**: Processes flight bookings and generates confirmation details. Step 9: Process Hotel Booking Hotel Booking Processor**: Processes hotel bookings and generates confirmation details. Step 10: Generate Confirmation Message Confirmation Message Generator**: Creates structured confirmation messages for the user. Step 11: Send Confirmation Email Send Confirmation Email**: Sends booking confirmation via email to the user. Step 12: Send Response Send Response**: Returns structured booking data to the user, completing the workflow. How to Use the Workflow? Importing the workflow in n8n is a straightforward process. Follow these steps to import the Conversational Travel Booker workflow: Download the Workflow: Obtain the workflow file (e.g., JSON export from n8n). Open n8n: Log in to your n8n instance. Import Workflow: Navigate to the workflows section, click "Import," and upload the workflow file. Configure Nodes: Adjust settings (e.g., API keys, webhook URLs) as needed. Execute Workflow: Test and activate the workflow to start processing bookings. Requirements n8n account and instance setup. Access to flight and hotel search APIs (e.g., Aviationstack, Booking.com). Email service integration for sending confirmations. Webhook URL for receiving booking requests. Customizing this Workflow Modify the AI Request Parser to handle additional languages or booking types. Update API endpoints in Flight Search API and Hotel Search API nodes to match your preferred providers. Adjust the Send Confirmation Email node to include custom email templates or additional recipients. Schedule the Webhook Trigger to align with your business hours or demand peaks.