by Jason Krol
Notion Weekly Journal AI Summary This workflow will run on a weekly schedule and retrieve your Notion Daily Journal pages for the past week and aggregate them into a ChatGPT generated concise summary. It will save that weekly summary back to your Notion as a new Note in addition to posting to a personal Discord channel. Additionally it will also retrieve all of the Tasks you've completed in the past week and provide a quick total with a congratulatory message to a Discord channel as well. Requirements/Setup: You need Notion setup w/ a Notes database If you want the Discord messages, setup a Discord webhook for your channel as well, or simply delete the Discord nodes. One of the properties for the Notes db should be Type with a value of Journal The contents of your daily Journal pages can be whatever you want I've found what works best for me is the format of "What was a highlight of the day?", "What was a low point of the day?", and "What decisions did I delegate, delay, or dodge?" You should also create an additional Type for your Weekly summary page that gets created - in this case I used simply Weekly Automate this to run weekly on your day of choice. I tend to only journal on weekdays so I've set mine up to run every Friday retrieving the past week's Journal entries. Options: You don't have to use Discord, feel free to swap out with Slack or remove altogether. You also don't need to use the Tasks summary bottom half, simply remove that if you don't want it or need it. You can easily reuse this workflow to aggregate your Weekly Summary notes (that this workflow auto generates/saves) to generate a Quarterly or even Yearly summary!
by Oneclick AI Squad
In this guide, we’ll walk you through setting up a smart workflow that triggers on new restaurant orders, extracts and formats customer and dish details from Google Sheets, uses Gemini AI to recommend dishes or offers, and sends suggestions via Telegram. Ready to automate your order processing and enhance customer experience? Let’s dive in! What’s the Goal? Automatically trigger the workflow when a new order is placed. Extract and format customer information and order details from Google Sheets. Use Gemini AI to analyze orders and recommend dishes or offers. Send personalized suggestions to customers via Telegram. Enable real-time order processing and customer engagement. By the end, you’ll have a smart system that processes orders and suggests items effortlessly. Why Does It Matter? Manual order processing and suggestion generation are inefficient and miss opportunities. Here’s why this workflow is a game changer: Real-Time Efficiency**: Instantly process orders and suggest items. Personalized Engagement**: AI-driven suggestions enhance customer satisfaction. Time-Saving Automation**: Reduce manual effort in order management. Improved Sales**: Targeted recommendations can boost order value. Think of it as your intelligent assistant for orders and customer delight. How It Works Here’s the step-by-step magic behind the automation: Step 1: New Order Trigger Trigger the workflow when a new order is detected (e.g., via a form submission). Step 2: Extract & Format Order Extract and format dish ordering details from the customer order details sheet for further processing. Step 3: Save Customer Info Save customer information (e.g., ID, name, mobile number) from the customer details sheet. Step 4: Save Dish Info Save dish details (e.g., name, quantity, price) from the customer order details sheet. Step 5: Prepare Dish Details for AI Prepare the dish details for AI analysis to generate recommendations. Step 6: Clean Data for Input to Improve AI Understanding Clean and structure the data to enhance AI comprehension. Step 7: Use Gemini AI to Recommend Dishes or Offers Utilize Gemini AI (via Google Chat Model and Think Tool) to recommend dishes or offers based on order data. Step 8: Format AI Suggestions Format the AI-generated suggestions into a Telegram-friendly message. Step 9: Send Suggestions via Telegram Send the formatted suggestions directly to the customer via Telegram. How to Use the Workflow? Importing a workflow in n8n is a straightforward process that allows you to use pre-built workflows to save time. Below is a step-by-step guide to importing the Smart Restaurant Order & Suggestion System workflow in n8n. Steps to Import a Workflow in n8n Obtain the Workflow JSON Source the Workflow: Workflows are shared as JSON files or code snippets, e.g., from the n8n community, a colleague, or exported from another n8n instance. Format: Ensure you have the workflow in JSON format, either as a file (e.g., workflow.json) or copied text. Access the n8n Workflow Editor Log in to n8n (via n8n Cloud or self-hosted instance). Navigate to the Workflows tab in the n8n dashboard. Click Add Workflow to create a blank workflow. Import the Workflow Option 1: Import via JSON Code (Clipboard): Click the three dots (⋯) in the top-right corner to open the menu. Select Import from Clipboard. Paste the JSON code into the text box. Click Import to load the workflow. Option 2: Import via JSON File: Click the three dots (⋯) in the top-right corner. Select Import from File. Choose the .json file from your computer. Click Open to import. Setup Notes Google Sheet Columns**: Customer Details Sheet: Customer id, Customer name, Customer mobile number (e.g., CUST-JW4Z8Y, ajay, 9898989898; CUST-VEITPW, akash, 9898976898). Customer Order Details Sheet: Customer id, Dish name, Dish quantity, Per unit price, Actual price (e.g., CUST-JW4Z8Y, Tandoori Chicken, 1, 250, 250; CUST-VEITPW, Masala Dosa, 1, 150, 150). Google Sheets Credentials**: Configure OAuth2 settings in the extract and save nodes with your Google Sheet ID and credentials. Gemini AI**: Set up the Gemini AI node with Google Chat Model and Think Tool credentials. Telegram Integration**: Authorize the Send Suggestions node with Telegram API credentials and the customer’s chat ID or mobile number. Trigger Setup**: Configure the New Order Trigger node to detect new orders (e.g., via form or webhook).
by MattF
This workflow tracks week-over-week changes in Google Search Console performance and highlights the top movers across keyword segments like brand, nonbrand, and content categories. Instead of providing a routine check, it focuses on significant movements by: Sending a Slack alert only if a query crosses a defined movement threshold. Emailing a structured report with the Top 25 increases and Top 25 decreases for clicks, including % changes and linked URLs It’s designed to surface the most important shifts, helping SEO teams catch big wins, losses, or anomalies early. How it works Runs weekly (e.g. every Monday) to compare last week’s GSC data to the week prior. Segments traffic based on query and page (e.g. brand terms, category page URLs, etc.). Calculates delta and % change for clicks, CTR, impressions, and position. Filters and flags top movers with large shifts (default: ±200 clicks and ±30%). Sends Slack alerts only if meaningful changes are detected. Emails a full HTML table report showing the Top 25 up/down queries per segment. Setup steps Requires a connected Google Search Console account. Slack alert is included by default (can be replaced with email, webhook, or other tools). Customize your brand terms and URL filters to match your segments (e.g. recipes, blog, category pages). Typical setup time: 15–25 minutes depending on the number of segments and filters you want. Note: “Recipes” is used in the example to show how to segment by content type. You can update this to reflect your own site’s structure.
by Gain FLow AI
AI Latest News Content Script Writer Overview This workflow automates the daily generation of viral short-form video content ideas tailored for founders and business leaders. It scrapes fresh AI-related news and trends from various topics, synthesizes the information, and then uses AI to craft complete content packages—including video scripts, captivating captions, and punchy text overlays. All generated content is saved to a Google Sheet, ready for your review and use. Use Case This workflow is perfect for: Founders & Entrepreneurs**: Consistently produce engaging content to build authority and attract inbound leads without a dedicated content team. AI Thought Leaders**: Stay on top of the latest AI news and effortlessly create shareable insights. Content Marketing Teams**: Automate the ideation and initial drafting phases for short-form video strategies. Agencies**: Offer a unique AI-powered content generation service to your clients. How It Works Scheduled Daily Trigger: The workflow runs automatically every day at 6 AM IST, ensuring you always have fresh content ideas to start your day. AI-Powered News Gathering: It uses Perplexity AI to fetch the latest, most interesting, and relevant stories across three key AI topics: Topic 1: General AI News Topic 2: AI Market and Industry Trends Topic 3: AI Business Automation Organize and Combine Content: The information from each topic is organized, and then all content and their respective citations are combined into a single, comprehensive input. Personalize "About Me": Crucially, a configurable "About me" node allows you to define the personal brand of the founder (e.g., Name, Niche, Business Name, Business Type). This context is fed to the AI to ensure generated content aligns perfectly with your persona and business objectives. Generate Content Packages: Leveraging OpenAI (acting as "CreatorAI"), the workflow takes the combined news and your "About me" information to: Identify a Unique Angle: Finds a distinct, engaging angle from the input that aligns with key content pillars (e.g., AI solving business pain points, future of work with AI). Craft Video Scripts: Generates concise video scripts (under 700 characters) with powerful hooks, mini-narratives (problem → AI solution → impact), and a focus on tangible business benefits. It subtly references your business as a thought leader, not a direct pitch. Write Captions: Creates friendly, expert-toned captions with engaging hooks, more context, a clear call to action (e.g., "Comment 'Workflow' for more"), and relevant hashtags. Design Text Overlays: Produces short, punchy text overlays (3-7 words, ALL CAPS or Title Case) perfect for video thumbnails or initial screens. Save to Google Sheet: Each generated content package (Text Overlay, Video Script, Caption) is appended as a new row in your designated Google Sheet ("Content Idea" sheet within "Video Automation (Vansh)"). Notify User: Finally, you'll receive an email notification confirming that new content ideas have been generated and saved to your Google Sheet. How to Set It Up To set up this AI Viral Content Generator, follow these steps: API Keys & Credentials: Perplexity AI API Key: Obtain your API key from Perplexity AI and replace the Bearer token in the "Topic 1", "Topic 2", and "Topic 3" HTTP Request nodes. OpenAI API Key: Connect your OpenAI API key in n8n and link it to the "Content Generation" node. Google Sheets Account: Ensure your Google Sheets OAuth2 API credentials are set up and connected to the "Save Data" node. Gmail Account: Connect your Gmail OAuth2 credentials to the "Notify user" node. Google Sheet Setup: Copy the Google Sheet Template provided. This template has predefined columns for "Text Overlay", "Video Script", "Caption", "Approval", and "Published". Update the documentId in the "Save Data" Google Sheets node with the ID of your copied template. Personalize "About me": Open the "About me" node. Fill in your Name, Niche, Business Name, Business Type, Website, and detailed Key Services & Products. This is crucial for the AI to generate relevant and personalized content. Configure Notification Email: In the "Notify user" node, update the sendTo field with your email address where you want to receive notifications. Set Schedule: The "Schedule Trigger" is set to run daily at 6 AM IST. You can adjust the time to your preference. Activate and Monitor: Activate the workflow. It will now automatically generate content ideas daily. Check your Google Sheet regularly to review the new content, mark it for approval, and track its publication status. This workflow is your secret weapon for consistently creating engaging, AI-driven short-form video content!
by Davide
This workflow allows users to generate AI videos using the cheaper model Google Veo3 Fast, save them to Google Drive, generate optimized titles with GPT-4o, and automatically upload them to YouTube and TikTok with Upload-Post. The entire process is triggered from a Google Sheet that acts as the central interface for input and output. IT automates video creation, uploading, and tracking, ensuring seamless integration between Google Sheets, Google Drive, Google Veo3 Fast, TikTok and YouTube. Benefits of this Workflow 💡 No Code Interface**: Trigger and control the video production pipeline from a simple Google Sheet. ⚙️ Full Automation**: Once set up, the entire video generation and publishing process runs hands-free. 🧠 AI-Powered Creativity**: Generates engaging YouTube and TikTok titles using GPT-4o. Leverages advanced generative video AI from Google Veo3. 📁 Cloud Storage & Backup**: Stores all generated videos on Google Drive for safekeeping. 📈 YouTube Ready**: Automatically uploads to YouTube with correct metadata, saving time and boosting visibility. 📈 TikTok Ready**: Automatically uploads to TikTok with correct metadata, saving time and boosting visibility. 🧪 Scalable**: Designed to process multiple video prompts by looping through new entries in Google Sheets. 🔒 API-First**: Utilizes secure API-based communication for all services. How It Works Trigger: The workflow can be started manually ("When clicking ‘Test workflow’") or scheduled ("Schedule Trigger") to run at regular intervals (e.g., every 5 minutes). Fetch Data: The "Get new video" node retrieves unfilled video requests from a Google Sheet (rows where the "VIDEO" column is empty). Video Creation: The "Set data" node formats the prompt and duration from the Google Sheet. The "Create Video" node sends a request to the Fal.run API (Google Veo3 Fast) to generate a video based on the prompt. Status Check: The "Wait 60 sec." node pauses execution for 60 seconds. The "Get status" node checks the video generation status. If the status is "COMPLETED," the workflow proceeds; otherwise, it waits again. Video Processing: The "Get Url Video" node fetches the video URL. The "Generate title" node uses OpenAI (GPT-4.1) to create an SEO-optimized YouTube and TikTok title. The "Get File Video" node downloads the video file. Upload & Update: The "Upload Video" node saves the video to Google Drive. The "HTTP Request" node uploads the video to YouTube via the Upload-Post API. The "HTTP Request" node uploads the video to TikTok via the Upload-Post API. The "Update Youtube URL" and "Update result" nodes update the Google Sheet with the video URL and YouTube link. Set Up Steps Google Sheet Setup: Create a Google Sheet with columns: PROMPT, DURATION, VIDEO, and YOUTUBE_URL. Share the Sheet link in the "Get new video" node. API Keys: Obtain a Fal.run API key (for Veo3) and set it in the "Create Video" node (Header: Authorization: Key YOURAPIKEY). Get an Upload-Post API key (for YouTube uploads) and configure the "HTTP Request" node (Header: Authorization: Apikey YOUR_API_KEY). Get an Upload-Post API key (for TikTok uploads) and configure the "HTTP Request" node (Header: Authorization: Apikey YOUR_API_KEY). YouTube Upload Configuration: Replace YOUR_USERNAME in the "HTTP Request" node with your Upload-Post profile name. Schedule Trigger: Configure the "Schedule Trigger" node to run periodically (e.g., every 5 minutes). Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Greg Evseev
This workflow template provides a robust solution for efficiently sending multiple prompts to Anthropic's Claude models in a single batch request and retrieving the results. It leverages the Anthropic Batch API endpoint (/v1/messages/batches) for optimized processing and outputs each result as a separate item. Core Functionality & Example Usage Included This template includes: The Core Batch Processing Workflow: Designed to be called by another n8n workflow. An Example Usage Workflow: A separate branch demonstrating how to prepare data and trigger the core workflow, including examples using simple strings and n8n's Langchain Chat Memory nodes. Who is this for? This template is designed for: Developers, data scientists, and researchers** who need to process large volumes of text prompts using Claude models via n8n. Content creators** looking to generate multiple pieces of content (e.g., summaries, Q&As, creative text) based on different inputs simultaneously. n8n users** who want to automate interactions with the Anthropic API beyond single requests, improve efficiency, and integrate batch processing into larger automation sequences. Anyone needing to perform bulk text generation or analysis tasks with Claude programmatically. What problem does this workflow solve? Sending prompts to language models one by one can be slow and inefficient, especially when dealing with hundreds or thousands of requests. This workflow addresses that by: Batching:** Grouping multiple prompts into a single API call to Anthropic's dedicated batch endpoint (/v1/messages/batches). Efficiency:** Significantly reducing the time required compared to sequential processing. Scalability:** Handling large numbers of prompts (up to API limits) systematically. Automation:** Providing a ready-to-use, callable n8n structure for batch interactions with Claude. Structured Output:** Parsing the results and outputting each individual prompt's result as a separate n8n item. Use Cases: Bulk content generation (e.g., product descriptions, summaries). Large-scale question answering based on different contexts. Sentiment analysis or data extraction across multiple text snippets. Running the same prompt against many different inputs for research or testing. What the Core Workflow does (Triggered by the 'When Executed by Another Workflow' node) Receive Input: The workflow starts when called by another workflow (e.g., using the 'Execute Workflow' node). It expects input data containing: anthropic-version (string, e.g., "2023-06-01") requests (JSON array, where each object represents a single prompt request conforming to the Anthropic Batch API schema). Submit Batch Job: Sends the formatted requests data via POST to the Anthropic API /v1/messages/batches endpoint to create a new batch job. Requires Anthropic credentials. Wait & Poll: Enters a loop: Checks if the processing_status of the batch job is ended. If not ended, it waits for a set interval (10 seconds by default in the 'Batch Status Poll Interval' node). It then checks the batch job status again via GET to /v1/messages/batches/{batch_id}. Requires Anthropic credentials. This loop continues until the status is ended. Retrieve Results: Once the batch job is complete, it fetches the results file by making a GET request to the results_url provided in the batch status response. Requires Anthropic credentials. Parse Results: The results are typically returned in JSON Lines (.jsonl) format. The 'Parse response' Code node splits the response text by newlines and parses each line into a separate JSON object, storing them in an array field (e.g., parsed). Split Output: The 'Split Out Parsed Results' node takes the array of parsed results and outputs each result object as an individual item from the workflow. Prerequisites An active n8n instance (Cloud or self-hosted). An Anthropic API account with access granted to Claude models and the Batch API. Your Anthropic API Key. Basic understanding of n8n concepts (nodes, workflows, credentials, expressions, 'Execute Workflow' node). Familiarity with JSON data structures for providing input prompts and understanding the output. Understanding of the Anthropic Batch API request/response structure. (For Example Usage Branch) Familiarity with n8n's Langchain nodes (@n8n/n8n-nodes-langchain) if you plan to adapt that part. Setup Import Template: Add this template to your n8n instance. Configure Credentials: Navigate to the 'Credentials' section in your n8n instance. Click 'Add Credential'. Search for 'Anthropic' and select the Anthropic API credential type. Enter your Anthropic API Key and save the credential (e.g., name it "Anthropic account"). Assign Credentials: Open the workflow and locate the three HTTP Request nodes in the core workflow: Submit batch Check batch status Get results In each of these nodes, select the Anthropic credential you just configured from the 'Credential for Anthropic API' dropdown. Review Input Format: Understand the required input structure for the When Executed by Another Workflow trigger node. The primary inputs are anthropic-version (string) and requests (array). Refer to the Sticky Notes in the template and the Anthropic Batch API documentation for the exact schema required within the requests array. Activate Workflow: Save and activate the core workflow so it can be called by other workflows. ➡️ Quick Start & Input/Output Examples: Look for the Sticky Notes within the workflow canvas! They provide crucial information, including examples of the required input JSON structure and the expected output format. How to customize this workflow Input Source:* The core workflow is designed to be called. You will build *another workflow that prepares the anthropic-version and requests array and then uses the 'Execute Workflow' node to trigger this template. The included example branch shows how to prepare this data. Model Selection & Parameters:* Model (claude-3-opus-20240229, etc.), max_tokens, temperature, and other parameters are defined *within each object inside the requests array you pass to the workflow trigger. You configure these in the workflow calling this template. Polling Interval:** Modify the 'Wait' node ('Batch Status Poll Interval') duration if you need faster or slower status checks (default is 10 seconds). Be mindful of potential rate limits. Parsing Logic:** If Anthropic changes the result format or you have specific needs, modify the Javascript code within the 'Parse response' Code node. Error Handling:** Enhance the workflow with more specific error handling for API failures (e.g., using 'Error Trigger' or checking HTTP status codes) or batch processing issues (batch.status === 'failed'). Output Processing:* In the workflow that *calls this template, add nodes after the 'Execute Workflow' node to process the individual result items returned (e.g., save to a database, spreadsheet, send notifications). Example Usage Branch (Manual Trigger) This template also contains a separate branch starting with the Run example Manual Trigger node. Purpose:** This branch demonstrates how to construct the necessary anthropic-version and requests array payload. Methods Shown:** It includes steps for: Creating a request object from a simple query string. Creating a request object using data from n8n's Langchain Chat Memory nodes (@n8n/n8n-nodes-langchain). Execution:** It merges these examples, constructs the final payload, and then uses the Execute Workflow node to call the main batch processing logic described above. It finishes by filtering the results for demonstration. Note:** This branch is for demonstration and testing. You would typically build your own data preparation logic in a separate workflow. The use of Langchain nodes is optional for the core batch functionality. Notes API Limits:** According to the Anthropic API documentation, batches can contain up to 100,000 requests and be up to 256 MB in total size. Ensure your n8n instance has sufficient resources for large batches. API Costs:** Using the Anthropic API, including the Batch API, incurs costs based on token usage. Monitor your usage via the Anthropic dashboard. Completion Time:** Batch processing time depends on the number and complexity of prompts and current API load. The polling mechanism accounts for this variability. Versioning:** Always include the anthropic-version header in your requests, as shown in the workflow and examples. Refer to Anthropic API versioning documentation.
by Niranjan G
This workflow leverages AI to intelligently analyze incoming Gmail messages and automatically apply relevant labels based on the email content. The default configuration includes the following labels: Newsletter**: Subscription updates or promotional content. Inquiry**: Emails requesting information or responses. Invoice**: Billing and payment-related emails. Proposal**: Business offers or collaboration opportunities. Action Required**: Emails demanding immediate tasks or actions. Follow-up Reminder**: Emails prompting follow-up actions. Task**: Emails containing actionable tasks. Personal**: Non-work-related emails. Urgent**: Time-sensitive or critical communications. Bank**: Banking alerts and financial statements. Job Update**: Recruitment or job-related communications. Spam/Junk**: Unwanted or irrelevant bulk emails. Social/Networking**: Notifications from social platforms. Receipt**: Purchase confirmations and receipts. Event Invite**: Invitations or calendar-related messages. Subscription Renewal**: Reminders for subscription expirations. System Notification**: Technical alerts from services or systems. You can customize labels and definitions based on your specific use case. How it works: The workflow periodically retrieves new Gmail messages. Only emails without existing labels, regardless of read status, are sent to the AI for analysis. Email content (subject and body) is analyzed by an AI model to determine the appropriate label. Labels identified by the AI are applied to each email accordingly. Note: This workflow performs 100% better than the default Gmail trigger method, which is why the workflow was switched from Gmail trigger to a scheduled workflow. By selectively processing only unlabeled emails, it ensures comprehensive labeling while significantly reducing AI processing costs. Setup Steps: Configure credentials for Gmail and your chosen AI service (e.g., OpenAI). Ensure labels exist in your Gmail account matching the workflow definitions. Adjust the AI prompt to match your labeling needs. Optionally customize the polling interval (default: every 2 minutes). This workflow streamlines your email management, keeping your inbox organized effortlessly while optimizing resource usage.
by Dr. Firas
WhatsApp AI Agent: Auto-Train Product Data & Handle Customer Support Who Is This For This workflow is ideal for eCommerce founders, product managers, customer support teams, and automation builders who rely on WhatsApp to manage product information and interact with clients. It’s perfect for businesses that want to automate product data entry and support responses directly from WhatsApp messages using GPT-4 and Google Sheets. What Problem Does This Workflow Solve Manual Product Data Entry**: Collecting and organizing product data from links is tedious and error-prone. Slow Customer Response Times**: Responding to client questions manually leads to delays and inconsistent support. No Logging System for Issues**: Without automation, support issues often go undocumented, making it harder to learn and improve. What This Workflow Does Step 1 – Incoming Message Detection Listens for incoming messages via WhatsApp. If the message starts with train:, it routes to the product training process. Otherwise, it routes to the customer support assistant. Step 2 – Product Data Training Extracts URL** from the message using a regex script. Fetches HTML content** from the URL. Cleans HTML data** to extract readable product description. Saves raw data** (URL + description) into Google Sheets. Uses GPT-4** to enhance product data: → Name, price (one-time or subscription), topic, and FAQs. Updates the product row** in Google Sheets with structured information. Step 3 – Customer Support Flow Analyzes user messages with GPT-4 to understand the request or issue. Looks up relevant product info in Google Sheets. Detects potential problems (e.g. payment, login, delivery). Suggests an appropriate solution. Logs the problem, solution, and category to the Customer Issues sheet. Sends a response back to the client via WhatsApp. Step 4 – Client Response Sends the AI-generated response to the client via WhatsApp. Keeps the communication fast, clear, and professional. Setup Guide Prerequisites WhatsApp Business API access** OpenAI API Key** Google Account** with Google Sheets access A hosted instance of n8n (Cloud or self-hosted) Setup Steps Import the Workflow into your n8n instance. Connect your credentials for WhatsApp, OpenAI, and Google Sheets. Customize Google Sheet IDs and names as needed. Test by sending a train: message or a regular customer message to WhatsApp. Activate the workflow to make it live. How to Customize This Workflow Edit AI prompts** to reflect your product type, language style, or tone. Change the trigger keyword** (e.g. from train: to add: or anything else). Add integrations** like Notion, Airtable, or CRM tools. Expand the Sheets structure** with more product fields (e.g. stock status, image link). Add notifications** to Slack or email after product updates or issue logging. 📄 Documentation: Notion Guide Need help customizing? Contact me for consulting and support : Linkedin / Youtube
by Evoort Solutions
🚀 AI-Powered LinkedIn Post Automation 🧩 How It Works This workflow automatically generates LinkedIn posts based on a user-submitted topic, including both content creation and image generation, then publishes the post to LinkedIn. Ideal for marketers, content creators, or businesses looking to streamline their social media activity, without the need for manual post creation. High-Level Workflow: Trigger: The workflow is triggered when a user submits a form with a topic for the LinkedIn post. Data Mapping: The topic is mapped and prepared for the AI model. AI Content Generation: Calls the Google Gemini AI model to generate engaging post content and a visual image prompt. Image Creation: Sends the image prompt to the external API, gen-imager, to generate a professional image matching the topic. Post Creation: Publishes the text and image to LinkedIn, automatically updating the user's feed. ⚙️ Set Up Steps (Quick Overview) 🕐 Estimated Setup Time: ~10–20 minutes Connect Google Gemini: Set up your Google Gemini API credentials to interact with the AI model for content creation. Set Up External Image API: Configure the external image generation API (gen-imager API) for visual creation based on the post prompt. Connect LinkedIn: Set up OAuth2 credentials to authenticate your LinkedIn account and allow publishing posts. Form Submission Setup: Create a simple web form for users to submit the topic for LinkedIn posts. Activate the Workflow: Once everything is connected, activate the workflow. It will trigger automatically upon receiving form submissions. 💡 Important Notes: The flow uses Google Gemini (PaLM) for generating content based on the user's topic. Text to Image: The image generation process involves creating a professional, LinkedIn-appropriate image based on the post’s topic using the **gen-imager API. You can customize the visual elements of the posts and adjust the tone of the generated content based on preferences. 🛠 Detailed Node Breakdown: On Form Submission Trigger: Captures the user-submitted topic and initializes the workflow. Action: Start the process by gathering the topic information. Mapper (Field Mapping) Action: Maps the captured topic to a variable that is passed along for content generation. AI Agent (Content Generation) Action: Calls Google Gemini to generate professional LinkedIn post content and an image prompt based on the submitted topic. Key: Outputs content in a structured form — post text and image prompt. Google Gemini Chat Model Action: AI model that generates actionable insights, engaging copy, and an image prompt for LinkedIn post. Normalizer (Data Cleanup) Action: Cleans the output from the AI model to ensure the content and image prompt are correctly formatted for use in the next steps. Text to Image (Image Generation) Action: Sends the image prompt to the gen-imager API, which returns a custom image based on the post's topic. Decoder (Base64 Decoding) Action: Decodes the image from base64 format for easier uploading to LinkedIn. LinkedIn (Post Creation) Action: Publishes the generated text and image to LinkedIn, automatically creating a polished post for the user’s feed. ⏱ Execution Time Breakdown: Total Estimated Execution Time**: ~15–40 seconds per workflow run. On Form Submission: Instant (Trigger) Mapper (Field Mapping): ~1–2 seconds AI Content Generation: ~5–10 seconds (depending on server load) Text to Image: ~5–15 seconds (depends on external API) LinkedIn Post Creation: ~2–5 seconds 🚀 Ready to Get Started? Let’s get you started with automating your LinkedIn posts! Create your free n8n account and set up the workflow using this link. 📝 Notes & Customizations Form Fields**: Customize the form to gather more specific information for the LinkedIn posts (like audience targeting, post category, etc.). Image API Customization**: Adjust the image generation prompt to fit your brand’s style, or change the color palette as needed. Content Tone**: The tone can be adjusted by modifying the system message sent to Google Gemini for content generation.
by Muhammad Ashar
How It Works – Your AI Marketing Team in Action This automation acts as your AI-powered content and image marketing assistant inside Telegram. With just a voice note or text message, it can: 🧠 Understand your request – Whether you send a message or speak into Telegram, it transcribes and processes your input using GPT-4. 🎨 Create and edit content – Based on what you say, it can generate: ✍️ Blog posts 💼 LinkedIn posts 🎬 Faceless videos 🖼️ AI-generated images 🪄 Edits to existing images 🔎 Searches through your image database 💬 Replies directly in Telegram – It sends you back the result—whether that’s a post, image, or video link—without leaving the app. 🧩 Built using LangChain agent logic – It intelligently chooses the right tool from a suite of sub-workflows like "Create Image", "Blog Post", or "Video" using agent reasoning. 🛠️ Setup Steps – Get Started in Minutes! ⌛ Time Estimate: ~15–30 minutes (faster if you're familiar with n8n) 🔗 1. Import the Template Pack 📥 Download and install these workflows into your n8n: Create Image, Edit Image, Search Images Blog Post, LinkedIn Post, Video 🔐 2. Add Required Credentials Telegram Bot 🤖 OpenRouter AI 🧠 Tavily API (for smart research) 📚 ElevenLabs 🎙️ (for voice in videos) PiAPI & Runway 🎞️ (for faceless videos) 🧩 3. Link the Tools to the Agent Node – Make sure the "Marketing Team Agent" is connected to each of the content creation tools as shown in the workflow. 📎 4. Download Templates & Logs 🧾 Google Sheets Log Template (to track output) 🖼️ Creatomate Template (optional for enhanced image control – shared in Skool group) 📌 Pro Tip: All detailed step-by-step setup instructions are included as sticky notes inside the n8n canvas. Just follow along!
by Don Jayamaha Jr
📊 This AI sub-agent aggregates Tesla (TSLA) trading signals across multiple timeframes using real-time technical indicators and candlestick behavior. It is a core component of the Tesla Quant Trading AI system. Powered by GPT-4.1, it consolidates 15-minute, 1-hour, and 1-day indicators, adds candlestick pattern data, and produces a unified JSON signal for downstream use by the master agent. ⚠️ This agent is not standalone. It is triggered by the Tesla Quant Trading AI Agent via Execute Workflow. 🧠 Requires: 4 connected sub-agents and Alpha Vantage Premium API Key 🔌 Required Sub-Workflows To use this workflow, you must install: Tesla 15min Indicators Tool Tesla 1hour Indicators Tool Tesla 1day Indicators Tool Tesla 1hour and 1day Klines Tool Tesla Quant Technical Indicators Webhooks Tool (provides Alpha Vantage data) 🧠 What This Agent Does Fetches pre-cleaned 20-point JSON outputs from the 4 sub-agents listed above Analyzes each timeframe individually: 15m: momentum and short-term setups 1h: confirmation of emerging trends 1d: macro positioning and trend alignment Klines: candlestick reversal patterns and volume divergence Generates a structured final signal in JSON with: Trading stance: Buy, Sell, Hold, or Cautious Confidence score (0.0–1.0) Multi-timeframe indicator breakdown Candlestick and volume divergence annotations 📋 Sample Output { "summary": "TSLA momentum is weakening short-term. 1h MACD shows bearish crossover, RSI declining. 1d candles confirm potential reversal setup.", "signal": "Cautious Sell", "confidence": 0.81, "multiTimeframeInsights": { "15m": { "RSI": 68.3, "MACD": { "macd": 0.53, "signal": 0.61 }, ... }, "1h": { "RSI": 65.0, "MACD": { "macd": -0.32, "signal": 0.11 }, ... }, "1d": { "BBANDS": { ... }, ... }, "candlestickPatterns": { "1h": "Doji", "1d": "Bearish Engulfing" }, "volumeDivergence": { "1h": "Bearish", "1d": "Neutral" } } } 🛠️ Setup Instructions Import this workflow into n8n Name it: Tesla_Financial_Market_Data_Analyst_Tool Add Required API Credentials Alpha Vantage Premium (via HTTP Query Auth) OpenAI GPT-4.1 for reasoning and synthesis Link Required Sub-Agents Connect the 4 tool workflows listed above to their respective Tool Workflow nodes Connect the webhook provider for data fetches Set Up as Sub-Agent This workflow must be triggered using Execute Workflow from the parent agent Pass in: message (optional context) sessionId (used for memory continuity) 🧾 Sticky Notes Provided 📘 Tesla Financial Market Data Analyst — Core logic overview 📈 15m / 1h / 1d Tool Notes — Indicator lists + use cases 🕯️ Klines Tool Note — Candlestick and volume divergence patterns 🧠 GPT Reasoning Note — GPT-4.1 handles final synthesis 🧩 Sub-Workflow Trigger — Proper integration with parent agent 🧠 Memory Buffer — Maintains session context across evaluations 🔒 Licensing & Support © 2025 Treasurium Capital Limited Company The logic, prompt design, and multi-agent architecture are proprietary and IP-protected. For support or collaboration inquiries: 🔗 Don Jayamaha – LinkedIn 🔗 n8n Creator Profile 🚀 Unify your Tesla trading logic across timeframes—automated, AI-powered, and built for scalers and swing traders.
by Saswat Saubhagya Rout
📝 Use Case This n8n workflow automates the creation and publication of technical blog posts based on a list of topics stored in Google Sheets. It fetches context using Tavily and Wikipedia, generates Markdown-formatted content with Gemini AI, commits it to a GitHub repository, and updates a Jekyll-powered blog — all without manual intervention. Ideal for developers, bloggers, or content teams who want to streamline technical content creation and publishing. ⚙️ Setup Instructions 🔑 Prerequisites n8n (cloud or self-hosted) Tavily API key Google Sheets with blog topics Gemini (Google Palm) API key GitHub repository (Jekyll enabled) GitHub OAuth2 credentials Google OAuth2 credentials 🧩 Setup Steps Import the workflow JSON into your n8n instance. Set up the following credentials in n8n: Tavily API Google Sheets OAuth2 Google Palm/Gemini AI GitHub OAuth2 Prepare your Google Sheet: Columns: Title, status, row_number Set status to blank for topics to be picked up. Configure: GitHub repo and _posts/ path Jekyll setup (front matter, _config.yml, GitHub Pages) Adjust prompt/custom parameters if needed. Enable and deploy the workflow. Schedule it daily or trigger manually. 🔄 Workflow Details | Node | Function | |------|----------| | Schedule Trigger | Triggers the flow at a set interval | | Google Sheets (Get Topic) | Fetches the next incomplete blog topic | | Extract Topic | Parses topic text from the sheet | | Tavily Search | Gathers up-to-date content related to the topic | | Wikipedia Tool | Optionally adds more context or images | | Summarize Results | Formats the context for the AI | | Gemini AI Agent (LangChain) | Generates a Markdown blog post with YAML front matter | | Set File Parameters | Prepares the filename, content, and commit message | | GitHub Commit | Uploads the .md file to the _posts/ directory | | Update Google Sheet | Marks topic as done after successful commit | 🛠️ Customization Options Change LLM prompt (e.g. tone, depth, format). Use OpenAI instead of Gemini by switching nodes. Modify filename pattern or GitHub repo path. Add Slack/Discord notifications after publish. Extend flow to upload images or embed YouTube links. ⚠️ Community Nodes Used This workflow uses the following community nodes: @tavily/n8n-nodes-tavily.tavily – for deep search > ⚠️ Ensure these are installed and enabled in your n8n instance. 💡 Pro Tips Use GitHub Actions to trigger an automatic Jekyll build post-commit. Structure blog posts with front matter, headings, and table of contents for SEO. Set Schedule Trigger to daily at a fixed time to keep content flowing. Enhance formatting in AI output using code blocks, images, and lists. ✅ Example Output title: "How LLMs Are Changing Web Development" date: "2025-07-25" categories: [webdev, AI] tags: [LLM, Gemini, n8n, automation] excerpt: "Learn how LLMs like Gemini are transforming how we generate and deploy developer content." author: "Saswat Saubhagya" Table of Contents Introduction Understanding LLMs Use Cases in Web Development Challenges Conclusion ...