by DigiMetaLab
A reasoning agent that can think, search, calculate, and remember — powered by GROQ inference and ready to deploy in one click. Unlike traditional AI bots that only respond, this assistant reasons before replying, fetches real-time facts, does math, and keeps short-term memory of your conversation. 🔧 How it works This template builds a conversational AI agent using the GROQ LLaMA 3 or LLaMA 4 API, combined with modular tools like: 🧠 Think Tool – performs step-by-step logical reasoning 🔍 SerpAPI – fetches live data from Google search ➗ Calculator – handles arithmetic and math queries 💾 Memory Buffer – keeps track of the last 5 messages for context Everything is integrated inside n8n and optimized for blazing-fast replies using GROQ’s ultra-low latency. 🧠 Your Agent Will: Understand and analyze your queries Think through solutions before answering Pull real-time data via SerpAPI Perform calculations with the built-in math engine Recall prior context using short-term memory Respond clearly, conversationally — like a real assistant 🧑💼 Who is this template for? Perfect for: AI builders and creators using GROQ + n8n Teams needing a real-time LLaMA-powered assistant Beginners exploring LangChain + n8n workflows Developers combining LLMs + tools + memory 🚀 How to Set Up Plug in your GROQ API key Add your SerpAPI key Import and run — it’s ready to chat! All tools are pre-wired. You can expand the memory, customize prompts, or plug in more tools. 📬 Use Cases Connect this agent with: Telegram Bots 🤖 WhatsApp via Twilio 📱 Slack, Discord, or Gmail 💬 Manual triggers in n8n 🔁 👉 Check out more templates by this creator: https://n8n.io/creators/digimetalab
by Yaron Been
LinkedIn AI Agent: Auto-Post Creator & Multi-Group Distributor Transform simple topic ideas into engaging LinkedIn posts and automatically distribute them across your profile and multiple LinkedIn groups. This powerful n8n workflow combines AI content generation with intelligent distribution, helping you maintain a consistent professional presence while maximizing your reach across relevant communities. 🚀 How It Works This sophisticated 6-step automation turns content ideas into LinkedIn success: Step 1: Smart Content Monitoring The workflow continuously monitors your Google Sheets for new post topics marked as "Pending", checking every minute for fresh content to process. Step 2: AI-Powered Content Generation GPT-4 transforms your basic topic into a professionally crafted LinkedIn post featuring: Compelling opening hooks that grab attention 3-4 informative paragraphs with valuable insights Strategic questions to encourage engagement 4-6 relevant hashtags for discoverability Professional emoji placement for visual appeal Optimized formatting for LinkedIn's platform Step 3: Professional Formatting The workflow ensures your content meets LinkedIn's technical requirements with proper JSON formatting, character limits, and special character handling. Step 4: Personal Profile Publishing Your generated post is automatically published to your personal LinkedIn profile, maintaining your professional brand presence. Step 5: Multi-Group Distribution The same content is intelligently distributed across all your specified LinkedIn groups, amplifying your reach to targeted professional communities. Step 6: Status Management The workflow automatically updates your Google Sheets to mark posts as "Posted", providing clear tracking of your content pipeline. ⚙️ Setup Steps Prerequisites Active LinkedIn account with API access Google Sheets access for content management OpenAI API key with GPT-4 access LinkedIn group memberships with posting permissions n8n instance (cloud or self-hosted) Required Google Sheets Structure Sheet 1 - Main Content: | ID | LinkedIn Post Title | Status | |----|-------------------|--------| | 1 | AI Trends in 2024 | Pending | | 2 | Remote Work Tips | Posted | Sheet 2 - Groups: | GroupIds | |-------------| | 123456789 | | 987654321 | | 456789123 | Note: Collect LinkedIn group IDs from groups where you have posting permissions. These can be found in the group URL or through LinkedIn's API. Configuration Steps Credential Setup Google Sheets OAuth2: Access your content spreadsheet OpenAI API Key: Required for AI content generation LinkedIn OAuth2: Enable profile and group posting HTTP Authentication: Configure LinkedIn API headers Google Sheets Preparation Create spreadsheet with the required two-sheet structure Populate group IDs from your joined LinkedIn groups Add initial post topics with "Pending" status Ensure proper column naming and data types LinkedIn Group Setup Join relevant professional LinkedIn groups Verify posting permissions in each group Collect group IDs using LinkedIn's interface or API Test posting permissions before full automation AI Content Customization The default prompt generates professional LinkedIn content, but can be customized for: Industry-specific terminology and trends Company voice and brand guidelines Target audience preferences Content style (educational, promotional, thought leadership) Workflow Activation Import the workflow JSON into your n8n instance Configure all credential connections Test with sample content before going live Activate the Google Sheets trigger 🎯 Use Cases Content Creators & Influencers Consistent Posting: Maintain regular LinkedIn presence without daily manual work Audience Growth: Reach multiple professional communities simultaneously Content Scaling: Transform brief ideas into full-length engaging posts Brand Building: Establish thought leadership across industry groups Marketing Teams Lead Generation: Share valuable content across targeted professional groups Brand Awareness: Increase visibility in relevant industry communities Thought Leadership: Position company experts as industry authorities Content Distribution: Maximize reach of marketing messages and insights Sales Professionals Pipeline Building: Share insights that attract potential clients Network Expansion: Engage with prospects across multiple professional groups Authority Building: Establish credibility through valuable content sharing Relationship Nurturing: Maintain visibility with existing connections Consultants & Freelancers Client Acquisition: Demonstrate expertise to potential clients Professional Branding: Build reputation across industry-specific groups Service Promotion: Share case studies and success stories broadly Network Building: Connect with peers and potential collaborators Business Leaders & Executives Industry Influence: Share strategic insights across professional networks Talent Attraction: Showcase company culture and opportunities Partnership Development: Connect with potential business partners Market Education: Share expertise to influence industry conversations 🔧 Advanced Customization Options Content Strategy Enhancement Multi-Tone Generation: Create different content styles for various audiences Industry Templates: Pre-built prompts for specific professional sectors Engagement Optimization: A/B testing different post formats and styles Content Calendar Integration: Schedule posts for optimal timing Distribution Intelligence Group Performance Tracking: Monitor engagement across different groups Selective Distribution: Choose specific groups based on content type Audience Segmentation: Tailor content for different professional communities Engagement Analysis: Track which groups provide best ROI Content Quality Control Approval Workflows: Add human review before automatic posting Content Scoring: Rate post quality before distribution Brand Compliance: Ensure posts meet company guidelines Duplicate Detection: Avoid posting similar content too frequently Extended Integration Options CRM Integration: Track leads generated from LinkedIn posts Analytics Platforms: Monitor engagement and conversion metrics Content Management: Connect to existing content planning tools Social Media Expansion: Extend to other professional platforms 📊 Content Generation Features AI Writing Capabilities The workflow generates LinkedIn posts that include: Professional Structure: Attention-grabbing opening statements Well-organized multi-paragraph content Clear value propositions and insights Strategic calls-to-action for engagement LinkedIn Optimization: Platform-specific formatting and spacing Proper hashtag research and placement Emoji usage that enhances readability Character count optimization for maximum impact Engagement Drivers: Thought-provoking questions to encourage comments Industry insights that spark discussions Personal anecdotes that build connections Actionable tips that provide immediate value Sample Output Input Topic: "Remote Work Productivity Tips" Generated Post: 🏠 Working from home has taught me that productivity isn't about being busy—it's about being intentional. After managing remote teams for 3 years, I've discovered that the most successful remote workers share three key habits that transform their daily performance. First, they create physical boundaries even in small spaces. A dedicated workspace signals to your brain that it's time to focus, even if it's just a corner of your kitchen table. Second, they batch similar tasks together. Instead of jumping between emails, calls, and deep work, they group activities to maintain mental flow and reduce cognitive switching costs. Third, they communicate proactively rather than reactively. They share progress updates before being asked and clarify expectations upfront to avoid confusion later. What's your most effective remote work strategy? I'd love to hear what's working for your team! 💪 #RemoteWork #Productivity #WorkFromHome #Leadership #TeamManagement #ProfessionalDevelopment 🛠️ Troubleshooting & Best Practices Common Issues & Solutions LinkedIn API Limitations Respect posting frequency limits to avoid account restrictions Monitor API usage and implement appropriate delays between posts Ensure compliance with LinkedIn's terms of service Maintain authentic engagement rather than purely automated interactions Group Posting Permissions Verify membership status and posting rights before adding group IDs Some groups require administrator approval for posts Monitor group rules and community guidelines Remove inactive or restricted groups from your list Content Quality Control Review AI-generated content periodically for brand consistency Adjust prompts based on engagement performance Maintain a balance between automation and personal touch Monitor comments and engage authentically with your audience Optimization Strategies Performance Enhancement Track engagement metrics across different groups A/B test posting times and content formats Refine prompts based on successful post patterns Gradually expand to new groups based on performance Content Strategy Develop content themes that resonate with your target audience Create series of related posts for deeper engagement Balance promotional content with value-driven insights Maintain consistency in voice and messaging Network Growth Engage with comments on your automated posts Connect with active commenters to expand your network Participate in group discussions beyond your own posts Build genuine relationships through authentic interactions 📈 Success Metrics Engagement Indicators Post Reach: Total views across profile and groups Interaction Rate: Comments, likes, and shares per post Network Growth: New connections from content engagement Group Performance: Which communities provide best engagement Business Impact Measurements Lead Generation: Connections and inquiries from LinkedIn posts Brand Awareness: Mentions and sharing of your content Thought Leadership: Recognition as industry expert Professional Opportunities: Speaking, collaboration, or job opportunities 📞 Questions & Support Need help setting up or optimizing your LinkedIn AI Agent workflow? 📧 Direct Technical Support Email: Yaron@nofluff.online Response Time: Within 24 hours on business days Expertise: LinkedIn API integration, AI prompt optimization, workflow scaling 🎥 Comprehensive Learning Resources YouTube Channel: https://www.youtube.com/@YaronBeen/videos Complete setup walkthrough and configuration Advanced customization techniques and strategies LinkedIn API best practices and limitations Content strategy optimization for maximum engagement Troubleshooting common integration issues 🤝 Professional Networking & Updates LinkedIn: https://www.linkedin.com/in/yaronbeen/ Connect for ongoing automation support and advice Share your LinkedIn growth success stories Get early access to new workflow templates and features Join discussions about LinkedIn marketing automation 💬 Support Request Guidelines Include in your support message: Your current LinkedIn strategy and goals Target audience and industry focus Specific LinkedIn groups you want to target Any technical errors or integration issues Current content creation process and pain points
by Batu Öztürk
Extract the main idea and key takeaways from YouTube videos and turn them into Airtable content ideas 📝 Description Automatically turn YouTube videos into clear, structured content ideas stored in Airtable. This workflow pulls new video links from Airtable, extracts transcripts using a RapidAPI service, summarizes them with your favourite LLM, and logs the main idea and key takeaways—keeping your content pipeline fresh with minimal effort. ⚙️ What It Does Scans Airtable for new YouTube video links every 5 minutes. Extracts the transcript of the video using a third-party API via RapidAPI. Summarizes the content to generate a main idea and takeaways. Updates the original Airtable entry with the insights and marks it as completed. 🛠 Prerequisites Before using this template, make sure you have: ✅ A RapidAPI account with access to the youtube-video-summarizer-gpt-ai API. ✅ A valid RapidAPI key. ✅ An OpenAI, Claude or Gemini account connected to n8n. ✅ An Airtable account with a base and table ready. 🧰 Setup Instructions Clone this template into your n8n workspace. Open the Get YouTube Sources node and configure your Airtable credentials. In the Get video transcript node: Enter your X-RapidAPI-Key under headers. The API endpoint is pre-configured. Connect your LLM credentials to the Extract detailed summary node. (Optional) Adjust the summarization prompt in the LangChain node to better suit your tone. Set your preferred schedule in the Trigger node. 📋 Airtable Setup Create a base (e.g., Content Hub) with a table named Ideas and the following columns: | Column Name | Type | Required | Notes | |-------------|------------|----------|----------------------------| | Type | Single select | ✅ | Must be set to Youtube Video | | Source | URL | ✅ | The YouTube video URL | | Status | Checkbox | ✅ | Leave empty initially; updated after processing | | MainIdea | Single line text | ✅ | Summary generated by OpenAI | | Key Takeaways | Long text | ✅ | List of takeaways extracted from the transcript Activate the workflow—and you're done!
by Jimleuk
This n8n workflow demonstrates a simple approach to improve chat UX by staggering an AI Agent's reply for users who send in a sequence of partial messages and in short bursts. How it works Twilio webhook receives user's messages which are recorded in a message stack powered by Redis. The execution is immediately paused for 5 seconds and then another check is done against the message stack for the latest message. The purpose of this check lets use know if the user is sending more messages or if they are waiting for a reply. The execution is aborted if the latest message on the stack differs from the incoming message and continues if they are the same. For the latter, the agent receives the buffered messages up to that point and is able to respond to them in a single reply. Requirements A Twilio account and SMS-enabled phone number to receive messages. Redis instance for the messages stack. OpenAI account for the language model. Customising the workflow This workflow should work for other common messaging platforms such as Whatsapp and Telegram. 5 seconds too long or too short? Adjust the wait threshold to suit your customers.
by Zain Ali
🧠 Email real time RAG Assistant with Gmail, OpenAI & PGVector 📌 Who’s it for This workflow is ideal for: Professionals Project managers Sales and support teams Anyone managing high volumes of Gmail messages It enables fast and intelligent search through your email inbox using natural language queries. ⚙️ How it works / What it does Continuously monitors your Gmail inbox for new emails. Extracts email content and metadata (subject, body, sender, date). Converts email content into vector embeddings using OpenAI. Stores embeddings in a PostgreSQL database with PGVector. A conversational AI agent performs semantic search on your stored email history. Supports time-sensitive and context-aware responses via OpenAI Chat model. 🚀 How to set up Connect your Gmail account to the Gmail Trigger node (with API access enabled). Configure OpenAI credentials for the Embedding and Chat nodes. Set up a PostgreSQL database with the PGVector extension enabled. Import the workflow into your n8n instance (Cloud or Self-hosted). Customize parameters like polling frequency, embedding settings, or vector query depth. 📋 Requirements ✅ n8n instance (Self-hosted or Cloud) ✅ Gmail account with API access ✅ OpenAI API Key ✅ PostgreSQL database with PGVector extension installed 🛠️ How to customize the workflow Email Filtering**: Change filters in the Gmail Trigger to watch specific labels or senders. Text Splitting Granularity**: Adjust chunkSize and chunkOverlap in the text splitter node. Query Depth**: Modify topK in the vector search node to retrieve more or fewer similar results. Prompt Tuning**: Customize the system message or agent instructions in the RAG node. Workflow Extensions**: Add notifications, error logging, Slack/Telegram alerts, or data exports.
by Thomas Janssen
Build a 100% local RAG with n8n, Ollama and Qdrant. This agent uses a semantic database (Qdrant) to answer questions about PDF files. Tutorial Click here to view the YouTube Tutorial How it works Build a chatbot that answers based on documents you provide it (Retrieval Augmented Generation). You can upload as many PDF files as you want to the Qdrant database. The chatbot will use its retrieval tool to fetch the chunks and use them to answer questions. Installation Install n8n + Ollama + Qdrant using the Self-hosted AI starter kit Make sure to install Llama 3.2 and mxbai-embed-large as embeddings model. How to use it First run the "Data Ingestion" part and upload as many PDF files as you want Run the Chatbot and start asking questions about the documents you uploaded
by InfraNodus
Set Up ElevenLabs Voice Chat Agent using Graph RAG Knowledge Graphs as Experts This workflow creates an AI voice chatbot agent that has access to several knowledge bases at the same time (used as "experts"). These knowledge bases are provided using the InfraNodus GraphRAG using the knowledge graphs and providing high-quality responses without the need to set up complex RAG vector store workflows. We use ElevenLabs to set up a voice agent that can be embedded to any website or used via their API. The advantages of using GraphRAG instead of the standard vector stores for knowledge are: Easy and quick to set up (no complex data import workflows needed) and to update with new knowledge A knowledge graph has a holistic overview of your knowledge base Better retrieval of relations between the document chunks = higher quality responses Ability to reuse in other n8n workflows How it works This template uses the n8n AI agent node as an orchestrating agent that decides which tool (knowledge graph) to use based on the user's prompt. The user's prompt is received from the ElevenLabs Conversational AI agent via an n8n Webhook, which also takes care of the voice interaction. The response from n8n is then sent to the Webhook, which is polled by the ElevenLabs voice agent. This agent processes the response and provides the final answer. Here's a description step by step: The user submits a question using ElevenLabs voice interface The question is sent via the knowledge_base tool in ElevenLabs to the n8n Webhook with the POST request containing the user's prompt and sessionID for Chat Memory node in n8n. The n8n AI agent node checks a list of tools it has access to. Each tool has a description of the knowledge auto-generated by InfraNodus (we call each tool an "expert"). The n8n AI agent decides which tool should be used to generate a response. It may reformulate user's query to be more suitable for the expert. The query is then sent to the InfraNodus HTTP node endpoint, which will query the graph that corresponds to that expert. Each InfraNodus GraphRAG expert provides a rich response that takes the whole context into account and provides a response from each expert (graph) along with a list of relevant statements retrieved using a combination or RAG and GraphRAG. The n8n AI Agent node integrates the responses received from the experts to produce the final answer. The final answer is sent back to the Webhook endpoint ElevenLabs conversational AI agent picks up the response arriving from the knowledge_base tool via the webhook and then condenses it for conversational format and transforms text into voice. How to use You need an InfraNodus GraphRAG API account and key to use this workflow. Create an InfraNodus account Get the API key at https://infranodus.com/api-access and create a Bearer authorization key for the InfraNodus HTTP nodes. Create a separate knowledge graph for each expert (using PDF / content import options) in InfraNodus For each graph, go to the workflow, paste the name of the graph into the body name field. Keep other settings intact or learn more about them at the InfraNodus access points page. Once you add one or more graphs as experts to your flow, add the LLM key to the OpenAI node and launch the workflow You will also need to set up an ElevenLabs account and to set up a conversational AI agent there. See the Post note in the n8n workflow for a complete step-by-step description or our support article on setting up ElevenLabs AI voice agent Once the voice AI agent is ready, you might want to combine it with a text AI chatbot workflow so your users have a choice between the text and voice interaction. In that case, you may be interested to use our free open-source website popup chat widget popupchat.dev where you can create an embed code to add to your blog or website and allow the user to choose between the text and voice interaction. Requirements An InfraNodus account and API key An OpenAI (or any other LLM) API key An ElevenLabs account FAQ 1. How many "experts" should I aim for? We recommend to aim for the number of experts as the optimal number of people in a team, which is usually 2-7. If you add more experts, your AI orchestrating agent will have troubles choosing the most suitable "expert" tool for the user's query. You can mitigate this by specifying in the AI agent description that it can choose maximum 3-7 experts to provide a response. 2. Why use InfraNodus GraphRAG and not standard vector store for knowledge? First, vector stores are complex to set up and to update. You'd need a separate workflow for that, decide on the vector dimensions, add metadata to your knowledge, etc. With InfraNodus, you have a complete RAG / GraphRAG solution under the hood that is easy to set up and provides high-quality responses that takes the overall structure and the relations between your ideas into account. 3 Why not use ElevenLabs' own knowledge? One of the reasons is that you want your knowledge base to be in one place so you can reuse it in other n8n workflows. Another reason is that you will not have such a good separation between the "experts" when you converse with the agent. So the answers you get will be based on top matches from all the books / articles you upload, while with the InfraNodus GraphRAG setup you can better control which graphs are consulted as experts and have an explicit way to display this data. Customizing this workflow You can use this same workflow with a Telegram bot, so you can interact with it using Telegram. There are many more customizations available on our GitHub repo for n8n workflows. Check out the complete setup guide for this workflow at https://support.noduslabs.com/hc/en-us/articles/20318967066396-How-to-Build-a-Text-Voice-AI-Agent-Chatbot-with-n8n-Elevenlabs-and-InfraNodus Also check out the video tutorial with a demo:
by Nick Saraev
AI Upwork Application Agent with OpenAI & Google Docs Categories: AI Agents, Freelance Automation, Proposal Generation This workflow creates an intelligent AI agent that automates Upwork job applications by generating highly personalized proposals, professional Google Doc presentations, and visual workflow diagrams. Built by someone who earned over $500,000 on Upwork, this system demonstrates the exact templates and strategies that achieve superior response rates through perceived customization and value demonstration. Benefits Complete Application Automation** - Transform job descriptions into custom proposals, documents, and diagrams in minutes Proven Templates** - Based on $500K+ in Upwork earnings using exact strategies for high-converting applications Intelligent Personalization** - AI analyzes job requirements and customizes responses with relevant social proof Professional Asset Generation** - Creates Google Doc proposals and Mermaid workflow diagrams for enhanced perceived value Modular Architecture** - Three specialized sub-workflows handle different aspects of proposal generation High Response Rates** - Focuses on perceived customization and value demonstration over generic applications How It Works AI Agent Orchestration: Receives Upwork job descriptions through chat interface Maintains conversation context with window buffer memory Coordinates three specialized sub-workflows for comprehensive proposal generation Automatically integrates generated assets into cohesive application packages Application Copy Generation: Uses proven templates based on $500K+ Upwork success Follows structure: "Hi, I do [thing] all the time. So confident I created a demo: [link]" Incorporates personal social proof and achievements automatically Generates concise, spartan-toned applications that avoid generic AI language Google Doc Proposal Creation: Copies professional proposal template from Google Drive Generates structured content including system title, explanation, scope, and timeline Uses find-and-replace to populate template with AI-generated, personalized content Creates shareable documents with proper permissions for immediate client access Mermaid Diagram Visualization: Analyzes job requirements to create relevant workflow diagrams Generates Mermaid.js code for professional flowchart visualization Provides visual representation of proposed solutions Enhances perceived value through custom diagram creation Smart Template Integration: Automatically replaces placeholder text with generated Google Doc links Maintains consistent messaging across all generated assets Ensures cohesive presentation of application, proposal, and supporting materials Required Setup Configuration Personal Information Setup: Update the "aboutMe" variable in both Set Variable nodes with your credentials: Professional background and specializations Notable client achievements with specific revenue numbers Social proof elements (community size, subscriber count, etc.) Relevant project examples with quantified results Google Services Integration: Google Drive API Setup: Enable Google Drive API in Google Cloud Console Create OAuth2 credentials (Client ID and Client Secret) Connect n8n to Google Drive with proper permissions Google Docs Template: Copy the provided Google Docs proposal template to your Drive Update the template ID in the Google Drive node Customize template with your branding and standard language Google Docs API: Ensure Google Docs API is enabled in your Google Cloud project Test document creation and sharing permissions OpenAI API Configuration: Set up OpenAI API credentials across all OpenAI nodes Configure appropriate models (GPT-4O-mini recommended for speed) Set temperature to 0.7 for optimal personalization balance Monitor API usage to control costs Template Customization: Application Template**: Modify the proposal structure in OpenAI prompts to match your services Google Doc Template**: Update the document template with your standard proposal format Personal Details**: Replace all placeholder information with your actual achievements and social proof Business Use Cases Freelance Professionals** - Automate high-quality Upwork applications across multiple job categories Automation Specialists** - Demonstrate capabilities through automated proposal generation itself Service Providers** - Scale application volume while maintaining personalization quality Agency Owners** - Offer proposal automation services to freelance clients Consultants** - Streamline business development with automated custom proposals Content Creators** - Generate professional project proposals with visual workflow representations Revenue Potential This system transforms freelance business development: 10x Application Speed**: Generate comprehensive proposals in minutes vs. hours Higher Response Rates**: Perceived customization and value demonstration increase client engagement Scalable Outreach**: Apply to more jobs with maintained quality through automation Professional Positioning**: Visual diagrams and structured proposals demonstrate expertise Competitive Advantage**: Deliver proposals faster than competitors through intelligent automation Difficulty Level: Advanced Estimated Build Time: 3-4 hours Monthly Operating Cost: ~$30 (OpenAI + Google APIs) Watch My Complete Live Build Want to see me build this entire system from scratch? I walk through every component live - including the AI agent setup, prompt engineering strategies, Google Docs integration, and all the debugging that goes into creating a production-ready freelance automation system. 🎥 See My Live Build Process: "I Built An AI Agent That Automates Upwork ($500K+ Earned)" This comprehensive tutorial shows the real development process - including advanced prompt engineering, modular workflow design, and the exact business strategies that generated $500K+ in Upwork revenue. Set Up Steps AI Agent Foundation: Configure chat trigger and AI agent node with OpenAI integration Set up window buffer memory for conversation context Define system message with clear agent instructions and behavior rules Sub-Workflow Creation: Build three specialized workflows: Application Copy, Google Doc Proposal, Mermaid Code Configure execute workflow triggers for each sub-workflow Set up proper data passing between agent and sub-workflows Google Services Configuration: Create Google Cloud Console project with Drive and Docs APIs enabled Set up OAuth2 credentials and connect to n8n Copy and customize the proposal template document Personalization Setup: Update all "aboutMe" variables with your specific achievements and social proof Customize prompt templates to match your service offerings and communication style Test individual sub-workflows with sample job descriptions Agent Tool Integration: Connect sub-workflows as tools in the main AI agent Configure proper tool descriptions and response property names Test complete agent functionality with realistic job posting scenarios Template Optimization: Refine proposal templates based on your specific service offerings Adjust AI prompts for optimal personalization and response quality Test with various job types to ensure consistent quality output Advanced Optimizations Scale the system with: Job Scraping Integration:** Automatically discover and apply to relevant Upwork jobs Response Tracking:** Monitor application success rates and optimize templates Multi-Platform Support:** Extend to other freelance platforms (Fiverr, Freelancer, etc.) Client Communication:** Automate follow-up sequences for proposal responses Portfolio Integration:** Automatically include relevant portfolio pieces based on job requirements Important Considerations Template Authenticity:** Customize templates significantly to avoid detection as automated Upwork Compliance:** Ensure applications meet platform guidelines and quality standards Personal Branding:** Maintain consistent voice and positioning across all generated content Response Management:** Be prepared to handle increased application volume and client responses Quality Control:** Regularly review and refine generated content for accuracy and relevance Why This System Works The competitive advantage lies in proven strategies: Perceived Customization:** AI generates content that appears manually crafted for each job Value Demonstration:** Visual diagrams and structured proposals show immediate value Speed Advantage:** Deliver comprehensive proposals before competitors finish reading job posts Professional Presentation:** Consistent quality and formatting across all applications Scalable Personalization:** Maintain individual attention at volume through intelligent automation Check Out My Channel For more advanced automation systems and proven freelance business strategies that generate real revenue, explore my YouTube channel where I share the exact methodologies used to build successful automation agencies and scale to $72K+ monthly revenue.
by Yang
Who is this for? This workflow is perfect for eCommerce teams, market researchers, and product analysts who want to track or extract product information from websites that restrict scraping tools. It’s also useful for virtual assistants handling product comparison tasks. What problem is this workflow solving? Many eCommerce and retail sites use dynamic content or anti-bot protections that make traditional scraping methods unreliable. This workflow bypasses those issues by taking a screenshot of the full page, using OCR to extract visible text, and summarizing product information with GPT-4o—all fully automated. What this workflow does This workflow monitors a Google Sheet for new URLs. Once a new link is added, it performs the following steps: Trigger on New URL in Sheet – Watches for new rows added to a Google Sheet. Screenshot URL via Dumpling AI – Sends the URL to Dumpling AI’s screenshot endpoint to capture a full-page image of the product webpage. Save Screenshot to Drive Folder – Uploads the screenshot to a specific Google Drive folder for reference or logging. Extract Text from Screenshot with Dumpling AI – Uses Dumpling AI’s image-to-text endpoint to pull all visible content from the screenshot. Extract Product Info from Screenshot Text with GPT-4o – Sends the extracted raw text to GPT-4o, prompting it to identify structured product information such as product name, price, ratings, deals, and purchase options. Split Each Product Entry – Splits the GPT response (an array of product objects) so each product becomes an individual item for saving. Save Products info to Google Sheet – Appends each product’s structured details to a separate sheet in the same spreadsheet. Setup Google Sheet Create a Google Sheet with at least two sheets: Sheet1 should contain a header row with a column labeled URL. Sheet2 should contain headers: Product Name, price, purchased, ratings, deal, buyingOptions. Connect your Google account in both the trigger and final write-back node. Dumpling AI Sign up at Dumpling AI Create an API key and use it for both HTTP modules: Screenshot URL via Dumpling AI Extract Text from Screenshot with Dumpling AI The screenshot endpoint used is https://app.dumplingai.com/api/v1/screenshot. Google Drive Create a folder for storing screenshots. In the Save Screenshot to Drive Folder node, select the correct folder or provide the folder ID. Make sure permissions allow uploading from n8n. OpenAI Provide an API key for GPT-4o in the Extract Product Info from Screenshot Text with GPT-4o node. The prompt is structured to return structured product listings in JSON format. Split & Save Split Each Product Entry takes the array of product objects from GPT and makes each one a separate execution. Save Products info to Google Sheet writes structured fields into Sheet2 under: Product Name, price, purchased, ratings, deal, buyingOptions. How to customize this workflow Adjust the GPT prompt to return different product fields (e.g., shipping info, product categories). Use a filter node to limit which types of products get written to the final sheet. Add sentiment analysis to analyze review content if available. Replace Google Drive with Dropbox or another file storage app. Notes Make sure you monitor your API usage on both Dumpling AI and OpenAI to avoid rate limits. This setup is great for snapshot-based extraction where scraping is blocked or unreliable.
by assert
Who this template is for This template is for every engineer who wants to automate their code reviews or just get a 2nd opinion on their PR. How it works This workflow will automatically review your changes in a Gitlab PR using the power of AI. It will trigger whenever you comment with +0 to a Gitlab PR, get the code changes, analyze them with GPT, and reply to the PR discussion. Set up Steps Set up webhook of note_events in Gitlab repository (see here on how to do it) Configure ChatGPT credentials Note "+0" in MergeRequest to trigger automatic review by ChatGPT
by Jimleuk
This n8n template demonstrates how to calculate the evaluation metric "Similarity" which in this scenario, measures the consistency of the agent. The scoring approach is adapted from the open-source evaluations project RAGAS and you can see the source here https://github.com/explodinggradients/ragas/blob/main/ragas/src/ragas/metrics/_answer_similarity.py How it works This evaluation works best where questions are close-ended or about facts where the answer can have little to no deviation. For our scoring, we generate embeddings for both the AI's response and ground truth and calculate the cosine similarity between them. A high score indicates LLM consistency with expected results whereas a low score could signal model hallucination. Requirements n8n version 1.94+ Check out this Google Sheet for a sample data https://docs.google.com/spreadsheets/d/1YOnu2JJjlxd787AuYcg-wKbkjyjyZFgASYVV0jsij5Y/edit?usp=sharing
by Yaron Been
🧨 VIP Radar: Instantly Spot & Summarize High-Value Shopify Orders with AI + Slack Alerts Automatically detect when a new Shopify order exceeds $200, fetch the customer’s purchase history, generate an AI-powered summary, and alert your team in Slack—so no VIP goes unnoticed. 🛠️ Workflow Overview | Feature | Description | |------------------------|-----------------------------------------------------------------------------| | Trigger | Shopify “New Order” webhook | | Conditional Check | Filters for orders > $200 | | Data Enrichment | Pulls full order history for the customer from Shopify | | AI Summary | Uses OpenAI to summarize buying behavior | | Notification | Sends detailed alert to Slack with name, order total, and customer insights | | Fallback | Ignores low-value orders and terminates flow | 📘 What This Workflow Does This automation monitors your Shopify store and reacts to any high-value order (over $200). When triggered: It fetches all past orders of that customer, Summarizes the history using OpenAI, Sends a full alert with context to your Slack channel. No more guessing who’s worth a closer look. Your team gets instant insights, and your VIPs get the attention they deserve. 🧩 Node-by-Node Breakdown 🔔 1. Trigger: New Shopify Order Type**: Shopify Trigger Event**: orders/create Purpose**: Starts workflow on new order Pulls**: Order total, customer ID, name, etc. 🔣 2. Set: Convert Order Total to Number Ensures the total_price is treated as a number for comparison. ❓ 3. If: Is Order > $200? Condition**: $json.total_price > 200 Yes** → Continue No** → End workflow 🔗 4. HTTP: Fetch Customer Order History Uses the Shopify Admin API to retrieve all orders from this customer. Requires your Shopify access token. 🧾 5. Set: Convert Orders Array to String Formats the order data so it's prompt-friendly for OpenAI. 🧠 6. LangChain Agent: Summarize Order History Prompt**: "Summarize the customer's order history for Slack. Here is their order data: {{ $json.orders }}" Model**: GPT-4o Mini (customizable) 📨 7. Slack: Send VIP Alert Sends a rich message to a Slack channel. Includes: Customer name Order value Summary of past behavior 🧱 8. No-Op (Optional) Used to safely end workflow if the order is not high-value. 🔧 How to Customize | What | How | |--------------------------|----------------------------------------------------------------------| | Order threshold | Change 200 in the If node | | Slack channel | Update channelId in the Slack node | | AI prompt style | Edit text in LangChain Agent node | | Shopify auth token | Replace shpat_abc123xyz... with your actual private token | 🚀 Setup Instructions Open n8n editor. Go to Workflows → Import → Paste JSON. Paste this workflow JSON. Replace your Shopify token and Slack credentials. Save and activate. Place a test order in Shopify to watch it work. 💡 Real-World Use Cases 🎯 Notify sales team when a potential VIP buys 🛎️ Prep support reps with customer history 📈 Detect repeat buyers and upsell opportunities 🔗 Resources & Support 👨💻 Creator: Yaron Been 📺 YouTube: NoFluff with Yaron Been 🌐 Website: https://nofluff.online 📩 Contact: Yaron@nofluff.online 🏷️ Tags #shopify, #openai, #slack, #vip-customers, #automation, #n8n, #workflow, #ecommerce, #customer-insights, #ai-summaries, #gpt4o