by Oneclick AI Squad
Automate your post-event networking with this intelligent n8n workflow. Triggered instantly after an event, it collects attendee and interaction data, enriches profiles with LinkedIn insights, and uses GPT-4 to analyze engagement and generate tailored follow-up messages. High-value leads are prioritized, messages are sent via email, LinkedIn, or Slack, and all activity is logged in your CRM and database. Save hours of manual follow-up while boosting relationship-building and ROI. ๐คโจ Advanced Features Webhook automation** โ Starts instantly on event completion Multi-Source Enrichment** โ Combines event data, interactions, and LinkedIn profiles AI-Powered Insights** โ GPT-4 analyzes behavior and suggests personalized talking points Smart Priority Filtering** โ Routes leads into High, Medium, and Low priority paths Personalized Content Generation** โ AI crafts custom emails and LinkedIn messages Multi-Channel Outreach** โ Sends via Email, LinkedIn DM, and Slack CRM Integration** โ Automatically updates HubSpot with contact notes and engagement PostgreSQL Logging** โ Stores full interaction history and analytics ROI Dashboard** โ Tracks response rates, meetings booked, and pipeline impact What It Does Collects attendee data from your event platform Enriches with LinkedIn profiles & real-time interaction logs Scores networking potential using engagement algorithms Uses AI to analyze conversations, roles, and mutual interests Generates hyper-personalized follow-up emails and LinkedIn messages Sends messages through preferred channels (email, LinkedIn, Slack) Updates HubSpot CRM with follow-up status and next steps Logs all actions and tracks analytics for performance reporting Workflow Process The Webhook Trigger initiates the workflow via POST request with event and attendee data. Get Attendees** fetches participant list from the event platform. Get Interactions** pulls Q&A, chat, poll, and networking activity logs. Enrich LinkedIn Data** retrieves professional profiles, job titles, and company details via LinkedIn API. Merge & Enrich Data** combines all sources into a unified lead profile. AI Analyze Profile** uses GPT-4 to evaluate interaction depth, role relevance, and conversation context. Filter High Priority** routes top-tier leads (e.g., decision-makers with strong engagement). Filter Medium Priority** handles warm prospects for lighter follow-up. AI Agent1** generates personalized email content using chat model and memory. Generate Email** creates a professional, context-aware follow-up email. Send Email** delivers the message to the leadโs inbox. AI Agent2** crafts a concise, friendly LinkedIn connection message. Generate LinkedIn Msg** produces a tailored outreach note. Send LinkedIn** posts the message via LinkedIn API. Slack Notification** alerts your team in real-time about high-priority outreach. Update CRM (HubSpot)** adds contact, tags, and follow-up tasks automatically. Save to Database (Insert)** logs full lead journey and message content in PostgreSQL. Generate Analytics** compiles engagement metrics and success rates. Send Response** confirms completion back to the event system. Setup Instructions Import the workflow JSON into n8n Configure credentials: Event Platform API (for attendees & interactions) LinkedIn API (OAuth2) OpenAI (GPT-4) SMTP (for email) or Email Service (SendGrid, etc.) HubSpot API Key PostgreSQL Database Slack Webhook URL Trigger with a webhook POST containing event ID and settings Watch personalized outreach happen automatically! Prerequisites Event platform with webhook + attendee/interaction API LinkedIn Developer App with API access OpenAI API key with GPT-4 access HubSpot account with API enabled PostgreSQL database (table for leads & logs) Slack workspace (optional, for team alerts) Example Webhook Payload { "eventId": "evt_spring2025", "eventName": "Annual Growth Summit", "triggerFollowUp": true, "priorityThreshold": { "high": 75, "medium": 50 } } Modification Options Adjust scoring logic in AI Analyze Profile (e.g., weight Q&A participation higher) Add custom email templates in Generate Email with your brand voice Include meeting booking links (Calendly) in high-priority messages Route VIP leads to Send SMS via Twilio Export analytics to Google Sheets or BI tools (Looker, Tableau) Add approval step before sending LinkedIn messages Ready to 10x your event ROI? Get in touch with us for custom n8n automation!
by Sebastien
How to use Get a .csv file with your contacts (you can download this from any contact manager app) Set API key for Google Drive API, and Notion (you need to create a "connection" in Notion) Create Database for your contacts in Notion Choose which properties to extract from the .csv and pass it in to the Notion database. Right now, it transfer 4 pieces of information: full name, email, phone, and company.
by Oneclick AI Squad
This n8n workflow transforms simple chat requests into professional Center of Excellence (COE) blog posts using AI, then automatically publishes them to Google Drive. What Is This? An intelligent blog creation system that takes a topic from chat and produces executive-level blog posts. Uses three AI stages to ensure high-quality, professional content suitable for business consumption. Why Use This? Professional Content**: Creates sophisticated blogs with business insights Save Time**: Topic to published blog in 2-3 minutes No Writing Skills Needed**: AI handles all the writing and formatting Auto-Publishing**: Creates and shares Google Docs automatically Easy to Use**: Just chat your topic, get your blog How It Works 1. Blog Request & Planning Start Blog Request**: Chat interface receives your blog topic Create Blog Outline**: AI generates structured outline with sections AI Brain for Outline**: Powers the initial content planning 2. Content Review & Writing Review & Fix Outline**: AI improves outline structure and flow Write Full Blog Post**: Transforms outline into complete professional blog AI Brain for Review/Writing**: Advanced AI models handle content creation 3. Publish & Share Blog Clean Up Text Format**: Removes formatting issues for clean presentation Save Blog to Google Drive**: Creates properly formatted Google Doc Email Blog to Stakeholder**: Shares with specified team members Make Blog Public**: Creates shareable public link Send Blog Link to User**: Returns document URL via chat How to Use Start Chat: Access the chat interface and type your blog topic Wait: AI automatically processes (2-3 minutes) Get Your Blog: Receive Google Drive link to your published blog Good to Know Processing Time**: 2-3 minutes per blog Content Quality**: Uses advanced Gemini AI models for professional output Auto-Formatting**: Creates clean Google Docs ready for sharing Instant Sharing**: Stakeholders get email notifications automatically Public Access**: Generates shareable links for broader distribution Customizing This Workflow Content Style Modify AI prompts to match your company's writing tone Adjust content evaluation criteria for different audiences Change blog structure templates Publishing & Sharing Update stakeholder email addresses Change Google Drive folder destinations Modify sharing permissions (public/private) Add more distribution channels AI Enhancement Switch between different AI models for speed vs quality Add more review stages for specialized content Include company-specific knowledge sources
by Omar Akoudad
The workflow is well-designed for CRM analysis with a robust quality control mechanism. The dual-AI approach ensures reliable results, while the webhook integration makes it production-ready for real-time CRM data processing. Dual-AI Architecture: Uses DeepSeek Reasoner for analysis and DeepSeek Chat for verification. Flexible Input: Supports both manual testing and production webhook integration. Quality Assurance: Built-in verification system to ensure report accuracy. Comprehensive Analysis: Covers lead conversion, upsell metrics, agent ranking, and more. Professional Output: Generates structured markdown reports with actionable insights
by Sk developer
๐จ AI Image Generator with Flux AI Generate realistic, high-quality images from text prompts using the Flux AI Text-to-Image Generator API via RapidAPI, and seamlessly store the results in Google Drive and log them in Google Sheets โ all automated using n8n. ๐ง What This Workflow Does This no-code automation enables you to: ๐๏ธ Enter a custom text prompt using a web form. ๐ผ๏ธ Generate a photorealistic image using Flux AIโs Text-to-Image Generator via RapidAPI. โ๏ธ Upload the image to Google Drive. ๐ Log the prompt and result in a Google Sheet. โ ๏ธ Capture and log errors in a fallback sheet. ๐ก Use Case Ideal for: Digital artists and marketers Social media managers Brand mockup creators Rapid concept prototyping All without writing a single line of code. โ Benefits No-code automation** for AI-generated images Cloud storage** and structured logging Error handling** built-in Fast content creation** for design, branding, or concept testing Powered by* the Flux AI Text-to-Image Generator API via *RapidAPI** ๐งฉ Node-by-Node Breakdown 1. ๐ On Form Submission Accepts user input for a creative text prompt. ๐ Example: โA silver can with vapor and blue lightning background.โ ๐ก Benefit: No technical knowledge needed. 2. ๐ HTTP Request โ Flux AI API Sends the prompt to the Flux AI Text-to-Image Generator API via RapidAPI. ๐ฆ Returns an image encoded in base64. ๐ก Benefit: Seamless integration with cutting-edge image generation. 3. ๐งช Code Node โ Base64 Decoder Converts the base64 image to a binary .jpg file. ๐ก Benefit: Readies the image for upload/download/sharing. 4. ๐ Google Drive Uploads the generated image to your Google Drive folder. ๐ก Benefit: Secure, sharable cloud storage. 5. ๐ Google Sheets โ Success Log Appends a row with the original prompt, filename, and generation date. ๐ก Benefit: Tracks history of all generated images. 6. โ ๏ธ IF Node โ Error Detection Checks if the image generation failed. ๐ก Benefit: Prevents workflow from halting and routes to error logging. 7. ๐ Google Sheets โ Error Log Logs failed prompts and error messages. ๐ก Benefit: Helps identify what went wrong (e.g. malformed prompt). ๐ ๏ธ Challenges Solved | Problem | How This Workflow Fixes It | |--------|-----------------------------| | Manual prompt-based image generation is slow | Fully automated with Flux AI | | No storage pipeline for generated images | Integrated with Google Drive | | No audit trail for prompts/images | Logged into Google Sheets | | Errors go unnoticed in image generation | Built-in error check and logging | | Users lack API access or dev experience | Friendly web form UI | ๐ API Spotlight This workflow is powered by the Flux AI Text-to-Image Generator API โ available exclusively on RapidAPI. Why use this API? Ultra-fast text-to-image rendering High-resolution results Developer-friendly and cost-effective Great for branding, mockups, and visuals Weโve integrated this API to make advanced image generation accessible with just a prompt โ no AI or dev experience required.
by Jan Oberhauser
Triggers every day at 1pm Gets the current content from Hacker News Gets all the different submission items Extracts the rank, title and url Checks if it is a "Show HN" submission Combines the items into a simple email text Sends an email with the email text
by tanaypant
This is a workflow where a support channel on Telegram is being used to gather customer feedback. Depending on certain keywords in the customer's message, this workflow creates a ticket with a tag in your Freshdesk instance. The customer is then sent a message on Telegram and an item is created on Monday.com for tracking.
by dev
Every 10 minutes look at your published news in your Tiny tiny RSS public feed and make a toot on your mastodon. You'll need: Your mastondon URL instance Your mastondon access token Your Tiny Tiny RSS public published feed URL
by Harshil Agrawal
This workflow demonstrates the use of static data in n8n. The workflow is built on the concept of polling. Cron node: The Cron node triggers the workflow every minute. You can configure the time based on your use-case. HTTP Request node: This node makes an HTTP Request to an API that returns the position of the ISS. Set node: In the Set node we set the information that we need in the workflow. Since we only need the timestamp, latitude, and longitude we set this in the node. If you need other information, you can set them in this node. Function node: The Function node, checks if the incoming data is similar to the data that was returned in the previous execution or not. If the data is different the Function node returns this new node, otherwise, it returns a message 'No New Items'. The data is also stored as static data with the workflow. Based on your use-case, you can build the workflow further. For example, you can use it send updates to Mattermost or Slack
by Abdullahi Ahmed
Title RAG AI Agent for Documents in Google Drive โ Pinecone โ OpenAI Chat (n8n workflow) Short Description This n8n workflow implements a Retrieval-Augmented Generation (RAG) pipeline + AI agent, allowing users to drop documents into a Google Drive folder and then ask questions about them via a chatbot. New files are indexed automatically to a Pinecone vector store using OpenAI embeddings; the AI agent loads relevant chunks at query time and answers using context plus memory. Why this workflow matters / what problem it solves Large language models (LLMs) are powerful, but they lack up-to-date, domain-specific knowledge. RAG augments the LLM with relevant external documents, reducing hallucination and enabling precise answers. (Pinecone) This workflow automates the ingestion, embedding, storage, retrieval, and chat logic โ with minimal manual work. Itโs modular: you can swap data sources, vector DBs, or LLMs (with some adjustments). It leverages the built-in AI Agent node in n8n to tie all the parts together. (n8n) How to get the required credentials | Service | Purpose in Workflow | Setup Link | What you need / steps | | ------------------------- | ------------------------------------------- | -------------------------------------------------------------------------------------------------------------------------------------------------------- | ----------------------------------------------------------------------------------------------------------------------------------- | | Google Drive (OAuth2) | Trigger new file events & download the file | https://docs.n8n.io/integrations/builtin/credentials/google/oauth-generic/ | Create a Google Cloud OAuth app, grant it Drive scopes, get client ID & secret, configure redirect URI, paste into n8n credentials. | | Pinecone | Vector database for embeddings | https://docs.n8n.io/integrations/builtin/credentials/pinecone/ | Sign up at Pinecone, in dashboard create an index, get API key + environment, paste into n8n credential. | | OpenAI | Embeddings + chat model | https://docs.n8n.io/integrations/builtin/credentials/openai/ | Log in to OpenAI, generate a secret API key, paste into n8n credentials. | Youโll configure these under n8n โ Credentials โ New Credential, matching credential names referenced in your workflow nodes. Detailed Walkthrough: How the Workflow Works Hereโs a step-by-step of what happens inside your workflow (matching your JSON): 1. Google Drive Trigger Watches a specified folder in Google Drive. Whenever a new file appears (fileCreated event), the workflow is triggered (polling every minute). You must set the folder ID (in โfolderToWatchโ) to the Drive folder you want to monitor. 2. Download File Takes the file ID from the trigger and downloads the file content (binary). 3. Indexing Path: Embeddings + Storage (This path only runs when new files arrive) The file is sent to the Default Data Loader node (via the Recursive Character Text Splitter) to break it into chunks with overlap (so context is preserved). Each chunk is fed into Embeddings OpenAI to convert text into embedding vectors. Then Pinecone Vector Store (insert mode) ingests the vector + text metadata into your Pinecone index. This ensures your vector store stays up-to-date with files you drop into Drive. 4. Chat / Query Path (Triggered by user chat via webhook) When a chat message arrives via When Chat Message Received, it gets passed into the AI Agent node. Before generation, the AI Agent calls the Pinecone Vector Store1 set in โretrieve-as-toolโ mode, which runs a vector-based retrieval using the user query embedding. The relevant text chunks are pulled as tools/context. The OpenAI Chat Model node is linked as the language model for the agent. Simple Memory** node provides conversational memory (keeping history across messages). The agent combines retrieved context + memory + user input and instructs the model to produce a response. 5. Connections / Flow Logic The Embeddings OpenAI nodeโs output is wired into Pinecone Vector Store (insert) and also into Pinecone Vector Store1 (so the same embeddings can be used for retrieval). The AI Agent has tool access to Pinecone retrieval and memory. The Download File node triggers the insert path. The When chat message triggers the agent path. Similar Workflows / Inspirations & Comparisons To help understand how your workflow fits into whatโs already out there, here are a few analogues: n8n Blog: โBuild a custom knowledge RAG chatbotโ** โ they show a workflow that ingests documents from external sources, indexes them in Pinecone, and responds to queries via n8n + LLM. (n8n Blog) Index Documents from Google Drive to Pinecone** โ this is nearly identical for the ingestion part: trigger on Drive, split, embed, upload. (n8n) Build & Query RAG System with Google Drive, OpenAI, Pinecone** โ shows the full RAG + chat logic, same pattern. (n8n) Chat with GitHub API Documentation (RAG)** โ demonstrates converting API spec into chunks, embedding, retrieving, and chatting. (n8n) Community tutorials & forums** talk about using the AI Agent node with tools like Pinecone, and how the RAG part is often built as a sub-workflow feeding an agent. (n8n Community) What sets your workflow apart is your explicit combination: Google Drive โ automatic ingestion โ chat agent with tool integration + memory. Many templates show either ingestion or chat, but fewer show them combined cleanly with n8nโs AI Agent. Suggested Published Description (you can paste/adjust) > RAG AI Agent for Google Drive Documents (n8n workflow) > > This workflow turns a Google Drive folder into a live, queryable knowledge base. Drop PDF, docx, or text files into the folder โ new documents are automatically indexed into a Pinecone vector store using OpenAI embeddings โ you can ask questions via a webhook chat interface and the AI agent will retrieve relevant text, combine it with memory, and answer in context. > > Credentials needed > > * Google Drive OAuth2 (see: https://docs.n8n.io/integrations/builtin/credentials/google/oauth-generic/) > * Pinecone (see: https://docs.n8n.io/integrations/builtin/credentials/pinecone/) > * OpenAI (see: https://docs.n8n.io/integrations/builtin/credentials/openai/) > > How it works > > 1. Drive trigger picks up new files > 2. Download, split, embed, insert into Pinecone > 3. Chat webhook triggers AI Agent > 4. Agent retrieves relevant chunks + memory > 5. Agent uses OpenAI model to craft answer > > This is built on the core RAG pattern (ingest โ retrieve โ generate) and enhanced by n8nโs AI Agent node for clean tool integration. > > Inspiration & context > This approach follows best practices from existing n8n RAG tutorials and templates, such as the โIndex Documents from Google Drive to Pineconeโ ingestion workflow and โBuild & Query RAG Systemโ templates. (n8n) > > You're free to swap out the data source (e.g. Dropbox, S3) or vector DB (e.g. Qdrant) as long as you adjust the relevant nodes. If you like, I can generate a polished Markdown README for you (with badges, diagrams, instructions) ready for GitHub/n8n community publishing. Do you want me to build that? [1]: https://www.pinecone.io/learn/retrieval-augmented-generation/?utm_source=chatgpt.com "Retrieval-Augmented Generation (RAG) - Pinecone" [2]: https://n8n.io/integrations/agent/?utm_source=chatgpt.com "AI Agent integrations | Workflow automation with n8n" [3]: https://blog.n8n.io/rag-chatbot/?utm_source=chatgpt.com "Build a Custom Knowledge RAG Chatbot using n8n" [4]: https://n8n.io/workflows/4552-index-documents-from-google-drive-to-pinecone-with-openai-embeddings-for-rag/?utm_source=chatgpt.com "Index Documents from Google Drive to Pinecone with OpenAI ... - N8N" [5]: https://n8n.io/workflows/4501-build-and-query-rag-system-with-google-drive-openai-gpt-4o-mini-and-pinecone/?utm_source=chatgpt.com "Build & Query RAG System with Google Drive, OpenAI GPT-4o-mini ..." [6]: https://n8n.io/workflows/2705-chat-with-github-api-documentation-rag-powered-chatbot-with-pinecone-and-openai/?utm_source=chatgpt.com "Chat with GitHub API Documentation: RAG-Powered Chatbot ... - N8N"
by jason
This workflow takes a text file as input. It pulls the information from the text file and used it as a parameter to execute a command for each text line. This workflow references a file /home/n8n/filelist.txt in the Read Binary File node which will need to be changed to work properly. You can also edit the Execute Command node to modify what happens for each of these lines of text. Note: This workflow requires the Execute Command node which is only available on the on-premise version of n8n.
by Max Tkacz
This workflow is a generic example of how to load data from your workflow into a destination that stores tabular data. For example, a Google Sheets or Airtable sheet, a .CSV file, or any relational database like MySQL. Generally, you need to ensure that you send well-formatted data into the Spreadsheet or Database node. You can use the Set or Function node to transform data into the correct format for your destination. Key concepts Spreadsheets and databases have columns, like "Name" and "Email". The data you send into a Spreadsheet/ Database node needs to match these column names for each row of data that you want to insert. Data points need to be represented as key-value pairs. Specifically, each item of data needs to have a JSON key for each column in the sheet. For a spreadsheet with "Name" and "Email" columns, it would look like: {"Name" : "Karla", "Email" : "karla@email.com"} Before appending or inserting data to a spreadsheet or database, you might need to transform it into the correct format. You can preprocess the data with a Set or Function node. The Set node allows you to perform simple transforms when the data you want to load into spreadsheet rows is already represented as items. Use the Function node when you need to map nested data (like arrays) inside a single item to their own top-level items (Example in community forums). Spreadsheet and database nodes in n8n perform their configured action (like Append, Create Row, Write to File) on each item of input data. Workflow walkthrough