by L Hùng
Pre-Conditions A Facebook Developer account with an active app. Basic understanding of n8n workflows. Access to a database (optional, for storing tokens). Setup Webhook Activation: Configure the Webhook to receive user requests and process input data. Ensure the Webhook URL is correctly set in your Facebook App settings. Short-Lived Token Retrieval: Use Facebook OAuth to fetch a short-lived token from the authorization code. Long-Lived Token Conversion: Convert the short-lived token into a long-lived token (valid for ~60 days). Page Token Retrieval: Follow the provided instructions to retrieve Page Tokens for posting on managed Facebook Pages. Customizable Scopes: Edit the correctScopes array to include or exclude permissions as needed. Optional Database Storage: Extend the workflow to save tokens to a database instead of displaying them on-screen. Step-by-Step Instructions: Detailed guidance is provided via sticky notes for activating the app, configuring Webhook, and editing parameters like fb_redirect_uri, app_id, and app_secret. Who the Template is For Developers**: Integrating Facebook APIs into their applications. Social Media Managers**: Automating posting and engagement on Facebook Pages. n8n Users**: Looking for a ready-to-use workflow for Facebook Token management. Primary Use Automates Facebook Token retrieval and management. Supports posting to Facebook Pages via Page Tokens. Easily customizable and extendable for specific requirements.
by simonscrapes
Use Case Generate accurate search volume data for SEO keyword research: You have a list of potential keywords to target for your website SEO but don't know their actual search volume You need historical data to identify seasonal trends in keyword popularity You want to assess keyword difficulty to prioritize your content strategy You need data-driven insights for planning your SEO campaigns What this Workflow Does The workflow connects to Google's Keyword Planner API to retrieve keyword metrics for your SEO research: Fetches monthly search volume for each keyword Provides historical trends data for the past 12 months Calculates keyword difficulty scores Delivers competition metrics from Google Ads Setup Fill the Set 20 Keywords with up to 20 Keywords of your choosing in an array e.g. ["keyword 1", "keyword 2",...] Create a Google Ads API account and add credentials to Get Search Data node Replace the Connect to your own database with your own database for the output How to Adjust it to Your Needs Change the Set 20 Keywords node input to a source of your choosing e.g. Airtable database with 20 keywords Connect to output source of your choosing More templates and n8n workflows >>> @simonscrapes
by Airtop
Use Case Automatically responding to X (formerly Twitter) posts can help you engage with potential customers at scale, saving time while maintaining a personal touch. What This Automation Does This automation replies to specified X posts using the following input parameters: airtop_profile: The name of your Airtop Profile connected to X. thread_url: The URL of the X post to reply to. Example reply_text: The message you want to post as a reply. How It Works Creates a browser session using Airtop. Navigates to the specified X post. Types and submits the reply text. Setup Requirements Airtop API Key — free to generate. An Airtop Profile connected to X (requires one-time login). Next Steps Combine with X Monitoring**: Use this with the X monitoring automation to create a fully automated engagement pipeline. Extend to Other Platforms**: Adapt the automation for use on LinkedIn, Reddit, or any web community. Read more about this Airtop Automation.
by Joseph LePage
🔍 This n8n workflow integrates Tavily's search and extract APIs with AI summarization capabilities to process web content efficiently. Quick Setup Get your Tavily API key from https://app.tavily.com/home Replace tvly-YOUR_API_KEY in the "Tavily API Key" node Connect your OpenAI credentials to the "OpenAI Chat Model" node Deploy the workflow and start the chat trigger Core Features Search & Extract 🎯 Intelligent web searching with relevance filtering Automated content extraction from top results AI-powered content summarization in markdown format User Interaction 💬 Chat-based search topic input Real-time processing pipeline Structured markdown output The workflow demonstrates practical implementation of Tavily's API endpoints while handling the complete process from search to summarization in a single automated pipeline.
by Niklas Hatje
Use Case When trying to maximize your outreach, website visitors are often an overlooked source of qualified new leads. This workflow allows your to track and enrich new website visitors and saves them to a Google Sheet once they meet a pre-defined criteria. What this workflow does This workflow fires once a day and gets all your leads saved in Leadfeeder. It then takes the leads that meet a pre-defined engagement criteria, e.g. that they visited your site 3 times, and enriches them additionally with Clearbit. From there it filters the leads again by a criteria on the company, e.g. a minimum employee count, and saves matching leads into a Google Sheet document. Setup Add your Leedfeeder credentials. The name should be Authorization and the value Token token=yourapitoken. You can find your token via Settings -> Personal -> API-Token Add your Google Sheet credentials Save the Leedfeeder account names you want to use in the Setup node Copy the Google Sheets Template and add its URL to the Setup node How to adjust this to your needs Adjust and/or remove the engagement and company criteria Add more ways to enrich a company Potential ideas to enhance the use of this workflow Automatically reach out to users that meet the criteria / that get added to the sheet Create a workflow that finds the right employee in companies that are identified by this workflow
by Eric
Why use this You need to delete (many) posts on a WordPress website and also delete the featured image associated with each post. Hours of rote work cut into a fraction with this automation. How it works set your wordpress URL in the manual trigger node set your WP post search parameters (WP API returns 10 posts by default; you could also set up pagination for scaling this automation beyond 10 posts per execution) decide (and build) your filter/approval process What you can expect this automation is set up to run the 10 oldest pending posts, with oldest first if you remove the 'Filter' node from the workflow, after each run, another 10 posts will be returned from WP Notes on Filter/Approval This is arbitrary and depends on your own use case. Maybe you have an editor who needs to approve the post deletion. You might want to get approval by email, slack msg or ticketing system. Or maybe you just want to monitor the process and spare specific posts from deletion. I used the Filter node to only grab the first item (itemIndex < 1) which in this case was the oldest pending post. This could also be expanded to two separate workflows: One triggered when a pending post is created that sends an approval request A second triggered by the approval/rejection that either publishes or deletes the post, depending on the approval result This would require another HTTP request, similar to the DELETE post request, that instead publishes the post.
by Ranjan Dailata
Who this is for? Google SERP Tracker + Trends and Recommendations is an AI-powered n8n workflow that extracts Google search results via Bright Data, parses them into structured JSON using Google Gemini, and generates actionable recommendations and search trends. It outputs CSV reports and sends real-time Webhook notifications. This workflow is ideal for: SEO Agencies needing automated rank & trend tracking Growth Marketers seeking daily/weekly search-based insights Product Teams monitoring brand or competitor visibility Market Researchers performing search behavior analysis No-code Builders automating search intelligence workflows What problem is this workflow solving? Traditional tracking of search engine rankings and search trends is often fragmented and manual. Analyzing SERP changes and trends requires: Manual extraction or using unstable scrapers Unstructured or cluttered HTML data Lack of actionable insights or recommendations This workflow solves the problem by: Automating real-time Google SERP data extraction using Bright Data Structuring unstructured search data using Google Gemini LLM Generating actionable recommendations and trends Exporting both CSV reports automatically to disk for downstream use Notifying external systems via Webhook What this workflow does Accepts search input, zone name, and webhook notification URL Uses Bright Data to extract Google Search Results Uses Google Gemini LLM to parse the SERP data into structured JSON Loops over structured results to: Extract recommendations Extract trends Saves both as .csv files (example below): Google_SERP_Recommendations_Response_2025-06-10T23-01-50-650Z.csv Google_SERP_Trends_Response_2025-06-10T23-01-38-915Z.csv Sends a Webhook with the summary or file reference LLM Usage Google Gemini LLM handles: Parsing Google Search HTML into structured JSON Summarizing recommendation data Deriving trends from the extracted SERP metadata Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. A Google Gemini API key (or access through Vertex AI or proxy). Update the Set input fields with the search criteria, Bright Data Zone name, Webhook notification URL. How to customize this workflow to your needs Input Customization Set your target keyword/phrase in the search field Add your webhook_notification_url for external triggers or notifications SERP Source You can extend the Bright Data search logic to include other engines like Bing or DuckDuckGo. Output Format Edit the .csv structure in the Convert to File nodes if you want to include/exclude specific columns. LLM Prompt Tuning The Gemini LLM prompt inside the Recommendation or Trends extractor nodes can be fine-tuned for domain-specific insight (e.g., SEO vs eCommerce focus).
by Yaron Been
🚀 AI-Powered YouTube Video Summary Distributor: From Channel to Community! Workflow Overview This sophisticated n8n automation transforms YouTube content discovery into a seamless, multi-platform intelligence sharing process. By intelligently connecting YouTube RSS, AI summarization, and content distribution platforms, the workflow: Discovers New Content: Monitors YouTube channels via RSS feed Captures latest video uploads Tracks content in real-time AI-Powered Summarization: Extracts video metadata Generates concise, meaningful summaries Leverages GPT-4o for intelligent content analysis Intelligent Distribution: Logs summaries in Google Sheets Sends summaries to Slack for review Publishes approved content to Reddit Detailed Setup Instructions 1. YouTube Data API Configuration Prerequisites Google Cloud Console account Enabled YouTube Data API v3 Setup Steps: Go to Google Cloud Console Create a new project Enable YouTube Data API v3 Create credentials (API Key) Store API key securely in n8n credentials Obtain channel RSS feed URL 2. OpenAI API Setup Prerequisites OpenAI account Active API subscription Configuration: Visit OpenAI Platform Generate API key Select GPT-4o model Configure API key in n8n credentials Set up billing and usage limits 3. Slack Integration Prerequisites Slack workspace Slack app permissions Setup Process: Create a Slack app in your workspace Configure OAuth scopes for sending messages Install app to workspace Obtain webhook or OAuth token Configure in n8n Slack node 4. Reddit API Configuration Prerequisites Reddit account Reddit application created Steps: Go to Reddit's app preferences Create a new application Obtain client ID and secret Configure OAuth2 credentials in n8n Select target subreddit Workflow Customization Channel Modification Replace YouTube RSS feed URL in trigger node Adjust channel_id parameter Modify extraction logic if needed Subreddit Customization Change subreddit parameter in Reddit node Adjust title and text formatting AI Summarization Tuning Modify system message in Summarizer Agent Adjust prompt for different content types Implement custom filtering Key Customization Points Modify RSS feed URL Change target subreddit Adjust AI summarization prompt Add custom filtering logic Implement multi-channel support Technical Requirements n8n v0.220.0 or higher YouTube Data API v3 OpenAI API access Slack workspace Reddit application Stable internet connection Potential Use Cases Content creator content tracking Research and trend analysis Social media content distribution Automated content curation Community engagement Security Considerations Use environment variables for API keys Implement proper OAuth2 authentication Respect platform usage guidelines Maintain user privacy Future Enhancement Roadmap Multi-language support Advanced content filtering Sentiment analysis integration Expanded platform distribution Customizable summarization parameters Workflow Visualization [YouTube RSS Trigger] ⬇️ [Extract Channel ID] ⬇️ [Fetch Video Details] ⬇️ [AI Summarization] ⬇️ [Google Sheets Logging] ⬇️ [Slack Approval] ⬇️ [Reddit Publishing] Hashtag Performance Boost 🚀 #YouTubeAutomation #AIContentDistribution #WorkflowInnovation #ContentCuration #AIMarketing #DigitalMediaTech #AutomatedSummaries #CrossPlatformContent Connect With Me Ready to revolutionize your content workflow? 📧 Email: Yaron@nofluff.online 🎥 YouTube: @YaronBeen 💼 LinkedIn: Yaron Been Transform your content strategy with intelligent, automated workflows! Note: Always test and customize the workflow to fit your specific use case and comply with platform guidelines.
by Yaron Been
Automated outreach system that identifies and contacts potential leads from CrunchBase with personalized, timely messages. 🚀 What It Does Identifies target companies and contacts Personalizes email content Schedules follow-ups Tracks responses Integrates with email providers 🎯 Perfect For Sales development reps Business development teams Startup founders Investment professionals Partnership managers ⚙️ Key Benefits ✅ Automated lead generation ✅ Personalized outreach at scale ✅ Follow-up automation ✅ Response tracking ✅ Time-saving workflow 🔧 What You Need CrunchBase API access Email service (e.g., Gmail, SendGrid) n8n instance CRM (optional) 📊 Features Contact information extraction Email template personalization Send time optimization Open/click tracking Response handling 🛠️ Setup & Support Quick Setup Start sending in 30 minutes with our step-by-step guide 📺 Watch Tutorial 💼 Get Expert Support 📧 Direct Help Transform your outbound sales process with automated, personalized outreach to high-quality leads from CrunchBase.
by Sherlockes
What this template is made for: I have a personal Telegram channel and a bot inside it where I save interesting links that I want to save or read later. The idea is that n8n will take care of reading the new links added to this channel and send them, through the corresponding API, to the Hoarder and Readeck installations. How it works Since my server where n8n runs is not always on, a "Schedule Trigger" will be responsible for checking every so often if there is any new content in the Telegram channel where I store the links. This request is made through "http request" and the Telegram API. Next, a code block is responsible for filtering out everything that is not a hyperlink. At this point, the flow splits into two so that parallel and similar processes are performed for Hoarder and Readeck. The corresponding API is accessed to get a list of all the links saved in the corresponding service. A code block is responsible for filtering the list of hyperlinks previously obtained from Telegram so that only those that are not already saved in the service continue. Finally, another "Http Request" node is responsible for using the service API to save the link in the corresponding service. Configuration instructions The template makes use of the environment variables that I have declared in the n8n "docker-compose.yml" file through an external ".env" file. These are the variables I use: Telegram Bot Token Sherlink TG_SHERLINK_BOT_TOKEN=XXXXXXXX:XXXXXXXXXXXXXXXX Id Telegram Channel Sherlink TG_SHERLINK_ID=-XXXXXXXXXXXXX Readeck server READECK_SERVER=http://readeck.midomain.com READECK_API_KEY=xxxxxxxxxxxxx Hoarder server HOARDER_SERVER=http://hoarder.midomain.com HOARDER_API_KEY=xxxxxxxxxxxxxx Created in 1.85.4 n8n version
by PollupAI
Who is this for? This workflow is designed for Customer Satisfaction Managers (CSM), sales professionals, and operations managers who need to automate the analysis of client transcripts, save summarized notes to HubSpot, and route relevant feedback to the appropriate departments via email. What problem is this workflow solving? / Use Case Manually processing client conversations, extracting key insights, and distributing them to the right teams is time-consuming and error-prone. This workflow automates: Transcript analysis** using AI (OpenAI) to identify relevant content. HubSpot integration** to log meeting notes against client records. Email routing** to ensure feedback reaches the correct departments (e.g., support, sales, product, admin). What this workflow does Input Transcript: Accepts a client conversation transcript (e.g., from emails, calls, or chats). HubSpot Sync: Searches for the client’s HubSpot ID using their email. Uploads a summarized version of the conversation as meeting notes. AI-Powered Routing: Uses an OpenAI model to analyze the transcript and categorize content by department. Triggers emails (via Gmail) to route feedback to the relevant teams. Form Completion: Ends the workflow with optional user confirmation. Setup Prerequisites: n8n instance (cloud or self-hosted). HubSpot API credentials (for contact lookup and notes upload). OpenAI API key (for transcript analysis). Gmail account (for sending emails). Configuration: Replace placeholder nodes (e.g., HubSpot, OpenAI, Gmail) with your authenticated accounts. Define email templates and recipient addresses for routing. Adjust the OpenAI prompt to match your categorization criteria (e.g., "support," "billing"). How to customize this workflow to your needs Transcript Sources**: Extend the workflow to pull transcripts from other sources (e.g., Zoom, Slack). Departments**: Modify the routing logic to include additional teams or conditions. Notifications**: Add Slack/MS Teams alerts for urgent feedback. Error Handling**: Introduce retries or fallback actions for failed HubSpot/Gmail steps.
by Corentin Ribeyre
This template can be used to search for an email address with Icypeas. Be sure to have an active account to use this template. How it works This workflow can be divided into three steps : The workflow initiates with a manual trigger (On clicking 'execute'). It connects to your Icypeas account. It performs an HTTP request to search for an email address. Set up steps You will need a working icypeas account to run the workflow and get your API Key, API Secret and User ID. You will need a personn firstname, lastname and domain/company name to perform the search.