by AI Native
This workflow automates the process of retrieving Hugging Face paper summaries, analyzing them with OpenAI, and storing the results in Notion. Hereโs a breakdown of how it works: โฐ Scheduled Trigger: The flow is set to run automatically at 8 AM on weekdays. ๐ Fetching Paper Data: It fetches Hugging Face paper summaries using their API. ๐ Data Check: Before processing, the workflow checks if the paper already exists in Notion to avoid duplicates. ๐ค Content Analysis with OpenAI: If the paper is new, it extracts the summary and uses OpenAI to analyze the content. ๐ฅ Store Results in Notion: After analysis, the summarized data is saved in Notion for easy reference. โ๏ธ Set Up Steps for Automation Follow these steps to set up this automated workflow with Hugging Face, OpenAI, and Notion integration: ๐ Obtain API Tokens: Youโll need the Notion and OpenAI API tokens to authenticate and connect these services with n8n. ๐ Integration in n8n: Link Hugging Face, OpenAI, and Notion by configuring the appropriate nodes in n8n. ๐ง Configure Workflow Logic: Set up a cron trigger for automatic execution at 8 AM on weekdays. Use an HTTP request node to fetch Hugging Face paper data. Add logic to check if the data already exists in Notion. Set up the OpenAI integration to analyze the paperโs summary. Store the results in Notion for easy access and reference. Result:
by bangank36
This workflow retrieves all Shopify Orders and saves them into a Google Sheets spreadsheet using the Shopify Admin REST API. It uses pagination to ensure all orders are collected efficiently. I originally built this workflow for my own use and found it valuable for understanding how Shopify pagination works. Now, Iโm sharing it to help others automate their order retrieval process. How It Works Instead of relying on the built-in Shopify node (Get Orders Many), this workflow leverages the HTTP Request node to fetch paginated chunks manually. Shopify uses cursor-based pagination (page_info) instead of traditional page numbers. Pagination data is stored in the response headers, so we need to enable Include Response Headers and Status in the HTTP Request node. You can modify the limit parameter to control batch sizes and optimize for rate limits. This workflow can be run on demand or scheduled to keep your data up to date. Parameters You can adjust these parameters in the HTTP Request node: limit โ The number of orders per request (default: 50, max: 250). fields โ Comma-separated list of fields to retrieve. page_info โ Used for pagination; only limit and fields are allowed when paginating. ๐ Note: when you query the paginated chunks with page_info, only the limit and fields parameters are allowed Credentials Shopify API Key โ Required for authentication. Google Sheets API credentials โ Needed to insert data into the spreadsheet. ๐พ Clone the Google Sheets template here Who Is This For? Shopify store owners who need to export all orders to Google Sheets. Users who want full control over API parameters for optimized queries. Anyone looking for a flexible and scalable Shopify data extraction solution. Explore More Templates ๐ Check out my other n8n templates
by Yaron Been
This workflow automatically analyzes competitor content performance across various platforms to understand what content resonates with their audience. It saves you time by eliminating the need to manually track competitor content and provides insights into successful content strategies and engagement patterns. Overview This workflow automatically scrapes competitor websites, blogs, and social media to analyze content performance metrics including engagement rates, shares, comments, and audience response. It uses Bright Data to access competitor content without restrictions and AI to intelligently analyze performance data and extract actionable insights. Tools Used n8n**: The automation platform that orchestrates the workflow Bright Data**: For scraping competitor content platforms without being blocked OpenAI**: AI agent for intelligent content performance analysis Google Sheets**: For storing competitor content analysis and performance metrics How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Bright Data: Add your Bright Data credentials to the MCP Client node Set Up OpenAI: Configure your OpenAI API credentials Configure Google Sheets: Connect your Google Sheets account and set up your content analysis spreadsheet Customize: Define competitor URLs and content performance tracking parameters Use Cases Content Strategy**: Learn from high-performing competitor content to improve your own strategy Competitive Analysis**: Track competitor content trends and audience engagement patterns Content Optimization**: Identify content types and topics that drive the most engagement Market Research**: Understand what content resonates with your target audience Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Bright Data**: https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #competitoranalysis #contentperformance #brightdata #webscraping #contentmarketing #n8nworkflow #workflow #nocode #contentanalysis #competitormonitoring #contentresearch #engagementanalysis #marketresearch #contentintelligence #competitiveintelligence #contentoptimization #performancetracking #contentmetrics #marketanalysis #contentaudit #brandanalysis #contentstrategy #digitalmarketing #contentinsights #socialmediaanalysis #contentmonitoring #performanceanalysis #competitorresearch
by Yaron Been
This workflow provides automated access to the Vcollos Trefilio AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Vcollos Trefilio model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Vcollos/trefilio AI model Vcollos Trefilio**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Louis
Workflow Overview This workflow automates the process of updating a Spotify playlist with tracks from a YouTube playlist, ensuring no duplicates are added. Key Components Manual Trigger: Starts the workflow when you click โTest workflowโ. Spotify Integration: Retrieves tracks from a specified Spotify playlist. YouTube Integration: Fetches tracks from a designated YouTube playlist. Batch Processing: Processes tracks in batches to handle multiple items efficiently. Track Search: Searches for YouTube tracks on Spotify to find corresponding IDs. Comparison: Compares existing Spotify tracks with YouTube tracks to identify which ones to add. Track Addition: Adds new Spotify tracks to the playlist that are not already included. If you have any questions or need clarification, feel free to ask!
by Ficky
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. ๐ n8n Workflow: Meetup Registration & Giveaway Picker This n8n workflow is designed to handle both participant registration and giveaway winner selection, fully integrated with a frontend experience. ๐ Participant Registration Flow This part of the workflow automates the registration process for meetup attendees. ๐ Steps: ๐ Participant Form (Webhook Trigger) Triggered when a user submits the form. Captures fields like nama_lengkap, email, whatsapp, discord_username, and more. ๐ ๏ธ Data Mapping Maps raw form inputs into a structured format. Supports trimming, validation, and transformation as needed. ๐พ Save to Database Performs an upsert operation to store participant info in PostgreSQL. Prevents duplicate entries based on unique ID (e.g., WhatsApp or email). โ Confirmation Screen Returns a success message or thank-you page after registration is complete. ๐ Giveaway App This section serve frontend app to display and select random winners. ๐ Steps: ๐ Webhook GET (Giveaway App) Exposes a public endpoint that delivers a ready-to-use HTML app. Can be opened in a browser or projected during the event. ๐ฅ Fetch Participants Executes a SQL query to retrieve all participant records from the database. ๐งฎ Format Participant Data Redacts personal details (e.g., masks WhatsApp numbers). Encodes the id using Base64 for privacy and uniqueness. ๐ค Render Giveaway App Returns an HTML-based Single Page Application with the participant list included. Allows the host to click a button to pick random winners live. โ Use Case Highlights Streamlined participant collection and storage using n8n workflows Secure data handling with redaction and ID encoding Frontend integration for a fun, interactive giveaway experience Reusable for other community events, workshops, or internal team draws
by Aleksey Panov
Perplexity API Module This reusable workflow allows you to interact with the Perplexity API using the sonar or sonar-pro models. It is designed to be triggered from other workflows and accepts dynamic prompts via input parameters. Features ๐งฑ Modular: Triggered using Execute Workflow from any other workflow ๐ฅ Inputs: SystemPrompt: Set your system-level instruction for the LLM UserPrompt: The main user prompt for the conversation ๐ง Uses Perplexityโs chat/completions endpoint ๐ Built with API authentication in mind (Bearer token or header auth) ๐ Easily extendable for any model in the Perplexity suite Models supported sonar sonar-pro See full model list and capabilities Usage Set up your credentials in n8n under HTTP Bearer or Header Auth. Add this workflow as a subworkflow using the โExecute Workflowโ node. Pass the desired SystemPrompt and UserPrompt as input variables. Receive the model response in return. Notes This template is inactive by default and safe to import. Includes sticky notes with API references and model info.
by Cheng Siong Chin
How It Works This workflow automates property registration verification, fraud detection, and blockchain-based compliance tracking by systematically assessing fraud risk, validating transactions, ensuring data immutability through cryptographic hashing, and recording property records on the blockchain. It ingests property registration data, applies GPT-4โdriven fraud analysis with risk scoring, and verifies transaction legitimacy against regulatory and contractual criteria. The system generates cryptographic hashes for property and lease records, validates compliance requirements using AI-based analysis, queries the blockchain for verification, logs transactions on-chain, stores audit records in structured sheets, and securely archives all supporting documentation. Designed for real estate firms, legal practices, and property management companies, it enables transparent verification, fraud mitigation, and tamper-resistant compliance tracking across the property lifecycle. Setup Steps Configure property data source and set up OpenAI GPT-4 for fraud detection and compliance. Connect blockchain network credentials and configure hash generation parameters. Set up Google Sheets for audit logging and configure blockchain verification queries. Define fraud risk thresholds, compliance criteria, and transaction validation rules. Prerequisites Property registration data source; OpenAI API key; blockchain network access Use Cases Real estate firms automating fraud checks on property transactions; Customization Adjust fraud detection criteria and risk thresholds, modify blockchain network selection. Benefits Eliminates manual fraud detection, prevents title fraud and forgery
by Yaron Been
This workflow provides automated access to the Moicarmonas 2Ndmoises_Generator AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Moicarmonas 2Ndmoises_Generator model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Moicarmonas/2ndmoises_generator AI model Moicarmonas 2Ndmoises_Generator**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Taiki
disclaimer: This workflow contains community nodes that are only compatible with the self-hosted version of n8n. Quick Summary This template automates the creation of YouTube Shorts scripts. Simply send a YouTube video URL to a Telegram bot, and this workflow will use AI (Google Gemini) to generate a script optimized for Shorts and send it back to you. It streamlines your content creation process by eliminating the manual work of transcribing and scriptwriting. Target Audience & Problem Solved This template is ideal for: YouTube Content Creators Social Media Marketers Anyone looking to efficiently repurpose - - video content It solves the following problems: Quickly creating engaging short-form video content from existing long-form videos. Drastically reducing the time spent on video transcription, summarization, and scriptwriting. Enabling content idea generation on the go, even from a smartphone without access to a PC. Workflow Overview This workflow automatically generates a script for YouTube Shorts using AI (Google Gemini) from a YouTube video URL received via Telegram, and then sends the result back to Telegram. The process is as follows: Receive URL: A Telegram bot receives a YouTube video URL from a user. Get Transcript: It uses the community node n8n-nodes-supadata to retrieve the video's transcript via the Supadata service. Generate Script: Based on the retrieved transcript, Google Gemini creates a script (title and body) optimized for a YouTube Shorts video. Send Result: The generated script is sent back to the user who sent the original URL via Telegram. How to Use Prerequisites: Install the n8n-nodes-supadata community node as described below. Register your API keys for Telegram, Supadata, and Google Gemini as credentials in n8n. Set the appropriate credentials in each node within the workflow. For Telegram, create your bot with BotFarther. Execution: Activate the workflow. Send a YouTube video URL to your configured Telegram bot. Confirmation: After a short while, the workflow will run automatically, and you will receive the completed script from the Telegram bot. Customization Guide You can tailor this template to your specific needs: Modify the Prompt: Edit the prompt in the Google Gemini node to change the style of the generated script (e.g., make it more casual, use a specific tone, include emojis). Change the AI Model: You can switch to other AI models like OpenAI. Simply add or replace the corresponding n8n node. Change the Notification Channel: Replace the Telegram node with a Discord or Slack node to send the results to a different platform. Community Node Installation This workflow requires the n8n-nodes-supadata community node. Please install it using the following steps. Installation from the n8n UI (Recommended) Navigate to Settings > Community Nodes in your n8n interface. Click on Install. Enter n8n-nodes-supadata in the input field and click the Install button.
by Davide
This workflow integrates a chatbot frontend with a backend powered by Langflow, a visual low-code AI development tool. The flow is triggered whenever a chat message is received via the n8n chatbot widget embedded on a website. It then sends the message to a Langflow flow for processing and returns the generated response to the user. How It Works Chat Trigger: The workflow starts with a webhook trigger (When chat message received) that listens for incoming chat messages from the n8n Chat interface. Langflow Integration: The chat input is sent to a Langflow instance via an HTTP request (Langflow node). The request includes the user's message and expects a response from the Langflow flow. Response Processing: The output from Langflow is extracted and formatted using the Edit Fields node, ensuring the chatbot displays the response correctly. Customization: Sticky notes provide instructions for embedding the n8n Chat widget on a website and customizing its appearance, including welcome messages, language settings, and branding. Set Up Steps Configure Langflow Connection: Replace LANGFLOW_URL and FLOW_ID in the HTTP request node with your Langflow instance details. Ensure the API headers (e.g., Content-Type: application/json) and authentication (if required) are correctly set. Deploy n8n Chat: Add the provided CDN script to your website, replacing YOUR_PRODUCTION_WEBHOOK_URL with the webhook URL generated by the When chat message received node. Customize the chatbotโs UI (e.g., title, placeholder text, initial messages) using the JavaScript snippet in the sticky notes. Activate Workflow: Toggle the workflow to "Active" in n8n. Test the chatbot by sending a message and verifying the Langflow response is processed and displayed correctly. Advantages โ Seamless Langflow Integration It allows n8n to communicate directly with a Langflow flow via API, enabling AI responses using custom-designed Langflow logic. โ No-Code Chatbot Deployment With just a script snippet, the chatbot widget can be embedded into any website. Minimal coding is required to launch a fully functioning AI chatbot. โ Customizable UI/UX The included embed code offers full control over the chatbot's appearance, language, welcome message, input placeholder, and brandingโideal for white-label or customer-facing deployments. โ Modular and Extensible Because it's built in n8n, this chatbot can be easily extended with other services like CRMs, email alerts, or databases, without leaving the platform. โ Real-Time AI Interactions Thanks to Langflow's API and chat response support, users get immediate and dynamic AI-driven replies. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Oneclick AI Squad
This automated n8n workflow continuously tracks real-time flight fare changes by querying airline APIs (e.g., Amadeus, Skyscanner). It compares new prices with historical fares and sends instant notifications to users when a fare drop is detected. All tracked data is structured and logged for audit and analysis. Key Insights Works post-booking to track price fluctuations for booked routes. Supports multiple fare sources for improved accuracy and comparison. Users are notified instantly via email, SMS, or Slack for high-value drops. Historical pricing data is stored for trend analysis and refund eligibility checks. Can be extended to monitor specific routes or apply airline-specific refund rules. Workflow Process Schedule Trigger Initiates the fare check every 6 hours. Fetch Flight Fare Data Queries APIs (Amadeus, Skyscanner) for current flight fares. Get Tracked Bookings Retrieves tracked routes from the internal database. Compare Fares Detects price drops compared to original booking fares. Update Fare History Table Logs the new fare and timestamp into the fare_tracking table. Classify Drops Determines priority based on absolute and percentage savings. Notify Users Email Alerts: For all medium/high priority drops. SMS Alerts: For savings > \$100 or >15%. Slack Notifications: For internal alerts and rebooking suggestions. Log Activity Stores all sync actions and notifications in fare_alert_logs. Usage Guide Import** the workflow into your n8n instance. Set up** API credentials for Amadeus and Skyscanner. Configure email, SMS (Twilio), and Slack credentials. Update the booking database with valid records (with route, fare, timestamp). Set schedule frequency (e.g., every 6 hours). Review logs regularly to monitor fare alert activity and system health. Prerequisites Valid accounts and credentials for: Amadeus API Skyscanner API SendGrid (or SMTP) for email Twilio for SMS Slack workspace & bot token PostgreSQL or MySQL database for fare tracking Tracked booking dataset (with routes, fares, and user contacts) Customization Options Adjust alert thresholds in the comparison logic (e.g., trigger only if fare drops > \$50 or >10%). Add new notification channels (e.g., WhatsApp, Telegram). Extend logic to track multi-leg or roundtrip fares. Integrate airline refund APIs (where supported) to auto-initiate refund or credit requests. Excel Output Columns When exporting or logging fare tracking data to Excel or CSV, use the following structure: | flight\_number | airline | departure | arrival | departure\_time | arrival\_time | current\_fare | route | timestamp | | -------------- | --------------- | ---------------------------- | ------------------------- | ------------------------- | ------------------------- | ------------- | ------- | ------------------------ | | AT5049 | Royal Air Maroc | John F Kennedy International | Los Angeles International | 2025-07-21T06:00:00+00:00 | 2025-07-21T08:59:00+00:00 | 235 | JFK-LAX | 2025-07-21T13:04:14.000Z | | BA1905 | British Airways | John F Kennedy International | Los Angeles International | 2025-07-21T06:00:00+00:00 | 2025-07-21T08:59:00+00:00 | 479 | JFK-LAX | 2025-07-21T13:04:14.000Z |