by Sk developer
๐ฅ YouTube to MP3 Converter Workflow โ Auto Upload to Google Drive & Email Link with RapidAPI This n8n workflow allows users to convert any YouTube video to MP3 using the YouTube To MP3 Fast API on RapidAPI. It handles the entire process end-to-end โ from form input to automatic delivery via email with cloud storage. ๐ How It Works ๐ User Form Input Users submit a YouTube URL and their email through a simple form. ๐ฌ MP3 Conversion The video URL is sent to the YouTube To MP3 Fast API on RapidAPI to process and convert into MP3 format. โณ Wait for Processing Introduces a short delay to ensure the file is ready for download. ๐ฅ File Download The converted MP3 file is downloaded securely. โ๏ธ Google Drive Upload The MP3 is uploaded to Google Drive for cloud-based access and storage. ๐ Share Permissions Permissions are updated to allow anyone with the link to access the MP3. ๐ง Email Notification The final download link is emailed to the user. โ Benefits Completely Automated** โ From input to email, everything runs without manual effort. Cloud Storage Integration** โ Files are saved to your Google Drive for easy management. User-Friendly** โ Simple interface for users to get their MP3 in just a few clicks. Fast & Reliable Conversion** โ Leveraging the YouTube To MP3 Fast API on RapidAPI ensures quick and accurate results. Time-Saving** โ No need for manual downloads, converters, or email composition. โ Problems This Workflow Solves โ Manual Download Hassles โ No need to use third-party websites or shady downloaders. โ Slow or Unreliable Conversions โ The workflow uses the robust YouTube To MP3 Fast API on RapidAPI for reliable performance. โ Missing Files or Lost Links โ Files are securely stored on Google Drive with shareable links. โ User Follow-up โ Automatically sends download links via email, saving you support effort. ๐ง Tech Stack ๐ง n8n โ No-code automation engine ๐ YouTube To MP3 Fast API on RapidAPI โ Fast MP3 conversion service โ๏ธ Google Drive โ Cloud storage and file sharing ๐ค SMTP โ Email delivery system > ๐ Built for creators, educators, and automation fans who need fast YouTube-to-MP3 conversion with effortless delivery.
by Yaron Been
This workflow provides automated access to the Digitalhera Herathaisbragatto AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Digitalhera Herathaisbragatto model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Digitalhera/herathaisbragatto AI model Digitalhera Herathaisbragatto**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Rosh Ragel
Programatically Pull Square Report Data Into N8N What It Does This sub-workflow connects to the Square API and generates a daily sales summary report for all of your Square locations. The report matches the figures displayed in the Square Dashboard > Reports > Sales Summary. Itโs designed to be reused in other workflows, ideal for reporting, data storage, accounting, or automation. Prerequisites To use this workflow, you'll need: Square API credentials (configured as a Header Auth credential) How to Set Up Square Credentials: Go to Credentials > Create New Choose Header Auth Set the Name to "Authorization" Set the Value to your Square Access Token (e.g., Bearer <your-api-key>) How It Works Trigger: The workflow is triggered as a sub-workflow, requiring a report_date input. Fetch Locations: An HTTP request gets all Square locations linked to your account. Fetch Orders: For each location, an HTTP request pulls completed orders for the specified report_date. Filter Empty Locations: Locations with no sales are ignored. Aggregate Sales Data: A Code node processes the order data and produces a summary identical to Squareโs built-in Sales Summary report. Output: A cleaned, consistent summary that can be consumed by parent workflows or other nodes. Example Use Cases Automatically store daily sales data in Google Sheets, MySQL, or PostgreSQL for analysis and historical tracking Automatically send daily email or Slack reports to managers or finance teams Build weekly/monthly reports by looping over multiple dates Push sales data into accounting software like QuickBooks or Xero for automated bookkeeping Calculate commissions or rent payments based on sales volume How to Use Configure both HTTP Request nodes to use your Square API credential. If you are not in the Toronto/New York Timezone, please change the "start_at" and "end_at" parameters in the second HTTP node from "-05:00" to your local timezone Use as a sub-workflow inside a main workflow. Pass a report_date (formatted as YYYY-MM-DD) to the sub-workflow when you call it. Customization Options Add pagination to handle locations with more than 1,000 orders per day. Expand the workflow to save or send the report output via additional integrations (email, database, webhook, etc.). Why It's Useful This workflow saves time, reduces manual report pulling from Square, and enables smarter automation around sales dataโwhether for operations, finance, or performance monitoring.
by Marth
How It Works โ๏ธ This workflow acts as a communication bridge for your candidate pipeline: Webhook Trigger (Status Update): ๐ The workflow activates when it receives data indicating a candidate's status has changed. This data could come from an internal form, a custom script, or a webhook from a basic Applicant Tracking System (ATS). Extract & Prepare Data (Function): ๐งน This node processes the incoming data. It extracts key information such as the candidate's name, the position they applied for, their previous status (if available), and their new status. It then formats this information into a clear, concise message suitable for a notification. Send Slack Notification: ๐ข The prepared message is sent to a designated Slack channel (e.g., #recruitment-updates). This provides instant, real-time updates to your team, ensuring everyone is on the same page. (Alternative: Send Email Notification): This node can easily be swapped with a Gmail or SendGrid node to send email notifications to a predefined list of recipients instead of Slack. How to Set Up ๐ ๏ธ Follow these steps carefully to get your "Automated Candidate Status Notifier" workflow up and running: Import Workflow JSON: Open your n8n instance. Click on 'Workflows' in the left sidebar. Click the '+' button or 'New' to create a new workflow. Click the '...' (More Options) icon in the top right. Select 'Import from JSON' and paste the entire JSON code for this workflow. Configure Webhook Trigger (Status Update): Locate the 'Webhook Trigger (Status Update)' node (1. Webhook Trigger). Activate the workflow. n8n will provide a unique 'Webhook URL'. Crucial Step: Configure your data-sending system (e.g., a form submission, an ATS's webhook settings, or your custom script) to send candidate status update data (preferably in JSON format via POST request) to this n8n Webhook URL. Configure Extract & Prepare Data (Function): Locate the 'Extract & Prepare Data' node (2. Extract & Prepare Data). Adjust Field Names: Review the functionCode inside this node. You MUST adjust the variable assignments (e.g., inputData.candidateName, inputData.position) to accurately match the exact field names your sending system uses for candidate name, position, new status, old status, and notes. Use the 'Test Workflow' feature after sending a test webhook to inspect the incoming items[0].json.body data structure. The node automatically formats messages for Slack and Email. Configure Send Slack Notification: Locate the 'Send Slack Notification' node (3. Send Slack Notification). Credentials: Select your existing Slack API credential or click 'Create New' to set one up. Replace YOUR_SLACK_CREDENTIAL_ID with the actual ID or name of your credential from your n8n credentials. Channel: Replace YOUR_SLACK_CHANNEL_ID_OR_NAME with the exact ID or name of the Slack channel where you want to receive notifications (e.g., #recruitment-updates). OPTIONAL: Switch to Email Notification (Gmail/SendGrid/etc.): Delete the 'Send Slack Notification' node. Add a new 'Gmail' or 'SendGrid' (or your preferred email service) node. Configure its credentials. Set the 'To Email' field (e.g., your-team-email@example.com). Set the 'Subject' to ={{ $json.emailSubject }}. Set the 'HTML' body to ={{ $json.emailBody }}. Connect it from the 'Extract & Prepare Data' node. Review and Activate: Thoroughly review all node configurations. Ensure all placeholder values (like YOUR_...) are replaced and settings are correct. Click the 'Save' button in the top right corner. Finally, toggle the 'Inactive' switch to 'Active' to enable your workflow. ๐ข Your automated candidate status notifier is now live, keeping your team updated in real-time!
by Usman Liaqat
This workflow listens for incoming WhatsApp messages that contain media (e.g., images) and automatically downloads the media file using WhatsApp's private media URL. The trigger node activates when a WhatsApp message with media is received. The media ID is extracted from the message payload. A private media URL is retrieved using the media ID. The media file is downloaded using an authenticated HTTP request. Ideal for: Archiving WhatsApp media to external systems. Triggering further automations based on received media. Integrating with cloud storage like Google Drive, Dropbox, or Amazon S3. Set up steps Connect your WhatsApp Business API account. Add HTTP credentials for downloading media via private URL. Set up the webhook in your WhatsApp Business account. Extend the workflow as needed for your use case (e.g., file storage, alerts).
by Yaron Been
Description This workflow automatically searches multiple freelance platforms for new gigs matching your skills and requirements. It saves you time by eliminating the need to manually check multiple job boards and sends you alerts for relevant opportunities. Overview This workflow automatically scrapes freelance job boards and platforms for new gigs matching your skills and requirements. It uses Bright Data to access job listings and can notify you of new opportunities or save them to a database. Tools Used n8n:** The automation platform that orchestrates the workflow. Bright Data:** For scraping freelance platforms like Upwork, Fiverr, Freelancer, etc. without getting blocked. (Optional) Email/Slack/Database:** For notifications or data storage. How to Install Import the Workflow: Download the .json file and import it into your n8n instance. Configure Bright Data: Add your Bright Data credentials to the Bright Data node. Set Up Notifications: Configure how you want to receive job alerts. Customize: Add your skills, rate requirements, and other filters. Use Cases Freelancers:** Get notified of new gigs matching your skills. Agencies:** Monitor job boards for potential client opportunities. Remote Workers:** Track new remote job postings across multiple platforms. Connect with Me Website:** https://www.nofluff.online YouTube:** https://www.youtube.com/@YaronBeen/videos LinkedIn:** https://www.linkedin.com/in/yaronbeen/ Get Bright Data:** https://get.brightdata.com/1tndi4600b25 (Using this link supports my free workflows with a small commission) #n8n #automation #freelance #brightdata #webscraping #remotework #freelancejobs #gigeconomy #upwork #fiverr #freelancer #jobsearch #remotejobs #freelanceopportunities #n8nworkflow #workflow #nocode #jobhunting #freelancegigs #jobscraper #jobnotifications #freelancecareer #digitalnomad #workfromhome #jobopportunities #freelancetools #jobmonitoring #freelancesuccess
by Yassin Zehar
Context: This workflow automatically sends you a daily message on Telegram summarizing all your meetings and events for the day, straight from your Google Calendar. For who ? Perfect for anyone who: Uses Google Calendar to manage their schedule. Wants Telegram reminders for daily events. Loves automation and productivity tools. Requirements: Telegram. Google account. Google Calendar. Steps: ๐๏ธ Use the sticky notes in the n8n canvas to: Add your Telegram and Google credentials. Execute and test the workflow. Check if you receive your daily summary on Telegram. You'll get this: Tutorial video: Watch the Youtube Tutorial video How does it work? โฐ The trigger runs every day at 7AM. ๐ Your Google Calendar is checked. ๐ข If there are events or meetings, a number > 0 is returned. Otherwise, it's 0. ๐ A text message is generated with a summary of all your events, including all relevant details. โ If no events are found, a "no event" message is sent. About me : I'm Yassin, IT Project Manager, Agile & Data specialist. Scaling tech products with data-driven project management. ๐ฌ Feel free to connect with me on Linkedin
by digi-stud.io
Adobe developer API Did you know that Adobe provides an API to perform all sort of manipulation on PDF files : Split PDF, Combine PDF OCR Insert page, delete page, replace page, reorder page Content extraction (text content, tables, pictures) ... The free tier allows up to 500 PDF operation / month. As it comes directly from Adobe, it works often better than other alternatives. Adobe documentation: https://developer.adobe.com/document-services/docs/overview/pdf-services-api/howtos/ https://developer.adobe.com/document-services/docs/overview/pdf-extract-api/gettingstarted/ What does this workflow do The API is a bit painful to use. To perform a transformation on a PDF it requires to Authenticate and get a temporal token Register a new asset (file) Upload you PDF to the registered asset Perform a query according to the transformation requested Wait for the query to be proccessed by Adobe backend Download the result This workflow is a generic wrapper to perform all these steps for any transformation endpoint. I usually use it from other workflow with an Execute Workflow node. Examples are given in the workflow. Example use case This service is useful for example to clean PDF data for an AI / RAG system. My favorite use-case is to extract table as images and forward images to an AI for image recognition / description which is often more accuarate than feedind raw tabular data to a LLM.
by bangank36
This workflow retrieves all Shopify Orders and saves them into a Google Sheets spreadsheet using the Shopify Admin REST API. It uses pagination to ensure all orders are collected efficiently. I originally built this workflow for my own use and found it valuable for understanding how Shopify pagination works. Now, Iโm sharing it to help others automate their order retrieval process. How It Works Instead of relying on the built-in Shopify node (Get Orders Many), this workflow leverages the HTTP Request node to fetch paginated chunks manually. Shopify uses cursor-based pagination (page_info) instead of traditional page numbers. Pagination data is stored in the response headers, so we need to enable Include Response Headers and Status in the HTTP Request node. You can modify the limit parameter to control batch sizes and optimize for rate limits. This workflow can be run on demand or scheduled to keep your data up to date. Parameters You can adjust these parameters in the HTTP Request node: limit โ The number of orders per request (default: 50, max: 250). fields โ Comma-separated list of fields to retrieve. page_info โ Used for pagination; only limit and fields are allowed when paginating. ๐ Note: when you query the paginated chunks with page_info, only the limit and fields parameters are allowed Credentials Shopify API Key โ Required for authentication. Google Sheets API credentials โ Needed to insert data into the spreadsheet. ๐พ Clone the Google Sheets template here Who Is This For? Shopify store owners who need to export all orders to Google Sheets. Users who want full control over API parameters for optimized queries. Anyone looking for a flexible and scalable Shopify data extraction solution. Explore More Templates ๐ Check out my other n8n templates
by Yaron Been
This workflow provides automated access to the Vcollos Trefilio AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Vcollos Trefilio model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Vcollos/trefilio AI model Vcollos Trefilio**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Louis
Workflow Overview This workflow automates the process of updating a Spotify playlist with tracks from a YouTube playlist, ensuring no duplicates are added. Key Components Manual Trigger: Starts the workflow when you click โTest workflowโ. Spotify Integration: Retrieves tracks from a specified Spotify playlist. YouTube Integration: Fetches tracks from a designated YouTube playlist. Batch Processing: Processes tracks in batches to handle multiple items efficiently. Track Search: Searches for YouTube tracks on Spotify to find corresponding IDs. Comparison: Compares existing Spotify tracks with YouTube tracks to identify which ones to add. Track Addition: Adds new Spotify tracks to the playlist that are not already included. If you have any questions or need clarification, feel free to ask!
by Ficky
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. ๐ n8n Workflow: Meetup Registration & Giveaway Picker This n8n workflow is designed to handle both participant registration and giveaway winner selection, fully integrated with a frontend experience. ๐ Participant Registration Flow This part of the workflow automates the registration process for meetup attendees. ๐ Steps: ๐ Participant Form (Webhook Trigger) Triggered when a user submits the form. Captures fields like nama_lengkap, email, whatsapp, discord_username, and more. ๐ ๏ธ Data Mapping Maps raw form inputs into a structured format. Supports trimming, validation, and transformation as needed. ๐พ Save to Database Performs an upsert operation to store participant info in PostgreSQL. Prevents duplicate entries based on unique ID (e.g., WhatsApp or email). โ Confirmation Screen Returns a success message or thank-you page after registration is complete. ๐ Giveaway App This section serve frontend app to display and select random winners. ๐ Steps: ๐ Webhook GET (Giveaway App) Exposes a public endpoint that delivers a ready-to-use HTML app. Can be opened in a browser or projected during the event. ๐ฅ Fetch Participants Executes a SQL query to retrieve all participant records from the database. ๐งฎ Format Participant Data Redacts personal details (e.g., masks WhatsApp numbers). Encodes the id using Base64 for privacy and uniqueness. ๐ค Render Giveaway App Returns an HTML-based Single Page Application with the participant list included. Allows the host to click a button to pick random winners live. โ Use Case Highlights Streamlined participant collection and storage using n8n workflows Secure data handling with redaction and ID encoding Frontend integration for a fun, interactive giveaway experience Reusable for other community events, workshops, or internal team draws