by phil
This workflow is your ultimate solution for reliable image retrieval from any web source, including those heavily protected. It operates with a smart, cost-effective strategy: it first attempts to fetch the image using a Classic Image Getter node (a standard, free HTTP request). In approximately 80% of cases, this method will be sufficient. However, for the remaining instances where you encounter IP blocking, CAPTCHAs, rate limiting, or other advanced anti-bot measures, the workflow seamlessly switches to a robust BrightData Web Unblocker service as a fallback. It leverages BrightData’s Image Unblocker to retrieve these blocked images. This template is indispensable for anyone needing consistent and complete access to web images, ensuring you get the data you need without unnecessary overhead. Why Use This Image Scrapper Workflow? Maximum Success Rate**: Retrieves images even from the most challenging or protected websites. Cost-Optimized Strategy**: Prioritizes free, standard HTTP requests, only incurring costs when advanced unblocking is truly necessary. Automated Resilience**: Intelligently handles failed direct attempts by automatically engaging the BrightData failover via the "Unlock Image" node. Versatile Image Scraping**: Perfect for market research, content aggregation, or data enrichment that demands reliable image access. How It Works When clicking ‘Execute workflow’: The workflow is initiated manually, allowing for easy testing and integration into larger processes. image: A Set node defines the target image URL. This can be easily configured to accept dynamic URLs from preceding nodes. Classic Image Getter: This HTTP Request node performs a direct image download. It's the primary, free, and efficient method for readily accessible images. Unlock Image (BrightData Web Unblocker): Configured as an error handler and failover, this HTTP Request node activates only if the "Classic Image Getter" encounters an error. It then routes the image URL through BrightData's Web Unblocker, designed to bypass advanced protective measures and successfully retrieve the image data. 🔑 Prerequisites To enable the advanced capabilities of this workflow, specifically the BrightData Web Unblocker functionality, you will need a BrightData account and a correctly configured Web Unblocker zone. Setting Up Your BrightData Web Unblocker: BrightData Account: Ensure you have an active account with BrightData. If you don't, you can sign up on their website. Create a Web Unblocker Zone: Log in to your BrightData dashboard. Navigate to the "Proxy & Scraping Infrastructure" section, then "Zones." Click "Add new zone." Select "Web Unblocker" as the product type. Give your zone a clear name (e.g., n8n-image-unlocker). Confirm the creation of the zone. Retrieve API Key: Once your Web Unblocker zone is active, go to its settings. Locate your API Key (often referred to as "password" for proxy access) within the "Access Parameters" or "Credentials" section. Configure in n8n: In the Unlock Image HTTP Request node within this workflow, update the Authorization header. Replace "Bearer yourkey" with "Bearer YOUR_BRIGHTDATA_API_KEY". Important: For production workflows, it's highly recommended to use n8n credentials to store your BrightData API key securely, rather than hardcoding it directly in the node. This template uses a placeholder for demonstration purposes. Crucially, ensure that the zone parameter in the Unlock Image node matches the exact Zone ID you created in your BrightData account. You will need to replace the placeholder web_unlocker with your actual BrightData zone ID. Phil | Inforeole
by Sk developer
TikTok Transcript to OpenAI GPT-4 This automation workflow provides a seamless, efficient, and AI-powered solution for extracting, processing, and storing TikTok video subtitles. By combining TikTok Transcript API, OpenAI GPT-4 API, and Google Docs, this workflow transforms the process of transcription and text analysis into a smooth, automated experience. It's perfect for content creators, marketers, and businesses who need to process large volumes of TikTok videos and want to leverage AI for language processing and summarization. How It Works: User Form Submission: The process begins when a user submits a TikTok video URL and specifies the language in which they want the processed content. The data is captured via a simple form that triggers the entire workflow. The form is crucial for collecting the necessary parameters before processing, such as the video link and language preferences. Fetching Subtitles from TikTok: The workflow uses the TikTok Transcript API to retrieve subtitles from the specified TikTok video. This API extracts all textual data associated with the video (including spoken words, captions, etc.) in real-time. The TikTok Transcript API allows you to fetch subtitles efficiently, making it ideal for those who need to process content from TikTok quickly. Advanced Processing with OpenAI GPT-4: Once the TikTok subtitles are fetched, the workflow sends this text to OpenAI’s GPT-4 API. OpenAI's GPT-4 model is renowned for its powerful natural language processing capabilities, making it perfect for handling multi-lingual data. OpenAI GPT-4 API processes the raw transcript in several ways, including: Translation: If the subtitles are in a different language, GPT-4 API can translate them to the desired language. Summarization: GPT-4 API can summarize long TikTok video subtitles into concise points, saving you time and effort. Text Interpretation: You can configure GPT-4 API to generate insights, analyze emotions, or interpret context, which is ideal for detailed content analysis. Storing the Results in Google Docs: After processing the subtitles, the final output (whether it is a translated, summarized, or interpreted version) is automatically saved into a Google Doc. This integration allows the processed text to be stored in an easily editable and shareable format. The document can be accessed by anyone with permission, making it perfect for team collaboration or content management. Workflow Automation: The automation continues with a wait step to ensure that all data is fetched and processed before moving on to storing it in Google Docs. It ensures that the entire process is handled without needing manual intervention, from fetching subtitles to generating results and storing them. Key Features and Benefits: Efficient TikTok Subtitle Extraction: Automatically fetch TikTok video subtitles using the **TikTok Transcript API, eliminating the need for manual transcription. AI-Driven Text Processing: Use the power of **OpenAI GPT-4 API to process the extracted text. GPT-4 API can translate, summarize, or analyze the subtitles for advanced insights, making it far more than just a transcription tool. Seamless Multi-Language Support: **OpenAI GPT-4 API handles multiple languages, translating or summarizing the content based on the user’s input. This makes the workflow versatile for global content creators and marketers. Google Docs Integration: After processing the subtitles, the results are saved directly into **Google Docs for easy access, editing, and sharing. This ensures that all processed data is stored in an organized manner and ready for use in various projects. Time & Effort Savings**: The entire process is automated from start to finish, allowing users to bypass manual transcription and processing tasks. You can focus on creating content while the workflow handles all the repetitive tasks. Advanced Text Insights: By using **OpenAI’s GPT-4 API, you not only get the raw transcript, but you also get insights, summaries, translations, and other interpretations that enhance your content’s value. Challenges Solved: Manual Transcription: This workflow eliminates the need for manual transcription by automatically fetching subtitles from TikTok using the **TikTok Transcript API. Language Barriers: With **OpenAI GPT-4, users can translate TikTok video subtitles into any language, ensuring the content is accessible to a global audience. Content Management: By storing processed content in **Google Docs, this workflow makes it easier to manage and collaborate on transcriptions and analysis, providing a central hub for your data. Automation for Productivity**: This workflow automates every step of the process, from fetching subtitles to analyzing and storing them, freeing up time for higher-value tasks like content creation, strategy planning, or marketing. APIs Integrated: TikTok Transcript API**: Retrieves subtitles directly from TikTok videos, providing the base for further processing. OpenAI GPT-4 API**: Handles advanced text processing, including translation, summarization, and analysis of the TikTok video subtitles. Google Docs API**: Stores processed content into Google Docs, providing a clean, accessible format for viewing and collaboration. Use Cases: Content Creation**: Automatically process and summarize video subtitles for content creation, marketing, or research purposes. Market Research**: Extract and analyze content from TikTok to understand audience sentiment, trending topics, and engagement strategies. Education**: Teachers and educators can use the workflow to analyze educational TikTok videos and save the insights in Google Docs for lesson planning. Conclusion: This TikTok Transcript to OpenAI GPT-4 + Google Docs Automation workflow saves time, enhances content processing with AI, and organizes results into easily accessible documents. By integrating the TikTok Transcript API and OpenAI GPT-4 API, it provides a smart, automated solution for anyone working with TikTok content. Whether you're a content creator, researcher, or marketer, this workflow can help you streamline and optimize your content processing tasks.
by Sherlockes
Purpose of this Template: This template will help us combine multiple RSS sources, curate the content, and send the result to a Telegram channel for easy access. How it Works: We start with two RSS sources from which we primarily want to extract the title, link, and publication date. The workflow will periodically query both sources and use regular expressions and some JavaScript to edit several fields until we have the desired data in the proper format. Once the sources are formatted, we will combine them using the "Merge" node. Since the workflow will run twice a day, we will discard items older than two days. This will be achieved using a "Filter" node. A "Sort" node will arrange the items by publication date so that the most recent ones appear first. To create a flat list in "Markdown" format from the resulting items, we will use a small JavaScript function. This function must escape special characters to ensure they are correctly interpreted. Finally, the list will be sent to the appropriate channel using a "Telegram" node. Configuration Instructions: When opening the workflow for the first time, configure the Telegram credential. In the "RSS" nodes, insert the URLs of the sources to query. In the "Edit Fields" nodes we will have to adjust the regular expressions to obtain the desired result based on our RSS sources. In this case the expressions have been adjusted to obtain the name, size and link of each published file. In the "Sort" node we can modify the maximum age of the elements that we are going to send. In this case, 2 days have been included (2 * 24 * 60 * 60 * 1000 milliseconds) since it is necessary to take into account that Telegram has a maximum message length and if we exceed it, it will return an error instead of sending it. Lastly, include the channel ID in the "Telegram" node where the messages will be sent. Template was created in n8n v1.72.1
by Alessio
A simple node to send new YouTube videos from a channel to a Telegram chat (private, group or channel). CheckTime: set how often videos should be fetched from YouTube. Default is 30 minutes. GetVideosYT: this node will fetch the list of videos from a given channel. Here you need to specify on "Limit" the number of videos to fetch, and on "Channel ID" the ID of the desired channel (it should be the ending part of the URL). You need Google OAuth2 credentials to make it work. A guide is available here. (If you use n8n's tunneling, you may need to adjust the OAuth callback URL on Google Cloud Platform) Set: this node will set some variables to work easily with the next nodes. You shouldn't edit this. Function: this node checks if the video was seen previously by the workflow, so that it won't be published a second time on Telegram. You shouldn't edit this. SendVideo: this node sends the message to Telegram. You need to set your bot's credentials (guide here), specify the Chat ID to send the message (how to get) and personalize the Text of your message. This workflow works correctly only when it's activated. If you manually execute the workflow, it will send every time the latest videos.
by The O Suite
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. How it Works: - Accepts a domain from a web form - Generates a list of Google dorks targeting that domain - Scrapes Google search results for each dork - Filters out junk links (Google internal, non-http) - Formats valid results as a markdown report - Emails the report to your inbox For this workflow you will need to configure credentials in n8n as well as create a Parsera.io agent: 1. AI Scraper Account Create an account for: https://app.parsera.io Create an agent named "Google" and use the URL https://google.com Copy/Paste your Parsera.io API key to the 'Scrape with agent' node in the n8n workflow 2. Gmail OAuth2 Go to Google Cloud Console Enable Gmail API Create OAuth2 credentials Copy/Paste the Gmail API key to the Gmail node in the n8n workflow
by Samir Saci
Tags: Marketing, Image Processing, Automation Context Hey! I’m Samir, a Data Scientist from Paris and the founder of LogiGreen Consulting. We use AI, automation, and data to support sustainable business practices for small, medium and large companies. I implemented this workflow to support an event agency to automate image processing like background removal using Photoroom API. > Automate your photos processing with n8n! This n8n workflow collects all images in a Google Drive folder shared with multiple photographers. For each image, it calls the Photoroom API: A processed image w/o a background is saved in a subfolder Remove Background The original pictures are saved in the subfolder Original This workflow, triggered every morning, will process the backlog of images. 📬 For business inquiries, feel free to connect with me on LinkedIn Who is this template for? This workflow is useful for: Digital Marketing** teams that use images for content creation Photographs* or *Event Organisers** that collect large amounts of photos that need processing What does it do? This n8n workflow: ⏰ Triggers automatically every morning 🖼️ Collects the names and IDs of all images in the folder 🧹 HTTP POST request to Photoroom API to remove the background 📄 Stores the processed image and the original image in two separate sub-folders What do I need to get started? You’ll need: A Google Drive Account connected to your n8n instance with credentials A Photoroom API key that you can get for free (trial) here: Photoroom API Follow the Guide! Follow the sticky notes inside the workflow or check out my step-by-step tutorial on how to configure and deploy it. 🎥 Watch My Tutorial This workflow was built using n8n version 1.93.0 Submitted: May 26, 2025
by AlQaisi
Image Creation with OpenAI and Telegram Check this channel: AutoTechAi_bot Description: In the realm of automation and artificial intelligence, n8n offers a sophisticated platform for seamlessly integrating AI algorithms to enhance image creation and communication processes. This innovative workflow leverages the capabilities of OpenAI and Telegram to facilitate creative image generation and streamline communication channels, ultimately enhancing user engagement and interaction. How to Use: Set Up Credentials: Configure credentials for the Telegram account and OpenAI API to enable seamless integration. Configure Nodes: Telegram Trigger Node: Set up the node to initiate the workflow based on incoming messages from users on Telegram. OpenAI Node: Utilize advanced AI algorithms to analyze text content from messages and generate intelligent responses. Telegram Node: Send processed data, including images and responses, back to users on Telegram for seamless communication. Merge Node: Organize and combine processed data for efficient handling and integration within the workflow. Aggregate Node: Aggregate all item data, including binaries if specified, for comprehensive reporting and analysis purposes. Run Workflow: Initiate the workflow to leverage AI-enhanced image processing and communication capabilities for enhanced user interactions. Monitor Execution: Keep an eye on the workflow execution for any errors or issues that may occur during processing. Customize Workflow: Tailor the workflow nodes, parameters, or AI models to align with specific business objectives and user engagement strategies. Experience Benefits: Embrace the power of AI-driven image processing and interactive communication on Telegram to elevate user engagement and satisfaction levels. By following these steps, businesses can unlock the transformative potential of AI integration in image creation and communication workflows using n8n. Elevate your user engagement strategies and deliver exceptional experiences to your audience through innovative AI-driven solutions. Embark on a journey of innovation and efficiency with AI integration in image creation and communication workflows using n8n!
by Vitali
Workflow Description: This n8n workflow automates the drafting of email replies for Fastmail using OpenAI's GPT-4 model. Here’s the overall process: Email Monitoring: The workflow continuously monitors a specified IMAP inbox for new, unread emails. Email Data Extraction: When a new email is detected, it extracts relevant details such as the sender, subject, email body, and metadata. AI Response Generation: The extracted email content is sent to OpenAI's GPT-4, which generates a personalized draft response. Get Fastmail Session and Mailbox IDs: Connects to the Fastmail API to retrieve necessary session details and mailbox IDs. Draft Identification: Identifies the "Drafts" folder in the mailbox. Draft Preparation: Compiles all the necessary information to create the draft, including the generated response, original email details, and specified recipient. Draft Uploading: Uploads the prepared draft email to the "Drafts" folder in the Fastmail mailbox. Prerequisites: IMAP Email Account: You need to configure an IMAP email account in n8n to monitor incoming emails. Fastmail API Credentials: A Fastmail account with JMAP API enabled. You should set up HTTP Header authentication in n8n with your Fastmail API credentials. OpenAI API Key: An API key from OpenAI to access GPT-4. Make sure to configure the OpenAI credentials in n8n. Configuration Steps: Email Trigger (IMAP) Node: Provide your email server settings and credentials to monitor emails. HTTP Request Nodes for Fastmail: Set up HTTP Header authentication in n8n using your Fastmail API credentials. Replace the httpHeaderAuth credential IDs with your configured credential IDs. OpenAI Node: Configure the OpenAI API key in n8n. Replace the openAiApi credential ID with your configured credential ID. By following these steps and setting up the necessary credentials, you can seamlessly automate the creation of email drafts in response to new emails using AI-generated content. This workflow helps improve productivity and ensures timely, personalized communication.
by Eduard
🎲 Random Data Generator 🎲 This beginner-friendly workflow allows you to generate random mock data effortlessly, without any programming or external dependencies. 🚀 Here's how it works: 📝 Provide some default values for the random generator using the "Code" node. Don't worry, no coding is needed, it's just a simple JSON! You can even use the "Ask AI" in the cloud version of n8n to generate mock values for you. These values are permanently stored in the workflow ensuring swift and reliable execution. 🔀 The "Firstname," "Lastname," and "Occupation" nodes extract invividual variables. Two Merge nodes create a combination of all possible values. Adapt this section based on your needs. 🎲 The "Random" node shuffles the generated data. 🔢 Use the "Limit" node to control the number of output items. ✨ Perform additional post-processing, such as deriving email addresses based on the user's name and surname. And there you have it! With just a few clicks, you'll have a set of realistic mock data ready to test your workflows. 🎉 Want to learn more? Check out the full tutorial on YouTube! 📺 Happy data generating! 😄
by bangank36
This workflow scrapes Trustpilot reviews for a given profile and saves them into Google Sheets. How It Works Clone this Google Sheets template, which includes two sheets: trustpilot A raw collection of Trustpilot reviews. You can customize it as needed. helpfulcrowd This sheet follows the format from this HelpfulCrowd guide, with a slight modification: an added review_id column to support the upsert process. Once the workflow is complete, export the sheet as a CSV and upload it to HelpfulCrowd. For detailed steps, see this post. Running the Workflow You can trigger the workflow on-demand or schedule it to run at a set interval. Requirements Trustpilot business name** (e.g., n8n.io in https://www.trustpilot.com/review/n8n.io). Update this name and pagination settings in the Global node. Google Sheets API credentials** Check out my other templates: 👉 My n8n Templates
by Yaron Been
This workflow provides automated access to the Aihilums Sehatsanjha AI model through the Replicate API. It saves you time by eliminating the need to manually interact with AI models and provides a seamless integration for other generation tasks within your n8n automation workflows. Overview This workflow automatically handles the complete other generation process using the Aihilums Sehatsanjha model. It manages API authentication, parameter configuration, request processing, and result retrieval with built-in error handling and retry logic for reliable automation. Model Description: Advanced AI model for automated processing and generation tasks. Key Capabilities Specialized AI model with unique capabilities** Advanced processing and generation features** Custom AI-powered automation tools** Tools Used n8n**: The automation platform that orchestrates the workflow Replicate API**: Access to the Aihilums/sehatsanjha AI model Aihilums Sehatsanjha**: The core AI model for other generation Built-in Error Handling**: Automatic retry logic and comprehensive error management How to Install Import the Workflow: Download the .json file and import it into your n8n instance Configure Replicate API: Add your Replicate API token to the 'Set API Token' node Customize Parameters: Adjust the model parameters in the 'Set Other Parameters' node Test the Workflow: Run the workflow with your desired inputs Integrate: Connect this workflow to your existing automation pipelines Use Cases Specialized Processing**: Handle specific AI tasks and workflows Custom Automation**: Implement unique business logic and processing Data Processing**: Transform and analyze various types of data AI Integration**: Add AI capabilities to existing systems and workflows Connect with Me Website**: https://www.nofluff.online YouTube**: https://www.youtube.com/@YaronBeen/videos LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Get Replicate API**: https://replicate.com (Sign up to access powerful AI models) #n8n #automation #ai #replicate #aiautomation #workflow #nocode #aiprocessing #dataprocessing #machinelearning #artificialintelligence #aitools #automation #digitalart #contentcreation #productivity #innovation
by Zacharia Kimotho
This workflow is aimed to create new posts in wordpress automatically from an airtable dashboard. When creating content in bulk, we can save time by automating how we can post and publish this content. Usage Get the content from Airtable. Since we have this as a markdown, we will have to convert it to a html format to make it easier to publish and manage on WordPress Upload the blog post with the content, title and all other relevant information needed for an optimized blog Once the post is posted, we need to upload the image and set it as a features image for the blogs Happy productivity