by Keith Rumjahn
Who's this for? If you own a website and need to analyze your Matomo analytics data so you can increse the number of frequent visitors If you need to create an SEO report on what are the common trends amongst your most frequent visitors If you want to grow your site based on suggestions from data Matomo is an analytics tool that can give you details of each individual visitor. Much more powerful than Google analytics. Watch youtube tutorial here Get my SEO A.I. agent system here Read more -> How to create an A.I. Agent to analyze Matomo analytics using n8n for free Here's the A.I. output: Keywords showing the most improvement: Openrouter N8N. Keywords needing attention: Ai Generated Reference Letter Obsidian Second Brain Suggested actions for improvement: Optimize for "best Docker Synology" despite stable ranking, an improvement to top 10 is an achievable goal. Since "2nd brain app for developer" is of interest to a developer. Consider writing a blog post on how the app addresses the specific pain points of developers. Use case Instead of hiring an SEO expert, I run this report weekly. It looks at the data for the past week and looks for visitors with more than 3 visits and recommends ideas to convert more visitors into frequent visitors. How it works The workflow gathers matomo analytics for the past 7 days. We then parse the data The data is sent to Openrouter and using a FREE LLM, it analyses the data. It stores the results in baserow How to use this Input your Matomo analytics credentials Input your Matomo site ID Input your Openrouter.ai credentials Input your baserow credentials You will need to create a baserow database with columns: Dates, Notes, Blog. Created by Rumjahn
by Jimleuk
This n8n template demonstrates the easiest way to build a lead capture flow for your side project, startup or small business where simple works best! If Typeform's costs are getting you down or you feel Google form URLs are off-putting, then definitely give this a try. How it works Our flow begins with a form trigger to capture a newsletter signup and the user's email is captured into a google sheet. Google Sheet is used for demonstration purposes but this could be any database. Multi-page forms allow you to continue the onboarding experience with a short survey. 3 form nodes are chained to capture more details from the user which update the same row in the google sheet. Finally, a form ending node shows a customised completion screen for our user. Check out the example sheet here: https://docs.google.com/spreadsheets/d/15W1PiFjCoiEBHHKKCRVMLmpKg4AWIy9w1dQ2Dq8qxPs/edit?usp=sharing How to use Keeping forms simple may serve to increase form completion rates. If you feel the need to add additional fields, consider breaking them up into more forms and group them contextually. Requirements Google Sheets for data capture Slack for notifications Feel free to swap these out for services that you use! Customising this workflow Play with multi-form design to maximise the opportunity of getting to know the user better. That said, lengthy flows are likely to put people off. Instead of showing a static completion screen, perhaps redirecting to your blog or other more interesting page.
by Aitor | 1Node
This n8n workflow provides a robust error handling and notification system for your n8n workflows. When an error occurs, it automatically logs the error details to Google Sheets, sends a notification to a Telegram channel, and dispatches an email alert, ensuring you're immediately aware of any issues. How it works Error Trigger:** The workflow is activated whenever an error occurs in another n8n workflow. Log Error (Google Sheets):** Error details (e.g., workflow name, error message, timestamp) are appended to a specified Google Sheet, creating a centralized log for all errors. Edit Fields (Manual Configuration):** This node allows you to manually set the Telegram chat ID and recipient email for notifications. Notify in channel (Telegram):** An error notification containing relevant details is sent to your configured Telegram channel. Send email (Gmail):** An email alert with comprehensive error information is sent to the specified recipient. Set up steps This setup will take approximately 10-15 minutes. Download the workflow: Download this workflow and import it into your n8n instance. Configure the Error Trigger: This trigger will automatically activate when an error occurs in any workflow. Make sure you set this workflow as the "Error Workflow" inside the workflows where you want to be alerted. Configure Log error (Google Sheets): Connect your Google Sheets account credentials. Specify the Google Sheet ID and the sheet name where you want to log the errors. Ensure the sheet has appropriate headers (e.g., "Timestamp", "Workflow Name", "Error Message", "Error Details") to receive the data. Configure Edit Fields: In the "Edit Fields" node, manually enter your Telegram chat ID. This is the ID of the chat or channel where you want to receive Telegram notifications. Insert the recipient's email address where you want to receive email alerts. Configure Notify in channel (Telegram): Connect your Telegram account credentials. Ensure the "Chat ID" field is correctly linked to the output from the "Edit Fields" node. Configure Send email (Gmail): Connect your Gmail account credentials. Ensure the "To" email address is correctly linked to the output from the "Edit Fields" node. Customize the subject and body of the email to include relevant error information from the "Error Trigger" node. Test the workflow: To test, you can intentionally create an error in another simple n8n workflow. This error workflow should then trigger this error handling workflow, and you can verify if the log is updated, Telegram message is sent, and email is received. Make sure that the workflow you are testing has the "Error Workflow" selected in the workflow's settings. Requirements n8n instance:** An active n8n instance (self-hosted or cloud). Google Account:** A Google account with access to Google Sheets. Telegram Account:** A Telegram account and a chat/channel ID for notifications. Gmail Account:** A Gmail account to send email alerts. Need help? Feel free to contact us at 1 Node. Get instant access to a library of free resources we created.
by Mauricio Perera
Overview: This workflow is designed to handle user inputs via a webhook, process the inputs with the Google Gemini API (specifically the gemini-2.0-flash-thinking-exp-1219 model), and return a structured response to the user. The response includes three key elements: reasoning, the final answer, and citation URLs (if applicable). This workflow provides a robust solution for integrating AI reasoning into your processes. This workflow can be utilized as a tool for AI-based agents, intelligent email drafting systems, or as a standalone intelligent automation solution. Setup: Webhook Configuration: Ensure the webhook node is properly set up to accept GET requests with an input parameter. Verify that the webhook path matches your application requirements. Test the webhook using tools like Postman to ensure proper data formatting. Google Gemini API Credentials: Set up your Google Gemini API account credentials in the HTTP Request node. Ensure API access and permissions are valid. Parameter Adjustments: Customize the temperature, topK, topP, and maxOutputTokens parameters to fit your use case. Customization: Input Parameters: Modify the webhook path or parameters based on the data your application will send. Response Formatting: Adjust the JavaScript code in the "Process API Response" node to fit your desired output structure. Output Expectations: Test the response returned by the "Return Response to User" node to ensure it meets your application requirements. Workflow Steps: Receive User Input: Node Type: Webhook Purpose: Captures a GET request containing a user-provided input parameter. Acts as the starting point for the workflow. Send Request to Google Gemini: Node Type: HTTP Request Purpose: Sends the received input to the Gemini-2.0-flash-thinking-exp-1219 model for processing. The API configuration includes parameters for customizing the response. Process API Response: Node Type: Code Node Purpose: Extracts reasoning, the final answer, and citation URLs from the API response. Organizes the output for further use. Return Response to User: Node Type: Respond to Webhook Purpose: Sends the processed and structured response back to the user via the webhook. Ensures the response format meets expectations. Expected Outcomes: Input Handling:** Successfully captures user input via a webhook. AI Processing:* Generates a structured response using the *Gemini-2.0-flash-thinking-exp-1219** model, including reasoning, answers, and citations (if available). Output Delivery:** Returns a user-friendly response formatted to your specifications. Notes: The workflow is inactive by default. Each node is annotated with a Sticky Note to clarify its purpose. Ensure all API credentials are correctly configured before execution. Use this workflow to save time, improve accuracy, and automate repetitive tasks efficiently. Tags: Automation Google Gemini AI Agents Intelligent Automation Content Generation Workflow Integration
by WeblineIndia
Automate Telegram Chat Responses Using Google Gemini By WeblineIndia* β‘ TL;DR (Quick Steps) Create a Telegram bot using @BotFather and copy the API Token. Obtain Google Gemini API Key via Google Cloud. Set up the n8n workflow: Trigger: Telegram message received. AI Model: Google Gemini generates response. Output: AI reply sent back to user via Telegram. Customize the system prompt, model, or message handling to suit your use case. π§ Description This n8n workflow enables seamless automation of real-time chat replies in Telegram by integrating with Google Gemini's Chat Model. Every time a user sends a message to your Telegram bot, the workflow routes it through the Gemini AI, which analyzes and crafts a professional response. This reply is then automatically delivered back to the user. The setup acts as a lightweight but powerful chatbot system β ideal for businesses, customer service, or even personal productivity bots. You can easily modify its tone, intelligence level, or logging mechanisms to cater to specific domains such as sales, tech support, or general Q&A. π― Purpose of the Workflow The primary goal of this workflow is to automate intelligent, context-aware chat responses in Telegram using a robust AI model. It eliminates manual reply handling, enhances user engagement, and ensures 24/7 interaction capabilities β all through a no-code or low-code setup using n8n. π οΈ Steps to Configure and Use β Pre-Conditions / Requirements Telegram Bot Token**: Get it from @BotFather. Google Gemini API Key**: Available via Google Cloud PaLM/Gemini API access. n8n Instance**: Hosted or local instance with required nodes installed (Telegram, Basic LLM Chain, and Google Gemini support). π§ Setup Instructions Step 1: Telegram Trigger β Listen for Incoming Messages Add Telegram Trigger node. Select Trigger On: Message. Authenticate using your Telegram Bot Token. This will capture incoming messages from any user interacting with your bot. Step 2: Google Gemini AI β Generate a Smart Reply Add the Basic LLM Chain node. Connect the input message ({{$json.message.text}}) from the Telegram Trigger. System Prompt: > "You are an AI assistant. Reply to the following user message professionally:" Choose Google Gemini Chat Model (models/gemini-1.5-pro). Connect this node to receive the text input and pass it to Gemini for processing. Step 3: Telegram Reply β Send the AI Response Add a Telegram node (Operation: Send Message). Set Chat ID dynamically from the Telegram Trigger node. Input the generated message from the Gemini output. Enable Parse Mode as HTML for rich formatting. Final Step: Link All Nodes Receive Telegram Message β Generate AI Response β Send Telegram Reply. > Tip: Test the workflow by sending a message to your Telegram bot and ensure you receive an AI-generated reply. π§© Customization Guidance βοΈ Modify the AI tone by updating the system prompt. π€ Use other AI models (e.g., OpenAI GPT-4o). π Add filters to respond differently based on specific keywords. π Extend the workflow to store chats in Google Sheets, Airtable, or databases for audit or analytics. π Multi-language support: Add translation layers before and after AI processing. π οΈ Troubleshooting Guide No message received?** Check if your Telegram bot is active and webhook is working. AI not responding?** Validate your Google Gemini API key and usage quota. Wrong replies?** Refine the system prompt or validate message routing. Formatting issues?** Ensure Parse Mode is correctly set to HTML. π‘ Use Case Examples Customer Service Chatbot** for product queries. Educational Bots** for answering user questions on a topic. Mental Health Companion** that gives supportive replies. Event-based Announcers** or automatic responders during off-hours. > And many more! This workflow can be easily extended to support advanced use cases with just a few additional nodes. π¨βπ» About the Creator This workflow is developed by WeblineIndia, a trusted provider of AI development services and process automation solutions. If you're looking to build or customize intelligent workflows like this, we invite you to get in touch with our team. We also offer specialized Python development and AI developer hiring services to supercharge your automation needs.
by Yaron Been
Automated pipeline that exports technology stack data from BuiltWith to Google Sheets for analysis, reporting, and team collaboration. π What It Does Extracts technology stack data Organizes data in Google Sheets Updates automatically on schedule Supports multiple company tracking Enables easy data sharing π― Perfect For Sales teams Market researchers Business analysts Competitive intelligence Technology consultants βοΈ Key Benefits β Centralized technology database β Easy data analysis β Team collaboration β Historical tracking β Custom reporting π§ What You Need BuiltWith API access Google account n8n instance Google Sheets setup π Data Exported Company information Web technologies Hosting details Analytics tools Marketing technologies Contact information π οΈ Setup & Support Quick Setup Start exporting in 15 minutes with our step-by-step guide πΊ Watch Tutorial πΌ Get Expert Support π§ Direct Help Transform raw technology data into actionable business intelligence with automated exports.
by Marketing Canopy
Automate Sports Betting Data with TheOddsAPI This workflow enables you to create and update a table using TheOddsAPI for sports betting data. It automatically pulls upcoming Ice Hockey games at the start of the day and updates the table with results at the end of the day. You can modify it to retrieve odds and game data for any sport. This setup is particularly useful for sports betting applications, such as tracking the results of a predictive model. It leverages scheduled triggers to activate HTTP requests, which then create or update fields in Airtable by matching on the game ID. Prerequisites Before implementing this workflow, ensure you have the following: TheOddsAPI Account & API Key Sign up at TheOddsAPI and obtain an API key. Ensure you have the correct API permissions to access sports odds and results. Airtable Account & API Key Create an account at Airtable and set up a database. Obtain an API key from the Account Settings page. API Access & Rate Limits Review TheOddsAPIβs rate limits and ensure your account tier allows for scheduled API calls. Confirm that Airtable API limits align with your expected data retrieval frequency. Step-by-Step Guide to Integrating TheOddsAPI 1. Schedule API Requests Set up a trigger to automatically pull upcoming Ice Hockey games at the start of each day. 2. Fetch Data from TheOddsAPI Retrieve the latest sports betting data, including game details and odds, using TheOddsAPI. 3. Store Data in Airtable Insert or update records in Airtable by matching game IDs, ensuring data accuracy. Sample Airtable Template Column Setup for Ice Hockey (Table can adjust depending on sport and data needs. Reference TheOddsAPI for more documentation.) Game ID** Sport** League** Game Date (UTC)** Home Team** Away Team** Completed** (Boolean: TRUE/FALSE for game completion status) Scores** (JSON or String for final scores) Last Update** (Timestamp of the latest update) 4. Schedule an End-of-Day Update Configure another trigger to fetch final game results at the end of the day. 5. Update Records in Airtable Modify existing Airtable records with final scores and game outcomes for complete tracking. 6. Customize for Other Sports Adjust API parameters to retrieve data for different sports and betting odds, making the system flexible for multiple use cases. This structured workflow automates sports betting data collection and updates, ensuring accurate and real-time tracking of odds and game results. By integrating TheOddsAPI with Airtable, you can build scalable applications for predictive sports analytics and betting insights.
by Yaron Been
Workflow Overview This cutting-edge n8n automation is a powerful social media intelligence gathering tool designed to transform Instagram profile research into a seamless, automated process. By intelligently combining web scraping, data formatting, and cloud storage technologies, this workflow: Discovers Profile Insights: Automatically scrapes Instagram profile data Captures comprehensive profile metrics Extracts critical social media intelligence Intelligent Data Capture: Retrieves follower counts Collects biographical information Captures profile picture and external links Seamless Data Logging: Automatically stores data in Google Sheets Creates a living, updateable database Enables easy analysis and tracking Key Benefits π€ Full Automation: Instant profile intelligence π‘ Comprehensive Insights: Detailed social media metrics π Effortless Tracking: Automated data collection π Multi-Purpose Research: Flexible data gathering Workflow Architecture πΉ Stage 1: Trigger & Input Form-Based Trigger**: Manual username submission Webhook Support**: Flexible data entry methods User-Driven Initiation** πΉ Stage 2: Web Scraping Apify Integration**: Robust Instagram data extraction Comprehensive Profile Scanning**: Followers count Following count Profile biography Profile picture URL πΉ Stage 3: Data Formatting Intelligent Data Mapping** Standardized Data Structure** Preparation for Storage** πΉ Stage 4: Cloud Logging Google Sheets Integration** Persistent Data Storage** Easy Access and Analysis** Potential Use Cases Influencer Marketing**: Talent identification Competitive Intelligence**: Audience research Social Media Analysis**: Performance tracking Recruitment**: Talent scouting Brand Partnerships**: Collaboration opportunities Setup Requirements Apify Account Instagram scraping actor API token Configured scraping parameters Google Sheets Connected Google account Prepared tracking spreadsheet Appropriate sharing settings n8n Installation Cloud or self-hosted instance Workflow configuration API credential management Future Enhancement Suggestions π€ Advanced profile scoring π Engagement rate calculation π Real-time change alerts π Multi-platform profile tracking π§ AI-powered insights generation Technical Considerations Implement robust error handling Use exponential backoff for API calls Maintain flexible data extraction strategies Ensure compliance with platform terms of service Ethical Guidelines Respect user privacy Use data for legitimate research Maintain transparent data collection practices Provide opt-out mechanisms Connect With Me Ready to unlock social media insights? π§ Email: Yaron@nofluff.online π₯ YouTube: @YaronBeen πΌ LinkedIn: Yaron Been Transform your social media research with intelligent, automated workflows! #InstagramDataScraping #SocialMediaIntelligence #InfluencerMarketing #DataAutomation #AIResearch #MarketingTechnology #SocialMediaAnalytics #ProfileIntelligence #WebScraping #MarketingTech
by Prakash
Who is this for? This workflow is ideal for: Developers** who want to stay updated on issues without constantly checking GitHub. Managers** tracking issue progress in a Telegram group. DevOps teams that need automated notification alerts for new or updated issues. What problem does this workflow solve? Keeping track of GitHub issues manually can be tedious. Teams often miss critical updates because notifications are buried in emails or dashboards. This workflow automates the process by fetching new or open GitHub issues and instantly sending notifications to a specified Telegram chat. What this workflow does This workflow connects GitHub and Telegram to provide real-time issue notifications: Fetch GitHub Issues β Retrieves new or open issues from a selected GitHub repository. Format the Issue Details β Extracts key information like issue title, number, status, and URL. Send to Telegram β Posts the formatted issue details to a Telegram group or private chat. Setup Guide Prerequisites Before setting up the workflow, ensure you have: GitHub Personal Access Token**: Required to fetch issue details. Generate it under Developer Settings with repo or public_repo permissions. Telegram Bot Token**: Create a bot via BotFather on Telegram and obtain the token. Telegram Chat ID**: Find the chat ID where the bot should send messages using this method. Step-by-Step Setup Set Up GitHub Node Authenticate using your GitHub token. Choose the repository you want to track. Configure filters (e.g., fetch only open issues). Format Issue Details Extract key details like title, issue number, assignee, and status. Customize the message structure for better readability. Send Message to Telegram Add the Telegram node and enter your bot token. Use the Chat ID to define the recipient. Format the message to include issue details and links. Schedule the Workflow (Optional) Use the Cron node to run this workflow periodically (e.g., every hour). How to Customize This Workflow Filter Issues by Labels**: Modify the GitHub node to fetch only issues with specific labels. Include Additional Fields**: Add issue comments, priority, or assignee details in the message. Send Alerts Based on Priority**: Use conditional logic to send high-priority issues to a different chat. Trigger on Issue Events**: Instead of fetching periodically, use GitHub webhooks (if permitted in the repo) to trigger the workflow on issue creation or updates. Why Use This Workflow? Automates GitHub issue tracking** without manually checking repositories. Instant notifications in Telegram** ensure quick response times. Fully customizable** to fit different team workflows.
by lin@davoy.tech
This workflow template, "Daily Gratitude Reminder Bot for LINE," is designed to help users cultivate a habit of gratitude by sending personalized, AI-generated reminders every evening at 9:00 PM. Using Azure OpenAI , the bot generates varied and engaging messages to prompt users to reflect on the positive aspects of their day. The reminders are then sent directly to users via the LINE messaging platform , ensuring a seamless and impactful experience. Whether you're a developer, counselor, or business owner, this template offers a customizable and scalable solution for promoting mental wellness and fostering a culture of gratitude. profile Qwen2.5-Max 2:56 pm Hereβs a polished and detailed description for your "Daily Gratitude Reminder Bot for LINE" template, following the submission guidelines provided. This description is structured to meet all the requirements, including SEO optimization, clarity, and user convenience. Daily Gratitude Reminder Bot for LINE This workflow template, "Daily Gratitude Reminder Bot for LINE," is designed to help users cultivate a habit of gratitude by sending personalized, AI-generated reminders every evening at 9:00 PM. Using Azure OpenAI , the bot generates varied and engaging messages to prompt users to reflect on the positive aspects of their day. The reminders are then sent directly to users via the LINE messaging platform , ensuring a seamless and impactful experience. Whether you're a developer, counselor, or business owner, this template offers a customizable and scalable solution for promoting mental wellness and fostering a culture of gratitude. Who Is This Template For? Developers who want to integrate AI-powered workflows into messaging platforms like LINE. Counselors & Therapists looking to encourage mindfulness and emotional well-being among their clients. Businesses & Organizations focused on employee wellness or customer engagement through positive reinforcement. Educators & Nonprofits seeking tools to promote mental health awareness and self-care practices. What Problem Does This Workflow Solve? Gratitude journaling has been proven to improve mental health, reduce stress, and increase overall happiness. However, many people struggle to maintain the habit due to busy schedules or forgetfulness. This workflow solves that problem by automating daily reminders to reflect on positive experiences, making it easier for users to build and sustain a gratitude practice. What This Workflow Does Scheduled Trigger: The workflow is triggered every evening at 9:00 PM using a schedule node. AI-Powered Message Generation: An Azure OpenAI Chat Model generates a unique and engaging reminder message with a temperature setting of 0.9 to ensure variety and creativity. Message Formatting: The generated message is reformatted to comply with the LINE Push API requirements, ensuring smooth delivery. Push Notification via LINE: The formatted message is sent to the user via the LINE Push API , delivering the reminder directly to their chat. Setup Guide Pre-Requisites Access to an Azure OpenAI account with credentials. A LINE Developers Console account with access to the Push API. Basic knowledge of n8n workflows and JSON formatting. How to Customize This Workflow to Your Needs Change the Time: Adjust the schedule trigger to send reminders at a different time. Modify the Prompt: Edit the AI model's input prompt to generate messages tailored to your audience (e.g., focus on work achievements or personal growth). Expand Recipients: Update the LINE Push API node to send reminders to multiple users or groups. Integrate Additional Features: Add nodes to log user responses or track engagement metrics. Why Use This Template? Promotes Mental Wellness: Encourages users to reflect on positive experiences, improving emotional well-being. Highly Customizable: Easily adapt the workflow to suit different audiences and use cases. Scalable: Send reminders to one user or thousands, making it suitable for both personal and organizational use. AI-Powered Creativity: Avoid repetitive messages by leveraging AI to generate fresh and engaging content.
by n8n custom workflows
Introduction The namesilo Bulk Domain Availability workflow is a powerful automation solution designed to check the registration status of multiple domains simultaneously using the Namesilo API. This workflow efficiently processes large lists of domains by splitting them into manageable batches, adhering to API rate limits, and compiling the results into a convenient Excel spreadsheet. It eliminates the tedious process of manually checking domains one by one, saving significant time for domain investors, web developers, and digital marketers. The workflow is particularly valuable during brainstorming sessions for new projects, when conducting domain portfolio audits, or when preparing domain acquisition strategies. By automating the domain availability check process, users can quickly identify available domains for registration without the hassle of navigating through multiple web interfaces. Who is this for? This workflow is ideal for: Domain investors and flippers who need to check multiple domains quickly Web developers and agencies evaluating domain options for client projects Digital marketers researching domain availability for campaigns Business owners exploring domain options for new ventures IT professionals managing domain portfolios Users should have basic familiarity with n8n workflow concepts and a Namesilo account to obtain an API key. No coding knowledge is required, though understanding of domain name systems would be beneficial. What problem is this workflow solving? Checking domain availability one-by-one is a time-consuming and tedious process, especially when dealing with dozens or hundreds of potential domains. This workflow solves several key challenges: Manual Inefficiency: Eliminates the need to individually search for each domain through registrar websites. Rate Limiting: Handles API rate limits automatically with built-in waiting periods. Data Organization: Compiles availability results into a structured Excel file rather than scattered notes or multiple browser tabs. Bulk Processing: Processes up to 200 domains per batch, with the ability to handle unlimited domains across multiple batches. Time Management: Frees up valuable time that would otherwise be spent on repetitive manual checks. What this workflow does Overview The workflow takes a list of domains, processes them in batches of up to 200 domains per request (to comply with API limitations), checks their availability using the Namesilo API, and compiles the results into an Excel spreadsheet showing which domains are available for registration and which are already taken. Process Input Setup: The workflow begins with a manual trigger and uses the "Set Data" node to collect the list of domains to check and your Namesilo API key. Domain Processing: The "Convert & Split Domains" node transforms the input list into batches of up to 200 domains to comply with API limitations. Batch Processing: The workflow loops through each batch of domains. API Integration: For each batch, the "Namesilo Requests" node sends a request to the Namesilo API to check domain availability. Data Parsing: The "Parse Data" node processes the API response, extracting information about which domains are available and which are taken. Rate Limit Management: A 5-minute wait period is enforced between batches to respect Namesilo's API rate limits. Data Compilation: The "Merge Results" node combines all the availability data. Output Generation: Finally, the "Convert to Excel" node creates an Excel file with two columns: Domain and Availability (showing "Available" or "Unavailable" for each domain). Setup Import the workflow: Download the workflow JSON file and import it into your n8n instance. Get Namesilo API key: Create a free account at Namesilo and obtain your API key from https://www.namesilo.com/account/api-manager Configure the workflow: Open the "Set Data" node Enter your Namesilo API key in the "Namesilo API Key" field Enter your list of domains (one per line) in the "Domains" field Save and activate: Save the workflow and run it using the manual trigger. How to customize this workflow to your needs Modify domain input format**: You can adjust the code in the "Convert & Split Domains" node if your domain list comes in a different format. Change batch size**: If needed, you can modify the batch size (currently set to 200) in the "Convert & Split Domains" node to accommodate different API limitations. Adjust wait time**: If you have a premium API account with different rate limits, you can modify the wait time in the "Wait" node. Enhance output format**: Customize the "Convert to Excel" node to add additional columns or formatting to the output file. Add domain filtering**: You could add a node before the API request to filter domains based on specific criteria (length, keywords, TLDs). Integrate with other services**: Connect this workflow to domain registrars to automatically register available domains that meet your criteria.
by phil
This workflow automates the backup of your n8n workflows data to Google Drive every day. It ensures that important configurations and execution logs are securely stored, reducing the risk of data loss and improving workflow resilience. πΉ Why Use This? β Automates routine backups effortlessly. β Reduces manual intervention and potential data loss. β Securely stores critical workflow configurations in Google Drive. With this workflow, you can focus on innovation while n8n takes care of your backups. πβ¨ π How It Works This workflow operates seamlessly with a combination of scheduled triggers, JSON data transformation, and secure cloud storage. π Setup Steps Trigger the backup β Choose between manual execution or automated scheduling at 1:30 AM daily. Data preparation β Your workflow parameters define the backup location and organize files effectively. Transformation & Encoding β The data is processed and converted into a JSON file in base64 format. Cloud Storage β The backup is securely uploaded to your designated Google Drive folder. π§ Customization Options You can modify various aspects of the backup workflow to better suit your needs: 1οΈβ£ Adjusting Backup Frequency By default, the workflow runs daily at 1:30 AM. To change this: Open the Trigger Node in n8n. Modify the Cron Expression or select a different frequency (e.g., hourly, weekly, or custom intervals). 2οΈβ£ Selecting Specific Workflows to Backup Instead of backing up all workflows, you can filter which ones to include: Add a Filter Node before exporting data. Define specific workflow IDs or names to include in the backup. 3οΈβ£ Changing the Backup Destination The default destination is Google Drive, but you can change this: Replace the Google Drive Node with a different storage provider (e.g., Dropbox, AWS S3, or local storage via FTP/SFTP). Configure authentication for the new destination. 4οΈβ£ Modifying Data Format By default, the workflow stores data in JSON format. If you need a different format: Convert JSON to CSV using the Spreadsheet File Node. Store backups in a compressed format (ZIP) by adding a Compression Node. 5οΈβ£ Encrypting the Backup for Extra Security For added protection: Use the Crypto Node to encrypt the JSON file before uploading. Set up an Access-Controlled Folder in Google Drive with limited permissions. β Verify That Your Backup Works Before relying on this workflow for your automated backups, make sure it works correctly by performing a quick test: Manually trigger the workflow in n8n and check if the backup file appears in your Google Drive. Open Google Drive, navigate to the backup folder, and download the JSON file. Verify its content by checking if the data matches your workflowβs execution logs. Try to import the JSON file back into n8n using the βImport Fileβ function to ensure the workflow structure is intact. Alternatively, copy and paste a test file into Google Drive and confirm that it appears correctly in your workflow logs. This quick test will confirm that your backup is running smoothly and that your data is retrievable whenever needed. π How to Find Your Google Drive Directory ID To ensure that the backup is uploaded to the correct folder, you need to retrieve your Google Drive Directory ID. Follow these simple steps: Open Google Drive. Navigate to the folder where you want to store your backups. Click on the folder and check the URL in your browser. The Directory ID is the long string of characters at the end of the URL after /folders/. Example: π If your folder URL is: https://drive.google.com/drive/folders/14oUlH_LW_NT0Xb2woZWvuzRncV-bhla Then, your Directory ID is: 14oUlH_LW_NT0Xb2woZWvuzRncV-bhla Copy this Directory ID and use it in the workflow's parameters to ensure the backup is saved in the correct location. Phil | Inforeole