by Daniel Nolde
What it is: In version 1.78, n8n introduced a dedicated node to use the OpenRouter service, which lets you to use a lot of different LLM models and providers and change models on the fly in an agentic workflow. For prior n8n versions, there's a workaround to make OpenRouter accessible, by using the OpenAI node with a OpenRouter-specific BaseURL. This trivial workflow demonstrates this for version before 1.78, so that you can use different LLM model dynamically with the available n8n nodes for OpenAI LLM and OpenAI credentials. What you can do: Use any of the OpenRouter models Have the model even dynamically configured or changing (by some external config, some rule, or some specific chat message) Setup steps: Import the workflow Ensure you have registered and account, purchased some credits and created and API key for OpenRouter.ai Configure the "OpenRouter" credentials with your own credentials, using an OpenAI type credential, but making sure in the credential's config form its "Base URL" is set to https://openrouter.ai/api/v1 so OpenRouter is used instead of OpenAI. Open the "Settings" node and change the model value to any valid model id from the OpenRouter models list or even have the model property set dynamically
by Ria
This workflow demonstrates how to use the workflowStaticData() function to set any type of variable that will persist within workflow executions. https://docs.n8n.io/code/cookbook/builtin/get-workflow-static-data/ This can be useful for example when working with access tokens that expire after a certain time period. Using staticData we can keep a record of that access token and the expiry time and build our workflow logic around it. Important Static Data only persists across production executions, i.e. triggered by Webhooks or Schedule Triggers (not manual executions!) For this the workflow will have to be activated. Setup configure HTTP Request node to fetch access token from your API (optional) activate workflow test the workflow with the webhook production link you can check the population of the static data in the single executions Feedback If you found this useful or want to report some missing information - I'd be happy to hear from you at ria@n8n.io
by Cameron Wills
Who is this for? Content creators, social media managers, digital marketers, and researchers who need to download original TikTok videos without watermarks for analysis, repurposing, or archiving purposes. What problem does this workflow solve? Downloading TikTok videos without watermarks typically requires using questionable third-party websites that may have limitations, ads, or privacy concerns. This workflow provides a clean, automated solution that can be integrated into your own systems and processes. What this workflow does This workflow automates the process of downloading TikTok videos without watermarks in three simple steps: Fetch the TikTok video page by providing the video URL Extract the raw video URL from the page's HTML data Download the original video file without watermark (Optional) Upload to Google Drive with public sharing link generation The workflow uses web scraping techniques to extract the original video source directly from TikTok's own servers, maintaining the highest possible quality without any added watermarks or branding. Setup (Est. time: 5-10 minutes) Before getting started, you'll need: n8n installation The URL of a TikTok you want to download (Optional) Google Drive API enabled in Google Cloud Console with OAuth Client ID and Client Secret credentials if you want to use the upload feature How to customize this workflow to your needs Replace the example TikTok URL with your desired video links Modify the file naming convention for downloaded videos Integrate with other nodes to process videos after downloading Create a webhook to trigger the workflow from external applications Set up a schedule to regularly download videos from specific accounts This workflow can be extended to support various use cases like trending content analysis, competitor research, creating compilation videos, or building a content library for inspiration. It provides a foundation that can be customized to fit into larger automated workflows for content creation and social media management.
by Guillaume Duvernay
Description This template provides a simple and powerful backend for adding speech-to-text capabilities to any application. It creates a dedicated webhook that receives an audio file, transcribes it using OpenAI's gpt-4o-mini model, and returns the clean text. To help you get started immediately, you'll find a complete, ready-to-use HTML code example right inside the workflow in a sticky note. This code creates a functional recording interface you can use for testing or as a foundation for your own design. Who is this for? Developers:** Quickly add a transcription feature to your application by calling this webhook from your existing frontend or backend code. No-code/Low-code builders:** Embed a functional audio recorder and transcription service into your projects by using the example code found inside the workflow. API enthusiasts:** A lean, practical example of how to use n8n to wrap a service like OpenAI into your own secure and scalable API endpoint. What problem does this solve? Provides a ready-made API:** Instantly gives you a secure webhook to handle audio file uploads and transcription processing without any server setup. Decouples frontend from backend:** Your application only needs to know about one simple webhook URL, allowing you to change the backend logic in n8n without touching your app's code. Offers a clear implementation pattern:** The included example code provides a working demonstration of how to send an audio file from a browser and handle the response—a pattern you can replicate in any framework. How it works This solution works by defining a clear API contract between your application (the client) and the n8n workflow (the backend). The client-side technique: Your application's interface records or selects an audio file. It then makes a POST request to the n8n webhook URL, sending the audio file as multipart/form-data. It waits for the response from the webhook, parses the JSON body, and extracts the value of the Transcript key. You can see this exact pattern in action in the example code provided in the workflow's sticky note. The n8n workflow (backend): The Webhook node catches the incoming POST request and grabs the audio file. The HTTP Request node sends this file to the OpenAI API. The Set node isolates the transcript text from the API's response. The Respond to Webhook node sends a clean JSON object ({"Transcript": "your text here..."}) back to your application. Setup Configure the n8n workflow: In the Transcribe with OpenAI node, add your OpenAI API credentials. Activate the workflow to enable the endpoint. Click the "Copy" button on the Webhook node to get your unique Production Webhook URL. Integrate with the frontend: Inside the workflow, find the sticky note labeled "Example Frontend Code Below". Copy the complete HTML from the note below it. ⚠️ Important: In the code you just copied, find the line const WEBHOOK_URL = 'YOUR WEBHOOK URL'; and replace the placeholder with the Production Webhook URL from n8n. Save the code as an HTML file and open it in your browser to test. Taking it further Save transcripts:* Add an *Airtable* or *Google Sheets** node to log every transcript that comes through the workflow. Error handling:** Enhance the workflow to catch potential errors from the OpenAI API and respond with a clear error message. Analyze the transcript:* Add a *Language Model** node after the transcription step to summarize the text, classify its sentiment, or extract key entities before sending the response.
by bangank36
This workflow restores all n8n instance workflows from GitHub backups using the n8n API node. It complements the Backup Your Workflows to GitHub template by allowing users to seamlessly restore previously saved workflows. How It Works The workflow fetches workflows stored in a GitHub repository and imports them into your n8n instance. Setup Instructions To configure the workflow, update the Globals node with the following values: repo.owner** – Your GitHub username repo.name** – The name of your GitHub repository storing the workflows repo.path** – The folder path within the repository where workflows are stored For example, if your GitHub username is john-doe, your repository is named n8n-backups, and workflows are stored in a workflows/ folder, you would set: repo.owner → john-doe repo.name → n8n-backups repo.path → workflows/ Required Credentials GitHub API** – Access to your repository n8n API** – To import workflows into your n8n instance Who Is This For? This template is ideal for users who want to restore their workflows from GitHub backups, ensuring easy migration and recovery in case of data loss. Check out my other templates: 👉 My n8n Templates
by Airtop
Automating LinkedIn Company URL Verification Use Case This automation verifies that a given LinkedIn URL actually belongs to a company by comparing the website listed on their LinkedIn page against the expected company domain. It is essential for ensuring data accuracy in lead qualification, enrichment, and CRM updates. What This Automation Does Input Parameters Company LinkedIn**: The LinkedIn URL to be verified. Company Domain**: The expected domain (e.g., example.com) for validation. Airtop Profile (connected to LinkedIn)**: Airtop Profile with LinkedIn authentication. Output Confirmation whether the LinkedIn page corresponds to the provided domain. Returns the verified LinkedIn URL if the match is confirmed. How It Works Extracts the website URL from the specified LinkedIn company profile. Compares the extracted URL with the provided company domain. If the domain is contained in the extracted website, the LinkedIn profile is confirmed as valid. Returns the original LinkedIn URL if the match is successful. Setup Requirements Airtop API Key LinkedIn-authenticated Airtop Profile Next Steps Use for LinkedIn Discovery Validation**: Ensure correctness after automated LinkedIn page discovery. Combine with CRM Updates**: Prevent incorrect LinkedIn links from being stored in CRM. Automate in Data Pipelines**: Use this as a validation gate before enrichment or scoring steps.
by Ramsey Njire
Who Is This For? This workflow is perfect for content creators, marketers, and business professionals who receive regular newsletters and want to effortlessly convert them into engaging LinkedIn posts. By automating the extraction and repurposing process, you can save time and consistently share thoughtful updates with your network. What Problem Does This Workflow Solve? Manually reading newsletters, extracting the key points, and then formatting that content into professional, engaging LinkedIn posts can be time-consuming and error-prone. This workflow automates those steps by: Filtering Emails:** Uses the Gmail node to process only those emails from a specific sender (e.g., newsletter@example.com). Extracting Content:** Leverages OpenAI to identify and summarize the top news items in your newsletter. Generating Posts:** Crafts concise, insightful LinkedIn posts in a smart, deadpan style with a touch of subtle humor. Publishing:** Posts the generated content directly to LinkedIn. What This Workflow Does Filter Newsletters:** The Gmail node is set up to only handle emails from your chosen sender, ensuring that only relevant newsletters are processed. Extract Key Content:** An OpenAI node analyzes the newsletter text to pull out the most important news items, including headlines and summaries. Split Content:** A Split Out node divides the extracted content so each news item is processed on its own. Generate LinkedIn Posts:** Another OpenAI node takes each news item's details and produces a well-structured LinkedIn post that delivers practical insights and ends with a reflective observation or question. Publish to LinkedIn:** The LinkedIn node publishes the crafted posts directly to your account. Setup Gmail Node: Rename it to “Filter Gmail Newsletter” and configure it to filter emails by your newsletter sender. OpenAI Nodes: Ensure your OpenAI API credentials are set up correctly. Customize the prompt if needed to match your desired tone. LinkedIn Node: Rename it to “Post to LinkedIn” and confirm that your LinkedIn OAuth2 credentials are properly configured. How to Customize OpenAI Prompts:** Adjust the prompts in the OpenAI nodes to fine-tune the post tone and output formatting. Email Filter:** Change the Gmail filter to match the sender of your newsletters. Post Processing:** Optionally, add extra formatting (using Function nodes) to further enhance the readability of the generated LinkedIn posts. This template offers an automated, hands-off solution to transform your newsletter content into engaging LinkedIn updates, keeping your audience informed and inspired with minimal effort.
by Audun
Send structured logs to BetterStack from any workflow using HTTP Request Who is this for? This workflow is perfect for automation builders, developers, and DevOps teams using n8n who want to send structured log messages to BetterStack Logs. Whether you're monitoring mission-critical workflows or simply want centralized visibility into process execution, this reusable log template makes integration easy. What problem is this workflow solving? Logging failures or events across multiple workflows typically requires duplicated logic. This workflow solves that by acting as a shared log sender, letting you forward consistent log entries from any other workflow using the Execute Workflow node. What this workflow does Accepts level (e.g., "info", "warn", "error") and message fields via Execute Workflow Trigger Sends the structured log to your BetterStack ingestion endpoint via HTTP Request Uses HTTP Header Auth for secure delivery Includes a manual trigger for testing and a sample call to demonstrate usage Comes with clear sticky notes to help you get started Setup Copy your BetterStack Logs ingestion URL. Create a Header Auth credential in n8n with your Authorization: Bearer YOUR_API_KEY. Replace the URL in the HTTP Request node with your BetterStack endpoint. Optionally modify the test data or log levels for custom scenarios. Use Execute Workflow in any of your workflows to send logs here.
by Oneclick AI Squad
This n8n template demonstrates how to create a comprehensive voice-powered restaurant assistant that handles table reservations, food orders, and restaurant information requests through natural language processing. The system uses VAPI for voice interaction and PostgreSQL for data management, making it perfect for restaurants looking to automate customer service with voice AI technology. Good to know Voice processing requires active VAPI subscription with per-minute billing Database operations are handled in real-time with immediate confirmations The system can handle multiple simultaneous voice requests All customer data is stored securely in PostgreSQL with proper indexing How it works Table Booking & Order Handling Workflow Voice requests are captured through VAPI triggers when customers make booking or ordering requests The system processes natural language commands and extracts relevant details (party size, time, food items) Customer data is immediately saved to the bookings and orders tables in PostgreSQL Voice confirmations are sent back through VAPI with booking details and estimated wait times All transactions are logged with timestamps for restaurant management tracking Restaurant Info Provider Workflow Info requests trigger when customers ask about hours, menu, location, or services Restaurant details are retrieved from the restaurant_info table containing current information Wait nodes ensure proper data loading before voice response generation Structured restaurant information is delivered via VAPI in natural, conversational format Database Schema Bookings Table booking_id (PRIMARY KEY) - Unique identifier for each reservation customer_name - Customer's full name phone_number - Contact number for confirmation party_size - Number of guests booking_date - Requested reservation date booking_time - Requested time slot special_requests - Dietary restrictions or special occasions status - Booking status (confirmed, pending, cancelled) created_at - Timestamp of booking creation Orders Table order_id (PRIMARY KEY) - Unique order identifier customer_name - Customer's name phone_number - Contact for order updates order_items - JSON array of food items and quantities total_amount - Calculated order total order_type - Delivery, pickup, or dine-in special_instructions - Cooking preferences or allergies status - Order status (received, preparing, ready, delivered) created_at - Order timestamp Restaurant_Info Table info_id (PRIMARY KEY) - Information entry identifier category - Type of info (hours, menu, location, contact) title - Information title description - Detailed information content is_active - Whether info is currently valid updated_at - Last modification timestamp How to use The manual trigger can be replaced with webhook triggers for integration with existing restaurant systems Import the workflow into your n8n instance and configure VAPI credentials Set up PostgreSQL database with the required tables using the schema provided above Configure restaurant information in the restaurant_info table Test voice commands such as "Book a table for 4 people at 7 PM" or "What are your opening hours?" Customize voice responses in VAPI nodes to match your restaurant's tone and branding The system can handle multiple concurrent voice requests and scales with your restaurant's needs Requirements VAPI account for voice processing and natural language understanding PostgreSQL database for storing booking, order, and restaurant information n8n instance with database and VAPI integrations enabled Customising this workflow Voice AI automation can be adapted for various restaurant types - from quick service to fine dining establishments Try popular use-cases such as multi-location booking management, dietary restriction handling, or integration with existing POS systems The workflow can be extended to include payment processing, SMS notifications, and third-party delivery platform integration
by Wyeth
Encode JSON to Base64 String in n8n This example workflow demonstrates how to convert a JSON object into a base64-encoded string using n8n’s built-in file processing capabilities. This is a common requirement when working with APIs, webhooks, or SaaS integrations that expect payloads to be base64-encoded. > Tip: The three green-highlighted nodes (Stringify → Convert to File → Extract from File) can be wrapped in a Subworkflow to create a reusable Base64 encoder in your own projects. 🔧 Requirements Any running n8n instance (local or cloud) No credentials or external services required What This Workflow Does Generates example JSON data Converts the JSON to a string Saves the string as a binary file Extracts the file’s contents as a base64 string Outputs the base64 string on the final node Step-by-Step Setup Manual Trigger Start the workflow using the Manual Execution node. This is useful for testing and development. Create JSON Data The Create Json Data node uses raw mode to construct a sample object with all major JSON types: strings, numbers, booleans, nulls, arrays, nested objects, etc. Convert to String The Convert to String node uses the expression ={{ JSON.stringify($json) }} to flatten the object into a single string field named json_text. Convert to File The Convert to File node takes the json_text value and saves it to a UTF-8 encoded binary file in the property encoded_text. Extract from File This node takes the binary file and extracts its contents as a base64-encoded string. The result is saved in the base64_text field. Customization Tips Replace the sample JSON in the Create Json Data node with your own payload structure. To make this reusable, extract the three core nodes into a Subworkflow or wrap them in a custom Function. Use the base64_text output field to post to APIs, store in databases, or include in webhook responses.
by Yaron Been
⚡ AI Copywriter Pro: Instant Ad Copy & CTA Generator Transform product details into compelling marketing copy in seconds. This intelligent n8n workflow takes basic product information and generates professional ad copy with powerful calls-to-action using GPT-4, then distributes the results to your team via Slack or stores them in Airtable for future use. 🚀 How It Works This streamlined 4-step automation delivers professional marketing copy instantly: Step 1: Simple Form Input Users submit basic product information through a clean web form - just product name and key features required. Step 2: AI-Powered Generation The workflow sends product details to GPT-4 with a specialized copywriting prompt that creates: Compelling 2-sentence ad copy optimized for conversions 3 powerful call-to-action phrases with different urgency levels Professional tone tailored for marketing campaigns Step 3: Structured Processing The AI response is automatically parsed into clean, labeled fields ready for immediate use across marketing channels. Step 4: Multi-Channel Distribution Generated content is simultaneously: Posted to your designated Slack channel for team review Logged in Airtable for campaign tracking and reuse ⚙️ Setup Steps Prerequisites Active OpenAI API account with GPT-4 access Slack workspace with admin permissions Airtable account for content storage n8n instance (cloud or self-hosted) Configuration Steps 1. Credential Setup OpenAI API Key**: Required for GPT-4 content generation Slack OAuth2**: Needed for posting messages to channels Airtable Personal Access Token**: Essential for database operations 2. Form Configuration The workflow automatically generates a web form at: your-n8n-instance.com/form/[webhook-id] Form fields are pre-configured for: Product Name (text input) Product Features (textarea for detailed descriptions) Form can be embedded on websites or shared as standalone link 3. Slack Integration Setup Connect your Slack workspace to n8n Create or select target channel for ad copy notifications Update the channelId parameter with your channel's ID Test connection with sample message 4. Airtable Database Preparation Create new base or use existing one Set up table with columns: Product Name (Single line text) Product Features (Long text) Ad Copy (Long text) CTAs (Long text) Copy base and table IDs from Airtable URLs Update workflow parameters accordingly 5. AI Prompt Customization The default prompt generates versatile marketing copy, but you can customize for: Industry-specific terminology Brand voice and tone Target audience demographics Campaign objectives (awareness, conversion, retention) 🎯 Use Cases E-commerce & Retail Product Launch Campaigns**: Generate copy for new product announcements Seasonal Promotions**: Create urgency-driven messaging for sales events Category Pages**: Develop compelling descriptions for product collections Social Media Ads**: Produce scroll-stopping copy for Facebook and Instagram SaaS & Technology Feature Announcements**: Craft clear, benefit-focused messaging Free Trial Campaigns**: Generate conversion-optimized trial signup copy Case Study Promotions**: Create compelling success story headlines Webinar Marketing**: Develop registration-driving event descriptions Agency & Freelance Client Presentations**: Rapidly prototype copy concepts for pitches A/B Testing**: Generate multiple copy variations for campaign testing Campaign Brainstorming**: Kickstart creative sessions with AI-generated ideas Content Calendar Filling**: Populate social media schedules efficiently Startup & Small Business MVP Marketing**: Create professional copy without hiring agencies Investor Pitches**: Generate compelling product descriptions Website Content**: Populate landing pages with conversion-focused copy Email Campaigns**: Develop subject lines and promotional content Content Marketing Blog Post Promotions**: Generate social media copy for article shares Newsletter CTAs**: Create compelling subscription and engagement prompts Video Descriptions**: Develop YouTube and social video copy Podcast Promotions**: Generate episode descriptions and listener CTAs 🔧 Advanced Customization Options Multi-Tone Generation Modify the prompt to generate different copy styles: Professional & Corporate: "Generate formal, trust-building copy..." Casual & Friendly: "Create conversational, approachable messaging..." Urgent & Scarcity: "Develop time-sensitive, action-driving copy..." Luxury & Premium: "Craft sophisticated, high-end positioning..." Batch Processing Enhancement Add nodes for processing multiple products: Split in Batches**: Handle bulk product lists Merge**: Combine results for team review Item Lists**: Process CSV uploads of product catalogs Quality Control Integration Implement approval workflows: Human Review Step**: Add manual approval before publishing Sentiment Analysis**: Score copy for emotional impact Brand Guidelines Check**: Validate against company voice standards A/B Test Setup**: Generate multiple variations automatically Extended Distribution Connect additional platforms: Email Marketing**: Send to Mailchimp, ConvertKit, or HubSpot Social Media**: Post directly to Facebook, Twitter, LinkedIn CMS Integration**: Push to WordPress, Webflow, or custom sites Project Management**: Create tasks in Asana, Trello, or Monday.com 📊 Output Examples Sample Input: Product Name: EcoSmart Water Bottle Product Features: Self-cleaning UV technology, 24-hour temperature retention, BPA-free stainless steel, leak-proof design, 500ml capacity Generated Ad Copy: Stay hydrated and eco-conscious with the EcoSmart Water Bottle—featuring revolutionary self-cleaning UV technology that eliminates 99.9% of bacteria while keeping your drinks perfectly chilled for 24 hours! Experience the future of hydration with premium stainless steel construction that's as durable as it is sustainable. CTAs: • Order Your EcoSmart Bottle Today • Join the Clean Hydration Revolution • Get 24-Hour Freshness Now Slack Notification Format: 🎯 NEW AD COPY GENERATED Product: EcoSmart Water Bottle 📝 Ad Copy: Stay hydrated and eco-conscious with the EcoSmart Water Bottle—featuring revolutionary self-cleaning UV technology that eliminates 99.9% of bacteria while keeping your drinks perfectly chilled for 24 hours! Experience the future of hydration with premium stainless steel construction that's as durable as it is sustainable. 🔥 Call-to-Actions: • Order Your EcoSmart Bottle Today • Join the Clean Hydration Revolution • Get 24-Hour Freshness Now Ready to use in campaigns! 🚀 🛠️ Troubleshooting & Best Practices Common Issues & Solutions API Rate Limits Monitor OpenAI usage dashboard Implement delay nodes for high-volume processing Consider upgrading to higher tier plans Inconsistent Output Quality Refine prompts with specific examples Add output validation rules Include brand guidelines in system prompt Integration Failures Verify all API credentials are current Test connections individually before full workflow Check service status pages for outages Optimization Strategies Cost Management Use GPT-4o-mini for routine copy generation Reserve GPT-4 for premium campaigns Cache common responses to reduce API calls Quality Enhancement A/B test different prompt variations Collect team feedback on generated copy Maintain library of high-performing examples Workflow Efficiency Set up monitoring alerts for failures Create backup workflows for critical campaigns Document customizations for team reference 📈 Performance Tracking Key Metrics to Monitor Generation Speed**: Average time from input to output Approval Rate**: Percentage of AI copy used without edits Campaign Performance**: Click-through rates of AI-generated content Cost Per Copy**: API costs per generated piece Success Indicators Reduced copywriting turnaround time (target: under 30 seconds) Increased campaign launch frequency Higher team satisfaction with copy quality Measurable improvement in conversion rates 📞 Questions & Support Need help setting up or customizing your AI Copywriter Pro workflow? 📧 Direct Support Email**: Yaron@nofluff.online Response Time**: Within 24 hours on business days Specialties**: Custom prompt engineering, integration troubleshooting, workflow optimization 🎥 Video Tutorials YouTube Channel**: https://www.youtube.com/@YaronBeen/videos Complete setup walkthrough Advanced customization techniques Integration with popular marketing tools Prompt engineering masterclass 🤝 Professional Network LinkedIn**: https://www.linkedin.com/in/yaronbeen/ Connect for ongoing automation support Share your copywriting success stories Get early access to new workflow templates 💬 What to Include in Support Requests Your specific industry or use case Current marketing tools and platforms Sample products you want to test with Any custom requirements or brand guidelines Screenshots of any errors or unexpected outputs Ready to revolutionize your copywriting process? Deploy this workflow and watch your marketing team's productivity soar while maintaining professional, conversion-focused messaging!
by Niklas Hatje
Use Case In most companies, employees have a lot of great ideas. That was the same for us at n8n. We wanted to make it as easy as possible to allow everyone to add their ideas to some formatted database - it should be somewhere where everyone is all the time and could add a new idea without much extra effort. Since we're using Slack, this seemed to be the perfect place to easily add ideas and collect them in Notion. What this workflow does This workflow waits for a webhook call within Slack, that gets fired when users use the /idea command on a bot that you will create as part of this template. It then checks the command, adds the idea to Notion, and notifies the user about the newly added idea as you can see below: Creating your Slack bot Visit https://api.slack.com/apps, click on New App and choose a name and workspace. Click on OAuth & Permissions and scroll down to Scopes -> Bot token Scopes Add the chat:write scope Head over to Slash Commands and click on Create New Command Use /idea as the command Copy the test URL from the Webhook node into Request URL Add whatever feels best to the description and usage hint Go to Install app and click install Setup Add a Database in Notion with the columns Name and Creator Add your Notion credentials and add the integration to your Notion page. Fill the setup node below Create your Slack app (see other sticky) Click Test workflow and use the /idea comment in Slack Activate the workflow and exchange the Request URL with the production URL from the webhook How to adjust it to your needs You can adjust the table in Notion and for example, add different types of ideas or areas that they impact You might wanna add different templates in Notion to make it easier for users to fill their ideas with details Rename the Slack command as it works best for you How to enhance this workflow At n8n we use this workflow in combination with some others. E.g. we have the following things on top: We additionally have a /bug Slack command that adds a new bug to Linear. Here we're using AI to classify the bugs and move it to the right team. (see this template and this template) We also added other types, like /pain to be less solution-driven To make it easier for everyone to give input, we added a Votes column that allows everyone to vote on ideas/pain points in the list We're also running a workflow once a week that highlights the most popular new ideas and the most active voters (see here)