by Ranjan Dailata
Who this is for? The LinkedIn Company Story Generator is an automated workflow that extracts company profile data from LinkedIn using Bright Data's web scraping infrastructure, then transforms that data into a professionally written narrative or story using a language model (e.g., OpenAI, Gemini). The final output is sent via webhook notification, making it easy to publish, review, or further automate. This workflow is tailored for: Marketing Professionals**: Seeking to generate compelling company narratives for campaigns. Sales Teams**: Aiming to understand potential clients through summarized company insights. Content Creators**: Looking to craft stories or articles based on company data. Recruiters**: Interested in obtaining concise overviews of companies for talent acquisition strategies. What problem is this workflow solving? Manually gathering and summarizing company information from LinkedIn can be time-consuming and inconsistent. This workflow automates the process, ensuring: Efficiency**: Quick extraction and summarization of company data. Consistency**: Standardized summaries for uniformity across use cases. Scalability**: Ability to process multiple companies without additional manual effort. What this workflow does The workflow performs the following steps: Input Acquisition**: Receives a company's name or LinkedIn URL as input. Data Extraction**: Utilizes Bright Data to scrape the company's LinkedIn profile. Information Parsing**: Processes the extracted HTML content to retrieve relevant company details. Summarization**: Employs AI Google Gemini to generate a concise company story. Output Delivery**: Sends the summarized content to a specified webhook or email address. Setup Sign up at Bright Data. Navigate to Proxies & Scraping and create a new Web Unlocker zone by selecting Web Unlocker API under Scraping Solutions. In n8n, configure the Header Auth account under Credentials (Generic Auth Type: Header Authentication). The Value field should be set with the Bearer XXXXXXXXXXXXXX. The XXXXXXXXXXXXXX should be replaced by the Web Unlocker Token. In n8n, configure the Google Gemini(PaLM) Api account with the Google Gemini API key (or access through Vertex AI or proxy). Update the LinkedIn URL by navigating to the Set LinkedIn URL node. Update the Webhook HTTP Request node with the Webhook endpoint of your choice. How to customize this workflow to your needs Input Variations: Modify the **Set LinkedIn URL node to accept a different company LinkedIn URL. Data Points**: Adjust the HTML Data Extractor Node to retrieve additional details like employee count, industry, or headquarters location. Summarization Style**: Customize the AI prompt to generate summaries in different tones or formats (e.g., formal, casual, bullet points). Output Destinations**: Configure the output node to send summaries to various platforms, such as Slack, CRM systems, or databases.
by Khairul Muhtadin
The Page Speed Insight workflow automates website performance analysis by integrating Google PageSpeed Insights API with Discord messaging and Gemini. This n8n workflow provides expert-level performance audits and comparisons, delivering actionable insights for website owners, SEO professionals, and developers. Disclaimer: this workflow using community nodes Google PageSpeed Insights Community Node 💡 Why Use Page Speed Insight? Save Time:** Instantly analyze and compare website speeds without manual tool usage Eliminate Guesswork:** Receive expert audit reports that translate technical data into clear, actionable insights Improve Website Outcomes:** Identify critical bottlenecks and enhancements prioritized by AI-driven analysis Seamless Integration:** Pull URLs and deliver reports directly via Discord for team collaboration and immediate response ⚡ Who Is This For? Webmasters and website owners seeking fast, automated performance checks SEO analysts who need consistent, data-backed website comparisons Developers requiring clear, prioritized action points from performance audits Digital agencies managing multiple client sites with ongoing monitoring needs 🔧 What This Workflow Does ⏱ Trigger:** Discord message containing URLs or scheduled execution 📎 Parse:** Extracts URLs and determines analysis type (single/comparison) 🔍 Analyze:** Calls Google PageSpeed API for performance data 🤖 Process:** AI generates user-friendly reports from raw Lighthouse JSON 💌 Deliver:** Sends chunked reports to Discord channels 🗂 Log:** Stores execution data for review and improvement 🔐 Setup Instructions Import the provided JSON workflow into your n8n instance Set up credentials for: Google PageSpeed API (ensure you have a valid API key — get yours here) Discord Bot API with permissions to read messages and send messages in your chosen guild/channel Customize the workflow by adjusting: Discord guild and channel IDs where messages are monitored and results posted Scheduled trigger interval if needed Any prompt text or AI model parameters to tailor report tone and detail level Test thoroughly with real URLs and Discord interaction to confirm smooth data flow and output quality 🧩 Pre-Requirements Active n8n instance (Cloud or self-hosted) n8n Google PageSpeed community node Google PageSpeed Insights API key Discord Bot credentials with channel access Google Gemini AI credentials (recommended) 🛠️ Customize It Further Extend to analyze desktop performance or other device types easily by modifying the PageSpeed API call Integrate with Slack, email, or other team tools alongside Discord for broader notification Enhance report depth by adding more AI-driven insights like competitor site recommendations or historical trend tracking 🧠 Nodes Used Google PageSpeed Insights Community Node Discord (getAllMessages, sendMessage) Code (URL parsing, message chunking) AI Language Model (Google Gemini) Schedule Trigger Switch (message type handling) Sticky Notes (workflow guidance) 📞 Support Made by: khaisa Studio Tag: automation, performance, SEO, google-pagespeed, discord Category: Monitoring & Reporting Need a custom solution? Contact Me
by Andrew
Who is this for? This workflow is ideal for n8n self-hosted users, DevOps engineers, and automation developers who want to automatically back up their n8n workflows to GitHub on a regular basis. What problem is this workflow solving Manually backing up n8n workflows can be time-consuming and prone to human error. This workflow automates the backup process, ensuring that all workflows are safely stored in a version-controlled GitHub repository every 24 hours. What this workflow does This automation runs daily to back up all workflows from your n8n instance to a specified GitHub repository. Each workflow is saved as a .json file using its unique ID, organized into a folder path defined by repo_path. The workflow is designed to manage memory usage efficiently by recursively calling itself. Once the backup is complete, it optionally sends a Slack notification to confirm success. Setup Configure the Config node in the subworkflow to set: GitHub Repo Owner GitHub Repo Name Main folder path (repo_path) Connect your GitHub and (optionally) Slack credentials. Set the workflow to run on a daily cron schedule. Test the workflow manually to confirm the GitHub integration works. Sign up for a free consultation and find out how n8n can help you.
by Budi SJ
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. 🎯 Purpose This workflow helps you automatically monitor stock related news, extract the main content, summarize it using a LLM (via OpenRouter), and send real time alerts to Telegram and store them in Google Sheets. ⚙️ How It Works Trigger A Cron node triggers the workflow every 15 minutes (adjustable). RSS Feed node checks latest articles from Google Alerts RSS. The workflow filters duplicates using Google Sheets as a log. The article URL is sent to Jina AI Readability API to extract the main body text. The content is summarized using a model from OpenRouter (e.g., Gemini, Claude, GPT-4). You can customize the prompt to suit your tone and analysis needs. The result is appended to a Google Sheets file. Sends the title, summary, and reccomendation to Telegram chat. 🧾 Google Sheets Template Create a Google Sheet using this template: Stock Alert 🧰 Requirements Telegram Bot + your Chat ID OpenRouter account and API key Jina AI account for content extraction Google Account with access to Google Sheets Google Alerts RSS feed 🛠 Setup Instructions Install required credentials: Add OpenRouter API key to n8n credentials. Add Telegram Bot Token and Chat ID. Add Google Sheets credentials. Add Jina AI credentials. Create or copy the Google Sheet using the link above. Go to Google Alerts, create alerts, and copy the RSS feed URL. Replace placeholder API keys and URLs. Adjust Telegram Chat ID. 🔐 Security Note All sensitive credentials (e.g., API keys, personal chat IDs) have been removed from this template. Please replace them using the n8n credentials manager before activating the workflow.
by Khairul Muhtadin
The blogblizt: polylang workflow streamlines the creation and publication of high-quality blog content using powerful automation with n8n, OpenAI’s GPT and the WordPress API. It enables effortlessly generate SEO-friendly articles complete with metadata and optimized featured images, improving content freshness and search engine visibility. 💡 Why Use blogblizt? Automate content creation** to keep your blog fresh and engaging Generate SEO-optimized posts** with expert-crafted titles, meta descriptions, and focus keyphrases Save hours** of manual writing, image sourcing, and SEO configuration Leverage AI** for topic ideation and high-quality writing tailored to international student audiences Seamlessly publish and manage drafts** directly on your WordPress site via API Produce captivating, relevant featured images** without external tools Support multilingual content creation** with randomized language selection for diversity ⚡ Who Is This For? Content strategists managing WordPress blogs needing efficient topic generation SEO specialists wanting automated post creation with optimized metadata Website owners aiming to maintain active, multilingual content Marketers who want to leverage AI for high-quality, consistent article production ❓ What Problem Does It Solve? This workflow automates the entire editorial cycle—from generating engaging topics with AI, drafting full-length articles, producing featured images automatically, to posting drafts configured for SEO on WordPress—dramatically reducing editor workload and improving content output. 🔧 What This Workflow Does ⏱ Trigger Runs on manual trigger or a weekly schedule to ensure consistent content flow 📎 Fetch Site Context Retrieves recent posts, taxonomies, and WordPress API schema to understand site structure 🔍 Generate Topic Uses OpenAI GPT-4.1-mini to roll a random language and craft a targeted blog post topic + SEO metadata 🤖 Draft Article Composes a comprehensive, SEO-friendly article tailored to the generated topic 💌 Create Draft Posts the draft on WordPress with Yoast SEO fields populated 🖼 Generate Image Creates a high-quality, cinematic featured image via AI 📤 Upload & Attach Uploads the image to the WordPress media library and sets it as the post’s featured image 🔐 Setup Instructions Import the workflow file into n8n: Add credentials: WordPress API (with create-post & media permissions) OpenAI API key (for GPT and image models) Customize categories, languages, and schedule in the relevant nodes Adjust the Schedule Trigger timing as desired (e.g. every Monday at 9 AM) Test end-to-end on a staging WordPress site to verify drafts and images publish correctly 🧩 Pre-Requirements An operational n8n instance (Cloud or self-hosted) (self-hosted or n8n cloud) WordPress site with REST API access & proper authentication OpenAI account with API access for both language and image models (Optional) Yoast SEO plugin installed for metadata recognition 🛠️ Customize It Further Tweak OpenAI prompts for niche topics or additional languages Add social-media nodes to auto-share new posts Insert an editorial review step before publishing Refine image prompts for different visual styles (e.g., “modern infographic” vs. “cinematic portrait”) 🧠 Nodes Used Manual Trigger** Schedule Trigger** (weekly) HTTP Request** (fetch posts, taxonomies, schema; upload media) Code** (JavaScript analyzers for API schema & taxonomy parsing) OpenAI Chat** (GPT-4.1-mini for topics & articles) OpenAI Image Generation** (for featured images) WordPress** (create draft post) Sticky Notes** (in-flow documentation) 📞 Support Built by: Khaisa Studio Tags: wordpress, marketing, polylang Category: Content Creation Need a custom? contact me on LinkedIn or Web
by Lucas Peyrin
This workflow contains community nodes that are only compatible with the self-hosted version of n8n. How it works This template is your personal launchpad into the world of AI-powered automation. It provides a fully functional, interactive AI chatbot that you can set up in minutes, designed specifically for those new to AI Agents. What is an AI Agent? Think of it as a smart assistant that doesn't just talk—it acts. You give it a set of "tools" (like other n8n tool nodes), and it intelligently decides which tool to use to answer your questions or complete your tasks. This starter kit comes with a pre-built "toolbox" of superpowers, allowing your agent to: Get the Weather:** Ask for the forecast anywhere in the world. Get the News:** Fetch the latest headlines from n8n, CNN, and others. The workflow is designed to be a hands-on learning experience, with detailed sticky notes explaining every component, from the chat interface to the agent's "brain" and "memory." Set up steps Setup time: ~2-3 minutes This workflow is designed to be incredibly easy to start. You only need one free API key to get it working. Add Your AI Key: The workflow uses Google's Gemini model by default. You will need a free Gemini API key. Find the Gemini node on the canvas. The sticky note right below it (How to Get Google Gemini Credentials) provides a link and simple instructions to get your key. In the Gemini node, click the Credential dropdown and select + Create New Credential to add your key. Activate the Workflow: At the top-right of the screen, click the "Inactive" toggle switch. It will turn green and say "Active". Your agent is now live! Start Chatting: Open the Example Chat Window node (it has a 💬 icon). In its parameter panel, you will see a Chat URL. Click the link to copy it. Paste the URL into a new browser tab and start asking your agent questions! Optional: The template also includes disabled OpenAI chat model node and tools for Google Calendar, and Gmail. You can enable and configure these later to change the underlying AI model or give your agent even more superpowers!
by RedOne
🎙️ AI Audio Assistant with Voice-to-Voice Response Who is this for? Businesses, customer service teams, content creators, and organizations who want to provide intelligent voice-based interactions through Telegram. Perfect for accessibility-focused services, multilingual support, or hands-free customer assistance. What problem does this solve? Enables natural voice conversations with AI Breaks down language and accessibility barriers Provides instant voice responses to customer queries Reduces typing requirements for users Offers 24/7 voice-based customer support Maintains conversation context across voice interactions What this workflow does: Receives voice messages via Telegram bot Transcribes audio using Deepgram's advanced speech-to-text Processes transcribed text through AI agent with knowledge base access Generates intelligent responses based on conversation context Converts AI response to natural-sounding speech using Deepgram TTS Sends audio response back to user via Telegram Maintains conversation memory for contextual interactions 🔧 Technical Architecture Core Components: Telegram Bot**: Receives and sends voice messages Deepgram STT**: Transcribes voice to text with high accuracy OpenAI GPT**: Processes queries and generates responses Supabase Knowledge Base**: Stores and retrieves business information Memory Management**: Maintains conversation context Deepgram TTS**: Converts text responses to natural speech Data Flow: Voice Message → Telegram API → File Download Audio File → Deepgram STT → Transcript Transcript → AI Agent → Response Generation Response → Deepgram TTS → Audio File Audio Response → Telegram → User 🛠️ Setup Instructions Prerequisites Telegram Bot Token Create bot via @BotFather Get bot token and configure webhook Deepgram API Key Sign up at deepgram.com Get API key for STT and TTS services Note: Currently hardcoded in workflow OpenAI API Key OpenAI account with API access Configure in OpenAI Chat Model node Supabase Database Create Supabase project Set up knowledge_base table Configure API credentials Step-by-Step Setup Configure Telegram Bot Update telegramToken in "Prepare Voice Message Data" node Set correct bot token in Telegram nodes Test bot connectivity Set Up Deepgram Integration Replace API key in "Transcribe with Deepgram" node Update TTS endpoint in "HTTP Request" node Test voice transcription accuracy Configure Knowledge Base -- Create knowledge_base table in Supabase CREATE TABLE knowledge_base ( id UUID DEFAULT gen_random_uuid() PRIMARY KEY, question TEXT NOT NULL, answer TEXT NOT NULL, category VARCHAR(100), keywords TEXT[], created_at TIMESTAMP WITH TIME ZONE DEFAULT NOW() ); Customize AI Prompts Update system message in "Telegram AI Agent" node Adjust temperature and max tokens in OpenAI model Configure memory session keys Test End-to-End Flow Send test voice message to bot Verify transcription accuracy Check AI response quality Validate audio output clarity 🎛️ Configuration Options Voice Recognition Settings Model**: nova-2 (Deepgram's latest model) Language**: English (en) - can be changed Smart Format**: Enabled for better punctuation AI Response Settings Temperature**: 0.3 (conservative responses) Max Tokens**: 100 (adjust based on needs) Memory**: Session-based conversation context Text-to-Speech Settings Model**: aura-2-thalia-en (natural female voice) Alternative voices**: Available in Deepgram TTS API Audio Format**: Optimized for Telegram 🔒 Security Considerations API Key Management // Current implementation has hardcoded tokens // Recommended: Use environment variables const telegramToken = process.env.TELEGRAM_BOT_TOKEN; const deepgramKey = process.env.DEEPGRAM_API_KEY; Data Privacy Voice messages are processed by external APIs Consider data retention policies Implement user consent mechanisms Ensure GDPR compliance if applicable 📊 Monitoring & Analytics Key Metrics to Track Voice message processing time Transcription accuracy rates AI response quality scores User engagement metrics Error rates and failure points Recommended Logging // Add to workflow for monitoring console.log({ timestamp: new Date().toISOString(), user_id: userData.user_id, transcript_confidence: transcriptData.confidence, response_length: aiResponse.length, processing_time: processingTime }); 🚀 Customization Ideas Enhanced Features Multi-language Support Add language detection Support multiple TTS voices Translate responses Voice Commands Implement wake words Add voice shortcuts Create voice menus Advanced AI Features Sentiment analysis Intent classification Escalation triggers Integration Expansions Connect to CRM systems Add calendar scheduling Integrate with help desk tools Performance Optimizations Implement audio preprocessing Add response caching Optimize API call sequences Implement retry mechanisms 🐛 Troubleshooting Common Issues Voice Not Transcribing Check Deepgram API key validity Verify audio format compatibility Test with shorter voice messages Poor Audio Quality Adjust TTS model settings Check network connectivity Verify Telegram audio limits AI Responses Too Generic Improve knowledge base content Adjust system prompts Increase context window Memory Not Working Check session key configuration Verify user ID extraction Test conversation continuity 💡 Best Practices Voice Interface Design Keep responses concise and clear Use natural speech patterns Avoid technical jargon Provide clear next steps Knowledge Base Management Regular content updates Clear categorization Keyword optimization Quality assurance testing User Experience Fast response times (<5 seconds) Consistent voice personality Graceful error handling Clear capability communication 📈 Success Metrics Technical KPIs Response time: <3 seconds average Transcription accuracy: >95% User satisfaction: >4.5/5 Uptime: >99.5% Business KPIs Customer query resolution rate Support ticket reduction User engagement increase Cost per interaction decrease 🔄 Maintenance Schedule Daily Monitor error logs Check API rate limits Verify service uptime Weekly Review conversation quality Update knowledge base Analyze usage patterns Monthly Performance optimization Security audit Feature updates User feedback review 📚 Additional Resources Documentation Links Deepgram STT API Deepgram TTS API Telegram Bot API OpenAI API Supabase Documentation Community Support n8n Community Forum Telegram Bot Developers Group Deepgram Developer Discord OpenAI Developer Community Note: This template requires active API subscriptions for Deepgram and OpenAI services. Costs may apply based on usage volume.
by Derek Cheung
How it works: Using a Crew of AI agents (Senior Researcher, Visionary, and Senior Editor), this crew will automatically determine the right questions to ask to produce a detailed fundamental stock analysis. This application has two components: a front-end and a Stock Q&A engine. The front end is the team of agents automatically figuring out the questions to ask, and the back-end part is the ability to answer those questions with the SEC 10K data. This template implements the Stock Q&A engine. For the front-end of the application, you can choose one of two options: using CrewAI with the Replit environment (code approach) fully visual approach with n8n template (AI-powered automated stock analysis) Setup steps: Use first workflow in template to upsert a company annual report PDF (such as from SEC 10K filling) Get URL for Webhook in second workflow template CrewAI front-end: Youtube overview video Fork this AI Agent environment Crew Agent Environment Set the webhook URL into N8N_WEBHOOK_URL variable Set OpenAI_API_KEY variable
by Miha
Combine Tech News in a Personalized Weekly Newsletter This n8n template automates the collection, storage, and summarization of technology news from top sites, turning it into a concise, personalized weekly newsletter. If you like staying informed but want to reduce daily distractions, this workflow is perfect for you. It leverages RSS feeds, vector databases, and LLMs to read and curate tech content on your behalf—so you only receive what truly matters. How it works A daily scheduled trigger fetches articles from multiple popular tech RSS feeds like Wired, TechCrunch, and The Verge. Fetched articles are: Normalized to extract titles, summaries, and publish dates. Converted to vector embeddings via OpenAI and stored in memory for fast semantic querying. A weekly scheduled trigger activates the AI summarization flow: The AI is provided with your interests (e.g., AI, games, gadgets) and the desired number of items (e.g., 15). It queries the vector store to retrieve relevant articles and summarizes the most newsworthy stories. The summary is converted into a clean, email-friendly format and sent to your inbox. How to use Connect your OpenAI and Gmail accounts to n8n. Customize the list of RSS feeds in the “Set Tech News RSS Feeds” node. Update your interests and number of desired news items in the “Your Topics of Interest” node. Activate the workflow and let the automation run on schedule. Requirements OpenAI** credentials for embeddings and summarization Gmail** (or another email service) for sending the newsletter Customizing this workflow Want to use different sources? Swap in your own RSS feeds, or use an API-based news aggregator. Replace the in-memory vector store with Pinecone, Weaviate, or another persistent vector DB for longer-term storage. Adjust the agent's summarization style to suit internal updates, industry-specific briefings, or even entertainment recaps. Prefer chat over email? Replace the email node with a Telegram bot to receive your personalized tech newsletter directly in a Telegram chat.
by Jimleuk
This n8n template monitors active support issues in Linear.app to track the mood of their ongoing conversation between reporter and assignee using Sentiment Analysis. When sentiment dips into the negative, a notification is sent via Slack to alert the team. How it works A scheduled trigger is used to fetch recently updated issues in Linear using the GraphQL node. Each issue's comments thread is passed into a simple Information Extractor node to identify the overall sentiment. The resulting sentiment analysis combined with the some issue details are uploaded to Airtable for review. When the template is re-run at a later date, each issue is re-analysed for sentiment Each issue's new sentiment state is saved to the airtable whilst its previous state is moved to the "previous sentiment" column. An Airtable trigger is used to watch for recently updated rows Each matching Airtable row is filtered to check if it has a previous non-negative state but now has a negative state in its current sentiment. The results are sent via notification to a team slack channel for priority. Check out the sample Airtable here: https://airtable.com/appViDaeaFw4qv9La/shrq6HgeYzpW6uwXL How to use Modify the GraphQL filter to fetch issues to a relevant issue type, team or person. Update the Slack channel to ensure messages are sent to the correct location or persons. The Airtable also serves to give a snapshot of Sentiment across support tickets for a given period. It's possible to use this to assess the daily operations. Requirements Linear for issue tracking (but feel free to use another system if preferred) Airtable for Database OpenAI for LLM and Sentiment Analysis Customising the workflow Add more granular levels of sentiment to reduce the number of alerts. Explore different types of sentiment based on issue types and customer types. This may help prioritise alerts and response. Run across teams or categories of issues to get an overview of sentiment across the support organisation.
by Davide
This automated workflow takes a static image and a textual prompt and transforms them into an animated video using the MiniMax Hailuo 02 model. It then uploads the generated video to YouTube and TikTok, and updates a Google Sheet with relevant links and metadata. Benefits of This Workflow Fully Automated Pipeline**: From prompt to video to social media publication — all without manual intervention. Scalable Content Creation**: Generate and distribute dozens of videos per hour with minimal human input. Cross-Platform Posting: Automatically pushes content to **YouTube and TikTok simultaneously. SEO Optimization**: Uses AI to generate catchy, keyword-rich video titles that improve visibility. Easy Integration**: Based on Google Sheets for input/output, making it accessible to non-technical users. Time-Efficient**: Batch-processing enabled with scheduled runs every few minutes. Customizable Duration**: Video duration can be adjusted (default is 6 seconds). How It Works Trigger & Data Fetching: The workflow starts either manually or via a scheduled trigger (e.g., every 5 minutes). It checks a Google Sheet for new entries where the "VIDEO" column is empty, indicating pending video generation tasks. Video Creation: For each entry, the workflow extracts the image URL and prompt from the Google Sheet. It sends these inputs to the MiniMax Hailuo 02 to generate a video. The API processes the image and prompt, optimizes the prompt, and creates a short video (default: 6 seconds). Status Monitoring: The workflow polls the API every 60 seconds to check if the video is COMPLETED. Once ready, it retrieves the video URL and uploads the file to Google Drive. YouTube & TikTok Upload: The video is sent to YouTube and TikTok via the Upload-Post.com API (The free plan allows uploads to all platforms except TikTok. To enable, upgrade to a paid plan.). A GPT-generated SEO-optimized title is created for the video. The Google Sheet is updated with the video URL and YouTube link. Set Up Steps Google Sheet Setup: Create a Google Sheet with columns: IMAGE (input image URL), PROMPT (video description), VIDEO (auto-filled), and YOUTUBE_URL (auto-filled). Link the sheet to the workflow using the Google Sheets node. API Keys: Obtain a fal.run API key (for MiniMax Hailuo) and configure the "Authorization" header in the "Create video" node. Get an Upload-Post.com API key (10 free uploads/month) and set it in the "Upload on YouTube/TikTok" nodes. Workflow Configuration: Replace YOUR_USERNAME in the Upload-Post nodes with your profile name (e.g., "test1"). Adjust the video duration (6 or 10 seconds) in the "Create video" node. Set the Schedule Trigger interval (e.g., 5 minutes) to automate checks for new tasks. Execution: Run the workflow manually or let the scheduler process new rows automatically. The system handles video generation, uploads, and Google Sheet updates end-to-end. Need help customizing? Contact me for consulting and support or add me on Linkedin.
by Mark Shcherbakov
Video Guide I prepared a detailed guide explaining how to build an AI-powered meeting assistant that provides real-time transcription and insights during virtual meetings. Youtube Link Who is this for? This workflow is ideal for business professionals, project managers, and team leaders who require effective transcription of meetings for improved documentation and note-taking. It's particularly beneficial for those who conduct frequent virtual meetings across various platforms like Zoom and Google Meet. What problem does this workflow solve? Transcribing meetings manually can be tedious and prone to error. This workflow automates the transcription process in real-time, ensuring that key discussions and decisions are accurately captured and easily accessible for later review, thus enhancing productivity and clarity in communications. What this workflow does The workflow employs an AI-powered assistant to join virtual meetings and capture discussions through real-time transcription. Key functionalities include: Automatic joining of meetings on platforms like Zoom, Google Meet, and others with the ability to provide real-time transcription. Integration with transcription APIs (e.g., AssemblyAI) to deliver seamless and accurate capture of dialogue. Structuring and storing transcriptions efficiently in a database for easy retrieval and analysis. Real-Time Transcription: The assistant captures audio during meetings and transcribes it in real-time, allowing participants to focus on discussions. Keyword Recognition: Key phrases can trigger specific actions, such as noting important points or making prompts to the assistant. Structured Data Management: The assistant maintains a database of transcriptions linked to meeting details for organized storage and quick access later. Setup Preparation Create Recall.ai API key Setup Supabase account and table create table public.data ( id uuid not null default gen_random_uuid (), date_created timestamp with time zone not null default (now() at time zone 'utc'::text), input jsonb null, output jsonb null, constraint data_pkey primary key (id), ) tablespace pg_default; Create OpenAI API key Development Bot Creation: Use a node to create the bot that will join meetings. Provide the meeting URL and set transcription options within the API request. Authentication: Configure authentication settings via a Bearer token for interacting with your transcription service. Webhook Setup: Create a webhook to receive real-time transcription updates, ensuring timely data capture during meetings. Join Meeting: Set the bot to join the specified meeting and actively listen to capture conversations. Transcription Handling: Combine transcription fragments into cohesive sentences and manage dialog arrays for coherence. Trigger Actions on Keywords: Set up keyword recognition that can initiate requests to the OpenAI API for additional interactions based on captured dialogue. Output and Summary Generation: Produce insights and summary notes from the transcriptions that can be stored back into the database for future reference.