by Automate With Marc
This n8n workflow template uses community nodes and is only compatible with the self-hosted version of n8n. 📈 StockPulse: AI-Picked Daily News for Your Portfolio Stay ahead of the market with this automated, AI-powered stock market news briefing delivered straight to your inbox — no code required. Watch Step-by-step Video Tutorial Here: https://www.youtube.com/watch?v=iZvPej9eLYE&t=201s ⚙️ What it does: This workflow runs every morning and: Triggers a scheduled prompt to a Langchain AI Agent (OpenAI) Uses the Tavily Web Search API to fetch fresh financial news relevant to your watchlist or portfolio Summarizes the top stories, highlighting: 🔍 Key headlines 💡 Investment opportunities ⚠️ Risks and macro trends 📊 Suggested trades Sends a clean, readable email via Gmail to your preferred address 🔧 Built with: 🧠 Langchain AI Agent (OpenAI GPT-4o) 🔍 Tavily Search Tool 📬 Gmail Node for Email Delivery ⏰ Daily Cron Trigger (customizable) 💼 Who it’s for: Investors and traders who want to save time on news gathering Financial creators looking for curated, actionable insights Non-technical users interested in automating stock news monitoring Anyone who wants to combine AI + automation + market data 🟢 Customize easily: Edit your stock list or news focus inside the Agent prompt 📨 Email ready: Just plug in your Gmail credentials and you’re good to go ⏱️ 10-minute setup — no coding required!
by Jonathan
This workflow will take an alert from Syncro, determine if it's an agent_offline_trigger type, then determine if it's a new alert or a close to an existing alert, and then submit it to OpsGenie. New alerts will create a new alert in OpsGenie and resolved alerts will close the alert in OpsGenie. It doesn't require any kind of Google Sheets because OpsGenie allows you to submit a unique ID (known as an alias) along with the alert, which can be referenced later when closing the alert. The trigger type can be changed to suit your needs. You will need to create an API integration in OpsGenie. In Syncro, in addition to setting up the appropriate notification to webhook, you will also need a script that closes the agent_offline_trigger alert and an automated remediation to trigger that script when the asset goes offline (the script is queued and run when the asset comes back online). > This workflow is part of an MSP collection, The original can be found here: https://github.com/bionemesis/n8nsyncro
by Agent Circle
This n8n template helps you automatically discover, analyze, and track trending topics and videos on YouTube using an AI-powered agent. Use cases are many: This workflow is perfect for YouTube creators needing fresh video ideas, digital marketers scouting new campaign topics, social media managers who want to catch trends early, and researchers who want to analyze what’s viral. How It Works The workflow starts whenever a chat message is received (e.g., a trend question, a topic prompt, or a request for insights). Incoming chat is routed to the AI Agent – Trend Explorer node: First, the agent triggers the Workflow – YouTube Search tool to gather the latest trending topics and keywords from YouTube. Next, the agent supplies this real-time YouTube data to the OpenAI Chat Model for deep analysis, trend interpretation, and unique insights. To provide context-aware answers and track ongoing interests, the agent also references a Simple Memory module, recalling past queries, and user instructions. Finally, the result is a fast, data-driven, and smart trend report delivered instantly to your chat. How To Set Up Download the workflow package (including 2 .json files) and import it into your n8n interface. Set up necessary access in the following components of the AI Agent - Trend Explorer node: OpenAI Chat Model: allows API connection for trend insights. Workflow – YouTube Search: searches for trending videos based on the query. Simple Memory (optional): enhances experience for ongoing sessions. Start by sending a chat message on n8n. Check the response from the AI agent in the same chat box. Ask follow-ups, explore deeper, or trigger new searches - all in one chat thread. Requirements n8n instance (self-hosted or cloud). Set up API access to OpenAI Chat Model for chat-based AI. How To Customize Connect to your favorite chat platforms**: Easily integrate with additional chat triggers such as Telegram, Slack, or your preferred messaging app. Choose your preferred AI model**: If you want a different viewpoint, simply swap OpenAI Chat Model for Google Gemini, Claude, or any compatible LLM model in your workflow. Upgrade memory for smarter conversations: For long-term recall or more advanced, context-aware chats, replace **Simple Memory with a vector database like Pinecone or Redis. Need Help? If you’d like this workflow customized to fit your tools and platforms availability, or if you’re looking to build a tailored AI Agent for your own business - please feel free to reach out to Agent Circle. We’re always here to support and help you to bring automation ideas to life. Join our community on different platforms for support, inspiration and tips from others. Website: https://www.agentcircle.ai/ Etsy: https://www.etsy.com/shop/AgentCircle Gumroad: http://agentcircle.gumroad.com/ Discord Global: https://discord.gg/d8SkCzKwnP FB Page Global: https://www.facebook.com/agentcircle/ FB Group Global: https://www.facebook.com/groups/aiagentcircle/ X: https://x.com/agent_circle YouTube: https://www.youtube.com/@agentcircle LinkedIn: https://www.linkedin.com/company/agentcircle
by Jonathan
This workflow checks for new emails in a mailbox and if the email body contains the word "invoice" it will send the attachment to Mindee. It then posts a message to Slack to let the team know a payment needs to be made, If the value of the invoice is over 1000 it will also email the finance manager. To use this workflow you will need to configure the IMAP node to select the correct mailbox to use then configure the Mindee node to use your credentials. Once that is done the Send Email node will need to be configured to use the correct mail server and to send to the correct people, The last thing to configure is the Slack node this will need your Slack credentials and the channel you want to post the message to.
by jason
Not sure what to eat tonight? Have recipes emailed to you daily based on your criterial. To run this workflow, you will need to have: A Recipe Search API key from Edamam An active email account with configured credentials To set up your credentials: Set your Edamam AppID and AppKey in the Search Criteria node Select (or create) your email credentials in the Send Recipes node (and set up the to: and from: email addresses while you are at it) To customize the recipes that you receive, open up the Search Criteria node and modify one or more of the following: RecipeCount** - the numner of recipes you would like to receive IngredientCount** - the maximum number of ingredients you would like each recipe to have CaloriesMin** - the minimum number of calories the recipe will have CaloriesMax** - the maximum number of calories the recipe will have TimeMin** - the minimum amount of time (in minutes) the recipe will take to prepare TimeMax** - the maximum amount of time (in minutes) the recipe will take to prepare Diet** - Select one of the following options: balanced - Protein/Fat/Carb values in 15/35/50 ratio high-fiber - More than 5g fiber per serving high-protein - More than 50% of total calories from proteins low-carb - Less than 20% of total calories from carbs low-fat - Less than 15% of total calories from fat low-sodium - Less than 140mg Na per serving random - selects a different random diet each day Health** - Select one of the following options: alcohol-free - No alcohol used or contained immuno-supportive - Recipes which fit a science-based approach to eating to strengthen the immune system celery-free - does not contain celery or derivatives crustacean-free - does not contain crustaceans (shrimp, lobster etc.) or derivatives dairy-free - No dairy; no lactose egg-free - No eggs or products containing eggs fish-free - No fish or fish derivatives fodmap-free - Does not contain FODMAP foods gluten-free - No ingredients containing gluten keto-friendly - Maximum 7 grams of net carbs per serving kidney-friendly - per serving – phosphorus less than 250 mg AND potassium less than 500 mg AND sodium: less than 500 mg kosher - contains only ingredients allowed by the kosher diet. However it does not guarantee kosher preparation of the ingredients themselves low-potassium - Less than 150mg per serving lupine-free - does not contain lupine or derivatives mustard-free - does not contain mustard or derivatives low-fat-abs - Less than 3g of fat per serving no-oil-added - No oil added except to what is contained in the basic ingredients low-sugar - No simple sugars – glucose, dextrose, galactose, fructose, sucrose, lactose, maltose paleo - Excludes what are perceived to be agricultural products; grains, legumes, dairy products, potatoes, refined salt, refined sugar, and processed oils peanut-free - No peanuts or products containing peanuts pecatarian - Does not contain meat or meat based products, can contain dairy and fish pork-free - does not contain pork or derivatives red-meat-free - does not contain beef, lamb, pork, duck, goose, game, horse, and other types of red meat or products containing red meat. sesame-free - does not contain sesame seed or derivatives shellfish-free - No shellfish or shellfish derivatives soy-free - No soy or products containing soy sugar-conscious - Less than 4g of sugar per serving tree-nut-free - No tree nuts or products containing tree nuts vegan - No meat, poultry, fish, dairy, eggs or honey vegetarian - No meat, poultry, or fish wheat-free - No wheat, can have gluten though random - selects a different random health option each day SearchItem* - the general term that you are looking for e.g. *chicken
by Yaron Been
Prunaai Flux Schnell Image Generator Description This is a 3x faster FLUX.1 [schnell] model from Black Forest Labs, optimised with pruna with minimal quality loss. Contact us for more at pruna.ai Overview This n8n workflow integrates with the Replicate API to use the prunaai/flux-schnell model. This powerful AI model can generate high-quality image content based on your inputs. Features Easy integration with Replicate API Automated status checking and result retrieval Support for all model parameters Error handling and retry logic Clean output formatting Parameters Required Parameters prompt** (string): Prompt for generated image Optional Parameters seed** (integer, default: None): Random seed. Set for reproducible generation megapixels** (string, default: 1): Approximate number of megapixels for generated image speed_mode** (string, default: Juiced 🔥 (default)): Run faster predictions with model optimized for speed num_outputs** (integer, default: 1): Number of outputs to generate aspect_ratio** (string, default: 1:1): Aspect ratio of the output image output_format** (string, default: jpg): Format of the output images output_quality** (integer, default: 80): Quality when saving the output images, from 0 to 100. 100 is best quality, 0 is lowest quality. Not relevant for .png outputs num_inference_steps** (integer, default: 4): Number of denoising steps. 4 is recommended, and lower number of steps produce lower quality outputs, faster. How to Use Set up your Replicate API key in the workflow Configure the required parameters for your use case Run the workflow to generate image content Access the generated output from the final node API Reference Model: prunaai/flux-schnell API Endpoint: https://api.replicate.com/v1/predictions Requirements Replicate API key n8n instance Basic understanding of image generation parameters
by Harshil Agrawal
This workflow automatically monitors the functionality of a factory. The workflow logs machine data coming from factory sensors in a CrateDB database, generates an incident report in PagerDuty, and notifies the responsible staff members when the temperature of a machine crosses the threshold value. This workflow builds on a workflow that generates factory data. Read more about this use case and how to build both workflows with step-by-step instructions in the blog post How to automate your factory's incident reporting. Prerequisites A PagerDuty account and credentials AMQP, an ActiveMQ connection, and credentials A CrateDB instance running locally or on a server, and credentials. Nodes AMQP Trigger node starts the workflow. IF node filters sensor values higher than 50°C. PagerDuty node creates an incident in the account. Set nodes set the required incident information and sensor data, respectively. CrateDB nodes ingest the information data and machine sensor data, respectively. Function node converts degrees from Celsius to Fahrenheit.
by Fahmi Oktafian
This workflow is designed for content creators or AI artists who want to automate posting unique AI-generated images to their Facebook Page multiple times a day. It uses Google Gemini via LangChain to generate imaginative image prompts, and Pollinations AI to generate the images. Posts are published with hashtags and a clean caption. Who Is It For AI artists Facebook page managers Digital marketers looking for automated creative content What It Does Triggers 3x daily at 7:00, 11:00, and 17:00 (local time) Generates random AI image prompts in a retro-futuristic, cinematic, or surreal style using Google Gemini Fetches images from Pollinations AI using custom prompts Posts images automatically to your Facebook Page with hashtags Requirements n8n self-hosted or desktop (workflow uses schedule trigger) Pollinations API (no auth needed) Facebook Page with Facebook Graph API token: Required scopes: pages_manage_posts, pages_read_engagement, pages_show_list Google Gemini API Key (used via LangChain node) How to Customize Change the prompt style in the Basic LLM Chain node (promptType: define) to suit your theme. Adjust Set The Generator Image node if you want: Different image sizes (width / height) Different seed randomness Other Pollinations models (&model=kontext) Add Telegram/Twitter nodes if you want cross-posting Use Set node to allow easy user-level configuration of models, hashtags, times, etc.
by Oriol Seguí
This fun workflow automates the generation and delivery of personalized jokes by email based on the names or objects entered in the form. The process works as follows: On form submission The workflow starts when someone submits a form with the required names or objects to create the joke. You can modify the form fields to make the jokes more creative. Set the output language Manually define the language in which you want to receive the joke. OpenAI Message Model Uses the OpenAI model to generate the joke based on the prompt and in the chosen language. (The response is limited to around 200 tokens.) Gmail: send message The generated joke is automatically sent to the specified email address via Gmail.
by Anthony
This n8n workflow automates the process of researching companies by gathering relevant data such as traffic volume, foundation details, funding information, founders, and more. The workflow leverages the ProspectLens API, which is particularly useful for researching companies commonly found on Crunchbase and LinkedIn. ProspectLens is an API that provides very detailed company data. All you need to do is supply the company's domain name. You can obtain your ProspectLens API key here: https://apiroad.net/marketplace/apis/prospectlens In n8n, create a new "HTTP Header" credential. Set x-apiroad-key as the "Name" and enter your APIRoad API key as the "Value". Use this credential in the HTTP Request node of the workflow.
by Peter
Store a key with a value in a local json file. Multiple keys could be saved in a single file. Related workflow: GetKey Create a subfolder in your n8n homedir: /home/node/.n8n/local-files. In docker look at the data path and create a subfolder local-files. Set the correct access rights chmod 1000.1000 local-files. Put the workflow code in a new workflow named WriteKey. Create another workflow with a function item: return { file: '/4711.json', // 4711 should be your workflow id key: 'MyKey', value: 'MyValue' } Pipe the function item to an Execution Workflow that calls the WriteKey workflow. It would be nice if we could get someday a shiny built-in n8n node that does the job. :)
by Zacharia Kimotho
How to scrap emails from websites This workflow shows how to quickly build an Email scraping API using n8n. Email marketing is at the core of most marketing strategies, be it content marketing, sales, etc. As such, being able to find contacts in bulk for your business on a large scale is key. There are available tools available in the market that can do this, but most are premium; why not build a custom one with n8n? Usage The workflow gets the data from a website and performs an extraction based on the date around on the website Copy the webhook URL to your browser Add a query parameter eg ?Website=https://mailsafi.com . This should give you a URL like this {{$n8nhostingurl/webhook/ea568868-5770-4b2a-8893-700b344c995e?Website=https://mailsafi.com Click on the URL and wait for the extracted email to be displayed. This will return the email address on the website, or if there is no email, the response will be "workflow successfully executed." Make sure to use HTTP:// for your domains Otherwise, you may get an error.