Compare LLM Token Costs Across 350+ Models with OpenRouter

This n8n template lets you run prompts against 350+ LLM models and see exactly what each request costs with real-time pricing from OpenRouter

Use cases are many: Compare costs across different models, plan your AI budget, optimize prompts for cost efficiency, or track expenses for client billing!

Good to know

OpenRouter charges a platform fee on top of model costs. See OpenRouter Pricing for details. You need an OpenRouter account with API credits. Free signup available with some free models included. Pricing data is fetched live from OpenRouter's API, so costs are always up-to-date.

How it works

All available models are fetched from OpenRouter's API when you start. You select a model and enter your prompt via the form (or just use the chat). The prompt is sent to OpenRouter and the response is captured. Token usage (input/output) is extracted from the response using a LangChain Code node. Real-time pricing for your selected model is fetched from OpenRouter. The exact cost is calculated and displayed alongside your AI response.

How to use

Chat interface: Quick testing - just type a prompt and get the response with costs. Form interface: Select from all available models via dropdown, enter your prompt, and get a detailed cost breakdown. Click "Show Details" on the result form to see the full breakdown (input tokens, output tokens, cost per type).

Requirements

OpenRouter account with API key (Get one here)

Customising this workflow

Add a database node to log all requests and costs over time Connect to Google Sheets for cost tracking and reporting Extend with LLM-as-Judge evaluation to also check response quality

0
Downloads
0
Views
8.1
Quality Score
beginner
Complexity
Author:Philflow(View Original →)
Created:12/26/2025
Updated:1/13/2026

🔒 Please log in to import templates to n8n and favorite templates

Workflow Visualization

Loading...

Preparing workflow renderer

Comments (0)

Login to post comments