BurnLens is a local proxy that tracks every AI API call — per feature, team, and customer. One install. Zero config. Nothing leaves your machine.
No SDK changes. No code rewrites. Just point your existing API calls through the proxy.
pip install burnlens then burnlens start.
Proxy runs on port 8420. Takes 30 seconds.
Set OPENAI_BASE_URL to the proxy. Add
X-BurnLens-Tag-Feature headers to your requests. Done.
Open the dashboard. Cost by model, feature, team, customer. Waste alerts. Per-call token counts and latency.
Built for engineers and teams who are serious about understanding their AI costs.
Tag requests with X-BurnLens-Tag-Feature and see exactly which part of your product is spending what.
Set monthly spend limits per team and per customer. Get warnings at 80%. Automatic 429 when a customer exceeds their cap.
Detect duplicate requests, context bloat, model overkill, and uncached system prompts. Dollar estimates included.
BurnLens analyses your usage patterns and recommends cheaper models where the data supports it. Projected savings shown.
burnlens export to CSV. burnlens report for weekly summaries. Email delivery supported.
Everything stored in a local SQLite file. No accounts. No SaaS. No prompt content ever leaves your machine.
Supports all major providers out of the box. More coming.
BurnLens is MIT licensed and always will be. No accounts, no usage limits, no data collection. Run it locally and own everything.