Capture every LLM API call during development. Full request/response details, token usage, cost estimation, and latency — all in a local web UI.
Features
Request & response bodies, headers with API keys automatically masked.
Input, output, cache read & write tokens for every call.
Real-time cost calculated via models.dev pricing data.
SSE streaming forwarded in real-time with zero buffering delay.
WebSocket-powered live updates as calls happen.
SQLite DB per project directory. Zero config needed.
Integration
Works with any LLM API — OpenAI, Anthropic, Google, Mistral, and more.
Install