Capture every LLM API call during development. Request/response bodies, token usage, cost estimation, and latency in a local web UI.
Features
Request & response bodies with visual flow visualization.
Input, output, cache read & write tokens for every call.
Real-time cost calculated via models.dev pricing data.
Captures both streaming and non-streaming calls automatically.
WebSocket-powered live updates as calls happen.
SQLite DB per project directory. Zero config needed.
Integration
pnpm add managepromptimport { generateText, wrapLanguageModel } from "ai";
import { openai } from "@ai-sdk/openai";
import { devToolsMiddleware } from "manageprompt";
const model = wrapLanguageModel({
model: openai("gpt-4o"),
middleware: devToolsMiddleware(),
});import OpenAI from "openai";
import { capture } from "manageprompt";
const openai = new OpenAI();
const response = await capture(
{ model: "gpt-4o-mini", messages },
(input) => openai.chat.completions.create(input),
);Works with any LLM API — OpenAI, Anthropic, Google, Mistral, and more.
Install
brew install techulus/tap/managepromptgo install github.com/techulus/manage-prompt/cmd/manageprompt@latestpnpm add manageprompt