See everything. Miss nothing.
Full visibility into every LLM request. Logs, cost tracking, latency monitoring, and token analytics across 9 providers.
Complete LLM observability
Every LLM request logged with full context: input, output, tokens, and timing.
Per-request cost calculation across all providers. See exactly where money goes.
Track response times by provider, model, and endpoint. Catch regressions early.
Input vs output token usage over time. Understand your consumption patterns.
See exactly how much Fold saves you per request, per session, and per month.
Attach quality scores to requests. Correlate prompt changes with response quality.
One flag to enable everything
Add platform options to your existing Fold setup. Logging starts automatically.
import { fold } from "@fold-run/sdk";
const ctx = fold("coding", {
platform: {
apiKey: process.env.FOLD_API_KEY,
projectId: "my-project",
// Auto-logs every request with cost, latency, and tokens
logging: true,
analytics: true,
},
});
// Your normal agent loop — Fold logs everything automatically
ctx.system("You are a coding assistant");
ctx.observe(searchResults, "web-search");
const messages = ctx.messages();
// Request logged: provider, model, tokens, cost, latency, fold savingsYour LLM operations at a glance
Built for every team
Identify expensive prompts and optimize them. Track spending by project, team, and model.
Replay any request with full context. See exactly what your LLM received and returned.
Complete audit trail of every LLM interaction. Filter and export logs by date, user, or project.
Set alerts for latency spikes, cost overruns, or error rate increases. Stay ahead of issues.