LumiqTrace exposes four metrics endpoints that power the analytics views in the dashboard. You can call these endpoints directly to build custom reports, integrate metrics into your own internal tooling, or export data to other systems. All four endpoints require a Bearer token and aDocumentation Index
Fetch the complete documentation index at: https://docs.lumiqtrace.com/llms.txt
Use this file to discover all available pages before exploring further.
projectId parameter.
GET /v1/projects/:id/overview
Returns the top-level KPIs for a project along with a timeline, top models by cost, and recent errors. This is the same data shown on the project overview page.Authorization: Bearer <token>
Path Parameters
The project ID.
Example Request
Response
Total LLM calls in the selected period.
Total spend in USD.
Average latency in milliseconds across all calls.
Fraction of calls that resulted in an error, timeout, or rate-limited status (0–1).
Hourly or daily bucketed metrics series.
Top 5 models by total cost descending.
The 10 most recent non-success events.
GET /v1/costs
Returns a detailed cost breakdown by model, a daily spending trend, month-to-date total, a 30-day linear forecast, and cache utilization rate.Authorization: Bearer <token>
Query Parameters
The project ID.
ISO 8601 UTC start of the reporting window. Defaults to 30 days ago.
ISO 8601 UTC end of the reporting window. Defaults to now.
Example Request
Response
Per-model cost breakdown sorted by total cost descending.
Day-by-day cost series for the requested window.
Month-to-date total spend in USD.
Projected 30-day spend in USD based on linear extrapolation of recent daily spending.
Fraction of input tokens served from the provider’s prompt cache (0–1).
GET /v1/performance
Returns latency percentiles, average time-to-first-token, tokens-per-second throughput, and timeout rate.Authorization: Bearer <token>
Query Parameters
The project ID.
ISO 8601 UTC start of the reporting window.
ISO 8601 UTC end of the reporting window.
Example Request
Response
Median end-to-end latency in milliseconds.
90th percentile latency in milliseconds.
99th percentile latency in milliseconds.
Average time-to-first-token in milliseconds across streaming calls.
Average output token generation rate across all calls.
Fraction of calls that timed out (0–1).
GET /v1/errors
Returns error events grouped by error code with counts and trend data, plus a rate limit heatmap for identifying bursty traffic patterns.Authorization: Bearer <token>
Query Parameters
The project ID.
ISO 8601 UTC start of the reporting window.
ISO 8601 UTC end of the reporting window.
Example Request
Response
Error events grouped by
error_code, sorted by count descending. Each group includes a trend field ("increasing", "stable", or "decreasing") based on recent trajectory.Hourly count of rate-limited calls. Use this to identify time windows where your application is hitting provider rate limits.