Operations

Latency

Latency is the time between request and response.

Quick definition

Latency is the time between request and response.

  • Category: Operations
  • Focus: performance and reliability
  • Used in: Reducing time-to-first-token with streaming.

What it means

Lower latency improves user experience in chat interfaces. In operations workflows, latency often shapes performance and reliability.

How it works

Operations covers latency, throughput, and cost. Systems often use caching, batching, and monitoring to scale reliably.

Why it matters

Operational choices impact cost, latency, and reliability.

Common use cases

  • Reducing time-to-first-token with streaming.
  • Managing costs with token budgets and caching.
  • Tracking usage and errors with logs and metrics.

Example

Aim for under 1 second time-to-first-token.

Pitfalls and tips

Ignoring limits can cause timeouts or rate limiting. Set budgets and monitor usage to avoid surprises.

In BoltAI

In BoltAI, this shows up in performance, logging, or usage views.