Operations
Latency
Latency is the time between request and response.
Quick definition
Latency is the time between request and response.
- Category: Operations
- Focus: performance and reliability
- Used in: Reducing time-to-first-token with streaming.
What it means
Lower latency improves user experience in chat interfaces. In operations workflows, latency often shapes performance and reliability.
How it works
Operations covers latency, throughput, and cost. Systems often use caching, batching, and monitoring to scale reliably.
Why it matters
Operational choices impact cost, latency, and reliability.
Common use cases
Example
Aim for under 1 second time-to-first-token.
Pitfalls and tips
Ignoring limits can cause timeouts or rate limiting. Set budgets and monitor usage to avoid surprises.
In BoltAI
In BoltAI, this shows up in performance, logging, or usage views.