Models
Context window
The context window is the maximum amount of text a model can consider at once.
Quick definition
The context window is the maximum amount of text a model can consider at once.
- Category: Models
- Focus: model capability and fit
- Used in: Choosing a model that fits latency and cost constraints.
What it means
It limits how much conversation history and documents can be included. In models workflows, context window often shapes model capability and fit.
How it works
Model architecture and scale determine capability. Context length, parameter count, and modality support vary across models.
Why it matters
Model architecture affects capability, context length, and speed.
Common use cases
- Choosing a model that fits latency and cost constraints.
- Selecting longer context for document-heavy workflows.
- Using specialized models for code, vision, or speech.
Example
A 128k context window can hold long docs in one request.
Pitfalls and tips
Bigger is not always better. Match the model to the task and evaluate in production.
In BoltAI
In BoltAI, this shows up in model selection and configuration.