Skip to main content

COMPEL Glossary / context-window

Context Window

A context window is the maximum amount of text (measured in tokens) that a large language model can process at one time.

What this means in practice

It encompasses the system prompt, conversation history, retrieved documents, tool outputs, and the model's own generated response. Modern LLMs have context windows ranging from thousands to millions of tokens. Context window size matters for enterprise applications because it determines how much information the model can consider when generating a response: a legal AI assistant analyzing a 200-page contract needs a large context window to reason about the full document. When conversations exceed the context window, older information is lost, which can cause the model to forget instructions, constraints, or earlier context -- a significant concern for agentic systems executing multi-step tasks.

Why it matters

Context window size determines how much information an LLM can consider simultaneously, directly affecting capability for enterprise applications like contract analysis, document summarization, and complex reasoning tasks. When conversations exceed the context window, critical instructions and earlier context are lost — a significant concern for agentic systems executing multi-step tasks. Understanding context window limitations prevents organizations from deploying LLMs in scenarios where they cannot maintain coherent reasoning.

How COMPEL uses it

Context window considerations are part of the Technology pillar's assessment during Calibrate, informing model selection decisions. The Model stage designs application architectures that account for context window limitations through techniques like RAG and summarization. During Produce, context management strategies are implemented. The Evaluate stage monitors whether context window constraints are causing quality issues in production AI applications.

Related Terms

Other glossary terms mentioned in this entry's definition and context.