AI Dictionary
Context Window
Definition
The amount of text an LLM can consider at one time when generating a response.
Deep Dive
The Context Window, within the realm of Large Language Models (LLMs), refers to the specific amount of text or tokens that the model can simultaneously consider when generating a response or processing a query. This window essentially defines the "memory" of the LLM, allowing it to maintain conversational coherence and understand the surrounding information relevant to its current task. A larger context window enables the model to recall more of the preceding dialogue or document, leading to more relevant, nuanced, and consistent outputs by taking a broader range of input into account.
Examples & Use Cases
- 1An LLM generating a lengthy report, where a larger context window ensures consistency across paragraphs and sections
- 2A chatbot maintaining a coherent conversation over multiple turns, remembering previous user questions and its own responses
- 3A code generation AI leveraging the context of an entire codebase to suggest relevant functions or complete blocks of code
Related Terms
Large Language Model (LLM)Transformer ArchitectureNatural Language Processing (NLP)