Token Limit (aka Context Window) is the maximum amount of text an AI model can process in a single request—both your input and its output combined. Tokens are chunks of text (roughly 4 characters or 0.75 words in English). GPT-4 has a 8k-128k token limit depending on version; Claude has up to 200k tokens. Hit the limit and the AI either truncates, errors, or forgets earlier parts of the conversation.
You don't "use" token limits—you work within them. Track token usage to avoid hitting limits, optimize prompts to fit more context, chunk large documents into smaller pieces, use summarization to compress info, or upgrade to models with larger context windows. Token limits directly affect cost (you pay per token) and what's possible (can't analyze a 500-page doc in one go on GPT-3.5).
Ai Vocabulary
Context window—how much AI can remember