Learn what token limits mean for AI chatbots, how they affect conversation length, and strategies to work within context window constraints.
More about Token Limit
Token Limit (also called context window or context window) is the maximum number of tokens an AI model can process in a single request, including both input and output. Tokens are roughly equivalent to words or word pieces—approximately 4 characters per token in English.
Token limits affect how much conversation history and context an AI chatbot can "remember" in a single interaction. Models like GPT-4 Turbo offer 128K tokens, while Claude 3 supports up to 200K tokens. Managing token limits effectively is crucial for long conversations and RAG applications.