fancyant
TTCTheo's Typesafe Cult
•Created by fancyant on 4/2/2025 in #questions
T3 Pro Plan Context Window Size
I received an official response regarding the context window size limits for the T3 Chat models and thought it might be helpful to share:
Most models on T3 Chat support the full context window size. There's no message limit, other than the model's overall context window limit. The exception is Anthropic models (Claude), which are capped at 32K tokens.
Here's a breakdown of the context window sizes for each model:
GPT-4o-mini - 128,000 Tokens
GPT-4o - 128,000 Tokens
o3-mini - 200,000 Tokens
Claude 3.5 Sonnet - 32,000 Tokens
Claude 3.7 Sonnet - 32,000 Tokens
Claude 3.7 Sonnet (Reasoning) - 32,000 Tokens
DeepSeek v3 (Fireworks) - 128,000 Tokens
DeepSeek v3 (0324) - 64,000 Tokens
DeepSeek R1 (OpenRouter) - 128,000 Tokens
DeepSeek R1 (Llama Distilled) - 32,000 Tokens
DeepSeek R1 (Qwen Distilled) - 16,000 Tokens
Gemini 2.0 Flash - 1,000,000 Tokens
Gemini 2.0 Flash Lite - 1,000,000 Tokens
Gemini 2.5 Pro - 1,000,000 Tokens
Llama 3.3 70b - 128,000 Tokens
Qwen 2.5 32b - 8,000 Tokens
Qwen qwq-32b - 128,000 Tokens
4 replies