The Quick Reference
You don’t need a tokenizer to estimate costs. These rough conversions work for English text with GPT-4o’s tokenizer:
// Token estimation rules of thumb
1 token ≈ 4 characters or 0.75 words
100 tokens ≈ 75 words (a short paragraph)
1,000 tokens ≈ 750 words (about 1.5 pages)
A tweet ≈ 40–70 tokens
An email ≈ 130–150 tokens
A 10-page doc≈ 3,000–4,000 tokens
1M tokens ≈ 750,000 words ≈ ~1,500 pages
Context Window Sizes
Context windows define the maximum number of tokens a model can process in a single request (input + output combined). As of March 2026: GPT-4o supports 128K tokens (~96,000 words), Claude 3.5 Sonnet supports 200K tokens, and Gemini 2.0 supports up to 2M tokens. But bigger isn’t always better — performance degrades well before hitting the limit.
Key insight: A 128K context window at GPT-4o pricing ($2.50/M input tokens) costs $0.32 if you fill it completely. Run 10,000 full-context requests per day and you’re spending $96,000/month on input tokens alone.