Tokens
10 articles about tokens.
Claude Pro vs API Cost Comparison: Actual Numbers, Breakeven Math, and When to Switch
Detailed cost comparison of Claude Pro subscription ($20/mo) vs API pay-per-token pricing. Includes breakeven calculations, token math, and real usage scenarios.
A/B Testing Claude Code Hooks - Optimizing Token Usage
Cache read jumps show that hooks front-load context effectively. How to A/B test Claude Code hooks for performance and measure the impact on token consumption.
Tokens Used Loading MCP Tools - Measuring and Reducing the Overhead
31 MCP tools can eat 3-5k tokens just loading schemas. Here is how to measure and optimize MCP tool token overhead in Cursor, Claude Code, and other AI
The Hidden Token Cost of MCP Tools in Cursor and How to Fix It
31 Atlassian MCP tools burn 2-3k tokens per request just from schema definitions. A 400-tool enterprise server can exceed Claude's entire context window before you ask anything. Here's how to cut tool overhead by 85-100x.
The Sanitization Tax
Raw accessibility tree data is messy but information-rich. The tradeoff between sanitizing it for cleanliness and keeping tokens low is harder than it looks.
100M Tokens Tracked: 99.4% Were Input and Parallel Agents Make It Worse
After tracking 100M tokens, 99.4% were input tokens. Running parallel Claude Code agents multiplies the input cost problem. Here is how CLAUDE.md scoping, prompt caching, and context architecture helps.
Browser Automation: Accessibility Snapshots vs Screenshots - Saving Tokens by Skipping Pixels
Switching from screenshots to accessibility snapshots for browser automation saved us massive token costs. Here is why structured data beats pixel analysis
Embeddings vs Tokens - How AI Agent Memory Actually Works
Embeddings aren't tokens. They're dense vector representations that capture semantic meaning and power similarity search for AI agent memory retrieval.
The 1M Context Trap: Why More Context Makes Claude Lazier
Research on 18 frontier models confirms every one degrades with more context. The 'lost-in-the-middle' effect causes 30%+ accuracy drops. The counterintuitive fix: use less context, not more.
Opus Token Burn Rate - Watching It Write, Delete, and Rewrite 200-Line Functions
Opus does not just burn tokens - it vaporizes them. The write-delete-rewrite cycle where Opus creates 200 lines, decides it does not like them, and starts over.