Optimizes token usage for Claude Code and Claude Desktop through intelligent caching and Brotli compression, significantly reducing LLM context window costs.
Token Optimizer serves as a Model Context Protocol (MCP) server designed to dramatically cut down on token consumption when interacting with Claude Code and Claude Desktop. It achieves this by intelligently combining Brotli compression, a persistent SQLite-based caching system, and accurate token counting using tiktoken (GPT-4 tokenizer). The server provides a suite of tools that allow users to compress and cache text, retrieve optimized content, analyze optimization potential, and monitor cache statistics, enabling up to 95%+ token reduction for more efficient LLM interactions.