Optimizes LLM context windows through compaction, masking, and caching strategies to maximize performance and minimize token costs.
The Context Optimization skill empowers Claude to handle complex, long-running tasks by strategically managing its context window through advanced engineering techniques. By implementing compaction (summarizing older messages), observation masking (eliding verbose tool outputs), and KV-cache optimization, this skill allows Claude to effectively double or triple its useful capacity. It is essential for production-grade agent systems where maintaining high reasoning quality, reducing latency, and controlling API costs are critical requirements.
主な機能
01Context partitioning via sub-agent task isolation
02KV-cache optimization for prefix stability and faster inference
030 GitHub stars
04Observation masking for verbose tool and API outputs
05Trigger-based budget management for token utilization
06Intelligent context compaction and summarization
ユースケース
01Reducing operational costs by eliding redundant data in large prompts
02Improving response latency in complex multi-step reasoning trajectories
03Scaling long-running agent sessions without losing critical history