概要
This skill provides a specialized framework for managing large-scale AI context windows, allowing developers to extend effective capacity by up to 3x without increasing model limits. It implements sophisticated techniques such as compaction of long histories, masking of verbose tool outputs, and KV-cache optimization to maximize token efficiency. By strategically partitioning tasks across sub-agents and managing context budgets, it ensures that production agent systems remain performant, cost-effective, and low-latency even during complex, long-running interactions.