01Lossless KV cache persistence for LLM sessions
02Clone contexts to explore alternative conversation branches
03Tiered storage management (GPU, CPU, Disk, Redis)
040 GitHub stars
05Freeze and thaw LLM contexts to/from persistent storage
06Session isolation with unique `cache_salt`