01Standardized patterns for parallel model execution and background processing.
02Context-aware model preference learning and automatic reuse via local storage.
03Compatibility with local AI runners including Ollama, LM Studio, vLLM, and MLX.
04Multi-backend routing support for OpenAI, Google, MiniMax, Kimi, and GLM.
05Deterministic CLI-based model invocation for 100% reliable external calls.
06230 GitHub stars