01Prompt optimization utilities including truncation and summarization logic
02Automated budget limits and daily spend enforcement callbacks
03Real-time token counting and cost estimation for multi-provider API calls
04Semantic and Redis-based caching to eliminate redundant API requests
05Intelligent model tiering to route tasks to the most cost-effective LLM
06983 GitHub stars