01Multi-provider configuration for OpenAI, Anthropic, Gemini, Groq, and Mistral
02Real-time streaming for both raw text and structured Pydantic model outputs
03Granular model settings control including temperature, token limits, and timeouts
04Resilience through automated fallback model chaining and custom error handling
05Integrated usage tracking and token limits for cost and performance monitoring
0614 GitHub stars