01LLM-as-judge (llm-rubric) implementation for qualitative output scoring
020 GitHub stars
03Cost-free preview mode using the Echo provider to verify prompt rendering
04Automated prompt benchmarking across multiple providers like Anthropic and OpenAI
05Custom Python assertion support for specialized metric and logic validation
06Comprehensive few-shot example management for complex prompt patterns