01Detailed performance tracking including token-per-second and eval metrics
02Local embedding generation for RAG and vector search workflows
03Support for both streaming and non-streaming response patterns
040 GitHub stars
05Complete coverage of Ollama /api/* endpoints for chat and generation
06Comprehensive model management including list, show, copy, and delete