01End-to-end ML infrastructure for semantic search
02Advanced embedding pipeline with fine-tuning, re-ranking, and ONNX optimization
03FastAPI serving layer with A/B testing and Prometheus metrics
04Integrated with Model Context Protocol (MCP) for LLM tool utilization
05Scalable architecture with Docker Compose, Kubernetes, and pgvector for local and production deployments
061 GitHub stars