Acerca de
This skill provides a comprehensive framework for implementing high-quality vector search within LLM applications. It guides developers through the critical decisions of selecting embedding models—such as OpenAI, Voyage, or local BGE models—and implementing efficient chunking strategies like token-based, semantic, or recursive splitting. By providing production-ready templates for both API-based and local embedding pipelines, it helps developers maximize retrieval accuracy, optimize vector storage costs, and handle domain-specific content like code or legal documents.