About
This skill provides a comprehensive framework for implementing production-grade RAG architectures, enabling LLMs to access and process domain-specific or private data with high precision. It covers the entire implementation lifecycle, from advanced document chunking and embedding generation to sophisticated retrieval strategies like hybrid search, multi-query expansion, and cross-encoder reranking. By following these patterns, developers can significantly reduce hallucinations and build reliable AI applications such as documentation assistants, enterprise search tools, and grounded Q&A systems that cite their sources.