Builds sophisticated Retrieval-Augmented Generation systems to ground LLM applications in external knowledge and proprietary data.
This skill empowers developers to architect and implement full-stack RAG pipelines, bridging the gap between large language models and private or domain-specific datasets. It provides expert guidance on vector database selection, embedding models, advanced retrieval strategies like hybrid search and reranking, and specialized document chunking methodologies. Whether building a corporate knowledge base assistant or a high-accuracy research tool, this skill ensures grounded, factually accurate responses while providing the tools to minimize hallucinations through structured evaluation and source citations.
主な機能
01Comprehensive implementation patterns for vector databases like Pinecone, Chroma, and Weaviate
02Reranking integration using Cross-Encoders and LLM-based scoring
03Advanced retrieval strategies including Hybrid Search and Multi-Query variations
04Optimized document chunking techniques including semantic and recursive splitting
050 GitHub stars
06Evaluation frameworks for measuring accuracy, groundedness, and retrieval quality
ユースケース
01Creating documentation assistants that access live, domain-specific knowledge bases
02Reducing model hallucinations by grounding AI responses in verified internal facts
03Building proprietary document Q&A systems with verifiable source citations