Implements Retrieval-Augmented Generation (RAG) workflows to ground AI responses with external document context and reduce hallucinations.
The RAG skill provides comprehensive patterns for building end-to-end Retrieval-Augmented Generation pipelines within Claude Code. It guides developers through critical stages of data preparation, including advanced document chunking and embedding generation using local models via Ollama. The skill covers various vector storage options ranging from lightweight Pandas-based solutions to persistent databases like ChromaDB and high-performance FAISS indices. This is an essential toolkit for anyone building Q&A systems, searchable knowledge bases, or AI applications that require high factual accuracy derived from private or specialized datasets.
주요 기능
01Advanced document chunking strategies including fixed-size, sentence-based, and overlapping segments
02Multi-tier vector store implementations using Pandas, ChromaDB, and FAISS
03End-to-end conversational RAG pipelines using LangChain for multi-turn interactions
04Similarity search logic utilizing cosine similarity for relevant context retrieval
050 GitHub stars
06Local embedding generation workflows using Ollama and OpenAI-compatible APIs
사용 사례
01Creating searchable knowledge bases for academic research papers or large legal datasets
02Building private Q&A systems over technical documentation or internal company wikis
03Reducing LLM hallucinations by forcing responses to be grounded in specific source material