About
This skill provides a comprehensive framework for building production-grade Retrieval-Augmented Generation (RAG) systems using the llmemory library. It streamlines the entire pipeline from document ingestion and hierarchical chunking to sophisticated retrieval methods like hybrid search and query expansion. By incorporating advanced features such as intelligent query routing to prevent hallucinations and multiple reranking options—including local cross-encoders and OpenAI models—it ensures that LLM responses are accurately grounded in specific knowledge bases. It is an essential tool for developers creating technical assistants, documentation search engines, or any AI application requiring reliable, source-backed answers.