概要
This tool functions as a robust Model Context Protocol (MCP) server designed to empower Large Language Models (LLMs) with Retrieval-Augmented Generation (RAG) capabilities. It supports ingesting and processing documents across a wide array of formats, including PDF, Word, Excel, HTML, and more. Utilizing LangChain for its core RAG pipeline and ChromaDB for persistent vector storage, it offers flexible embedding options with either OpenAI for high performance or Ollama for local, free processing. This enables semantic search and retrieval of relevant information from custom datasets, significantly improving the contextual understanding of LLMs.