Enables Retrieval Augmented Generation (RAG) capabilities for Large Language Models (LLMs) by indexing and retrieving relevant information from documents.
Rag empowers Large Language Models (LLMs) to answer questions based on your document content by indexing and retrieving relevant information efficiently. It parses documents into chunks, generates vector embeddings using various embedding providers like Ollama and OpenAI, and stores them in a local vector store. This enables downstream LLMs, via MCP clients, to generate contextually relevant responses by querying the stored embeddings.