Optimize AI coding agent context windows by providing local RAG memory, reducing prompt tokens with hybrid semantic and keyword search.
Sponsored
TokenKeeper acts as a local Retrieval Augmented Generation (RAG) memory system for AI coding agents, specifically designed to manage large project contexts without consuming excessive prompt tokens. It intelligently indexes project documents and code into a local vector database, enabling AI agents to query only the most relevant information rather than loading entire files. This process significantly reduces prompt token usage by up to 80%, allowing agents to maintain focus on the task, prevent context window overflow, and ultimately improve the quality of their reasoning and code generation.
主要功能
01Code-Aware Indexing (AST parsing for Python)
02Auto-Indexing with File Watcher
031 GitHub stars
04Hybrid Semantic + Keyword Search
05Local-First Processing (Ollama, ChromaDB)
06Heading-Aware Markdown Chunking
使用案例
01Optimizing AI coding agent context windows in large projects
02Reducing prompt token costs for AI-powered development
03Enhancing AI agent reasoning and code generation quality