01Optimized for high-speed, low-cost models like Claude Haiku and GPT-Mini
02Hypothetical document generation to bridge semantic vocabulary gaps
03Graceful fallback to direct embedding on generation timeouts
0469 GitHub stars
05Integrated caching mechanism to minimize latency and API consumption
06Batch processing support for multi-concept query decomposition