01Local text embedding generation via Ollama's nomic-embed-text model.
02Streamlined CLI for direct chat, system prompts, and structured JSON output.
03Ultra-fast chat inference using Llama-3.3-70b-versatile and Llama-3.1-8b.
04Support for piped stdin to facilitate batch processing of text embeddings.
050 GitHub stars
06Robust error handling for API rate limits and local service availability.