01NVIDIA and AMD GPU status and utilization tracking
02Detailed inference performance metrics including tokens per second
03Automated health checks for hardware and Ollama server status
040 GitHub stars
05Real-time VRAM monitoring for loaded Ollama models
06Troubleshooting guides for OOM errors and slow inference