概要
The Ollama GPU Monitor skill provides specialized tools for managing hardware performance during local AI model execution. It allows users to track NVIDIA and AMD GPU status, monitor real-time VRAM consumption, verify model loading status via the Ollama API, and calculate precise inference performance metrics such as tokens per second. This skill is essential for developers debugging slow inference, managing resource constraints, or benchmarking different LLMs within a Bazzite AI environment.