01Quantize Hugging Face models to GGUF, GPTQ, or AWQ with specified bit widths
02Receive hardware-aware recommendations for optimal format and bit width
03Perform all operations in a single tool call for automation
040 GitHub stars
05Inspect model parameters, architecture details, and size estimates
06Check available quantization backends installed on the system