概要
This skill enables Claude to efficiently manage the end-to-end lifecycle of Large Language Models on macOS hardware. It provides domain-specific guidance for utilizing the MLX framework to perform high-speed inference, convert Hugging Face models to optimized formats, and execute memory-efficient fine-tuning via LoRA and QLoRA. By leveraging Apple's unified memory architecture, it helps developers maximize local hardware performance for complex AI tasks without the need for external GPU clusters.