소개
This Claude Code skill provides a comprehensive framework for adapting pre-trained Large Language Models (LLMs) to follow specific instructions through Supervised Fine-Tuning (SFT). By leveraging the Unsloth library and TRL's SFTTrainer, it enables up to 2x faster training speeds and significantly reduced memory overhead, making it ideal for local GPU and Jupyter environments. The skill includes specialized patterns for training thinking or reasoning models, handles complex dataset formatting for various chat templates, and provides clear paths for exporting fine-tuned models to GGUF for Ollama deployment.