Discover Agent Skills for data science & ml. Browse 61 skills for Claude, ChatGPT & Codex.
Optimizes and deploys PyTorch models to Arm Cortex-M processors using the CMSIS-NN backend.
Connects and controls ComfyUI via API to generate and edit high-quality images with token-saving efficiency.
Connects Claude directly to ComfyUI via API to automate professional image generation and editing workflows with high token efficiency.
Integrates ComfyUI with Claude for token-efficient image generation and automated editing workflows.
Integrates LLaVA to enable sophisticated visual instruction following and multi-turn conversational image understanding.
Interprets and manipulates neural network internals across local and remote models using the nnsight library and NDIF execution.
Builds sophisticated LLM applications using agents, chains, and Retrieval-Augmented Generation (RAG) with a unified interface.
Integrates Weights & Biases into your workflow to track machine learning experiments, visualize training metrics, and manage model artifacts in real-time.
Deploys and manages high-performance RLHF training pipelines for large-scale language models using Ray and vLLM acceleration.
Evaluates AI code generation models using industry-standard benchmarks and pass@k metrics.
Accelerates LLM inference speed by up to 3.6x using speculative decoding, Medusa heads, and lookahead techniques without sacrificing model quality.
Standardizes and accelerates PyTorch model training with built-in support for distributed computing, logging, and engineering best practices.
Compresses large language models into efficient student models while retaining performance through advanced teacher-student transfer techniques.
Optimizes Large Language Models using 4-bit activation-aware weight quantization to achieve 3x faster inference with minimal accuracy loss.
Ensures guaranteed valid JSON, XML, and type-safe code generation from LLMs using constrained token sampling and Pydantic models.
Enforces structured outputs and grammatical constraints on Large Language Models to guarantee valid JSON, code, and regex-compliant text.
Accelerates large-scale LLM pretraining using PyTorch-native 4D parallelism and Float8 optimization.
Optimizes Transformer attention mechanisms using Flash Attention for significant speedups and 10-20x memory reduction.
Implements zero-shot image segmentation using Meta AI's SAM to identify and extract objects via points, boxes, or automatic mask generation.
Streamlines the fine-tuning of 100+ large language models using LLaMA-Factory with support for QLoRA and multimodal architectures.
Implements, fine-tunes, and deploys high-performance Large Language Models using Lightning AI's LitGPT framework.
Quantizes Large Language Models to 8-bit or 4-bit formats to reduce memory usage by up to 75% with minimal accuracy loss.
Optimizes large-scale Reinforcement Learning training for Mixture-of-Experts models using high-performance quantization and speculative rollout techniques.
Deploys high-performance LLM inference on CPU, Apple Silicon, and non-NVIDIA GPUs using GGUF quantization.
Reduces Large Language Model size and accelerates inference using advanced pruning techniques like Wanda and SparseGPT.
Trains large language models using advanced reinforcement learning algorithms like GRPO and PPO with the production-ready verl framework.
Optimizes LLM fine-tuning workflows by providing expert guidance on Unsloth implementation for faster training and reduced memory usage.
Implements a minimalist, educational GPT-2 architecture in PyTorch for learning and training transformer models from scratch.
Compresses large language models to 4-bit precision to enable high-speed inference and deployment on consumer-grade hardware.
Provides high-performance, Rust-based tokenization tools for building and training NLP models with support for BPE, WordPiece, and Unigram algorithms.
Scroll for more results...