Discover Agent Skills for data science & ml. Browse 61skills for Claude, ChatGPT & Codex.
Implements and optimizes RWKV architectures, a hybrid RNN-Transformer model offering linear-time inference and infinite context windows.
Streamlines deep learning development by decoupling research code from engineering boilerplate for automated distributed training and hardware scaling.
Evaluates Large Language Models across 60+ academic benchmarks to measure reasoning, coding, and mathematical capabilities using industry-standard metrics.
Enforces structured LLM outputs using regex and grammars to guarantee valid JSON, XML, and code generation.
Orchestrates teams of autonomous AI agents to collaborate on complex tasks through role-based delegation and memory.
Implements Meta AI's foundation model for high-precision zero-shot image segmentation using points, boxes, and masks.
Interprets and manipulates neural network internals for any PyTorch model, including massive foundation models via remote execution.
Extends transformer context windows using RoPE, YaRN, and ALiBi techniques to process documents exceeding 128k tokens.
Accelerates large-scale similarity search and clustering for dense vectors using Facebook AI's high-performance library.
Streamlines the fine-tuning of large language models using Axolotl through expert YAML configuration guidance and advanced training techniques.
Decomposes complex neural network activations into sparse, interpretable features to understand and steer model behavior.
Generates state-of-the-art text and image embeddings for RAG, semantic search, and clustering tasks.
Optimizes large-scale AI model training using PyTorch Fully Sharded Data Parallelism for efficient memory management and scaling.
Compresses Large Language Models using advanced techniques like Wanda and SparseGPT to reduce memory footprint and accelerate inference speeds.
Implements and optimizes Mamba-based Selective State Space Models for high-efficiency sequence modeling and long-context AI research.
Quantizes Large Language Models to 4-bit or 8-bit formats to reduce GPU memory usage by up to 75% with minimal accuracy loss.
Optimizes large-scale AI model training using DeepSpeed's ZeRO, pipeline parallelism, and high-performance DeepNVMe I/O handling.
Provides high-performance, Rust-optimized text tokenization for NLP research and production-grade machine learning pipelines.
Tracks machine learning experiments and manages model lifecycles with real-time visualization and collaborative tools.
Generates high-quality images and performs advanced image transformations using Stable Diffusion models and the HuggingFace Diffusers library.
Simplifies PyTorch distributed training by providing a unified API for DDP, DeepSpeed, and FSDP with minimal code changes.
Optimizes Large Language Models using activation-aware 4-bit quantization to achieve 3x inference speedups and significant memory reduction with minimal accuracy loss.
Evaluates AI code generation models across multiple programming languages and benchmarks using standardized pass@k metrics.
Generates high-fidelity music and sound effects from text descriptions using Meta's AudioCraft framework.
Optimizes Transformer models using Flash Attention to achieve significant speedups and memory reductions during training and inference.
Enables advanced vision-language capabilities for image understanding, multi-turn visual conversations, and document analysis.
Serves Large Language Models with maximum throughput and efficiency using vLLM's PagedAttention and continuous batching.
Compresses large language models using teacher-student learning techniques to reduce inference costs while maintaining high performance.
Accelerates LLM inference speeds by up to 3.6x using advanced decoding techniques like Medusa heads and lookahead decoding.
Performs declarative causal interventions and mechanistic interpretability experiments on PyTorch models.
Scroll for more results...