Discover Agent Skills for data science & ml. Browse 61 skills for Claude, ChatGPT & Codex.
Accelerates LLM data curation using GPU-powered deduplication, quality filtering, and PII redaction at scale.
Deploys and manages high-performance RLHF training pipelines for large-scale language models using Ray and vLLM acceleration.
Decomposes neural network activations into interpretable, sparse features using SAELens for deep mechanistic interpretability research.
Optimizes Large Language Model inference for maximum throughput and ultra-low latency on NVIDIA GPUs.
Facilitates causal interventions on PyTorch models using a declarative framework for mechanistic interpretability experiments.
Evaluates Large Language Models across 60+ academic benchmarks using standardized prompts and metrics for reproducible research.
Implements state-of-the-art vision-language pre-training to enable high-quality image captioning and visual question answering within AI workflows.
Generates high-quality sentence, text, and image embeddings for RAG, semantic search, and clustering using state-of-the-art transformer models.
Integrates Weights & Biases into your workflow to track machine learning experiments, visualize training metrics, and manage model artifacts in real-time.
Implements language-independent subword tokenization using BPE and Unigram algorithms for robust NLP model training and inference.
Implements and manages RWKV architectures for efficient, linear-time AI inference and long-context processing.
Build and optimize complex AI systems using declarative programming instead of manual prompt engineering.
Implements and optimizes Selective State Space Models (SSM) for high-performance sequence modeling and long-context AI applications.
Serves large language models with high throughput and low latency using PagedAttention and continuous batching.
Streamlines the fine-tuning of large language models using Axolotl through expert YAML configuration and advanced training pattern guidance.
Implements PyTorch-native agentic reinforcement learning workflows using Meta's torchforge library for scalable algorithm experimentation.
Compresses large language models to 4-bit precision to enable high-speed inference and deployment on consumer-grade hardware.
Implements Simple Preference Optimization to align Large Language Models without requiring a reference model.
Transcribes and translates audio across 99 languages using OpenAI's robust general-purpose speech recognition models.
Facilitates high-performance distributed data processing and streaming for large-scale machine learning workloads.
Optimizes large-scale model training using DeepSpeed configurations, ZeRO optimization stages, and high-performance I/O management.
Compresses large language models into efficient student models while retaining performance through advanced teacher-student transfer techniques.
Evaluates AI code generation models using industry-standard benchmarks and pass@k metrics.
Integrates LLaVA to enable sophisticated visual instruction following and multi-turn conversational image understanding.
Accelerates LLM inference speed by up to 3.6x using speculative decoding, Medusa heads, and lookahead techniques without sacrificing model quality.
Standardizes and accelerates PyTorch model training with built-in support for distributed computing, logging, and engineering best practices.
Builds sophisticated LLM applications using agents, chains, and Retrieval-Augmented Generation (RAG) with a unified interface.
Optimizes Large Language Models using 4-bit activation-aware weight quantization to achieve 3x faster inference with minimal accuracy loss.
Provides high-performance, Rust-based tokenization tools for building and training NLP models with support for BPE, WordPiece, and Unigram algorithms.
Integrates OpenAI's CLIP model to enable zero-shot image classification, semantic image search, and cross-modal retrieval without task-specific training.
Scroll for more results...