Discover Agent Skills for data science & ml. Browse 61 skills for Claude, ChatGPT & Codex.
Implements programmable safety rails and runtime validation for LLM applications using NVIDIA's NeMo Guardrails framework.
Implements language-independent subword tokenization using BPE and Unigram algorithms for robust NLP model training and inference.
Decomposes neural network activations into interpretable, sparse features using SAELens for deep mechanistic interpretability research.
Optimizes LLM serving and structured data generation with RadixAttention prefix caching for high-performance agentic workflows.
Extracts and validates structured data from LLM responses using Pydantic for reliable, type-safe outputs and automatic retries.
Simplifies PyTorch distributed training across multiple GPUs, TPUs, and nodes with minimal code changes and a unified API.
Deploys and manages high-performance RLHF training pipelines for large-scale language models using Ray and vLLM acceleration.
Implements Anthropic's Constitutional AI method to train harmless AI models through self-critique and reinforcement learning from AI feedback.
Quantizes Large Language Models to ultra-low bit precision without requiring calibration datasets for efficient inference and fine-tuning.
Scales LLM post-training via reinforcement learning by integrating Megatron-LM training with high-throughput SGLang inference.
Evaluates Large Language Models across 100+ industry-standard benchmarks using NVIDIA's enterprise-grade containerized architecture.
Orchestrates autonomous teams of specialized AI agents to collaborate on complex, multi-step tasks and production workflows.
Generates high-quality images from text and performs advanced image-to-image transformations using the HuggingFace Diffusers library.
Extends Transformer model context windows using advanced positional encoding and interpolation techniques like RoPE, YaRN, and ALiBi.
Manages high-performance vector embeddings and metadata for RAG applications and semantic search using the open-source Chroma database.
Evaluates Large Language Models across 60+ academic benchmarks using standardized prompts and metrics for reproducible research.
Implements state-of-the-art vision-language pre-training to enable high-quality image captioning and visual question answering within AI workflows.
Optimizes Large Language Model inference for maximum throughput and ultra-low latency on NVIDIA GPUs.
Generates high-quality sentence, text, and image embeddings for RAG, semantic search, and clustering using state-of-the-art transformer models.
Integrates Weights & Biases into your workflow to track machine learning experiments, visualize training metrics, and manage model artifacts in real-time.
Facilitates causal interventions on PyTorch models using a declarative framework for mechanistic interpretability experiments.
Implements efficient similarity search and clustering for dense vectors at scale using Facebook AI's high-performance library.
Combines multiple fine-tuned AI models into a single high-performance model without requiring additional training or expensive GPU resources.
Implements and optimizes Selective State Space Models (SSM) for high-performance sequence modeling and long-context AI applications.
Implements PyTorch-native agentic reinforcement learning workflows using Meta's torchforge library for scalable algorithm experimentation.
Integrates Pinecone's managed vector database to power high-performance RAG, semantic search, and recommendation systems.
Facilitates high-performance distributed data processing and streaming for large-scale machine learning workloads.
Build and optimize complex AI systems using declarative programming instead of manual prompt engineering.
Manage the complete machine learning lifecycle including experiment tracking, model versioning, and deployment using the MLflow framework.
Builds sophisticated RAG applications by connecting LLMs to private data through advanced indexing and retrieval patterns.
Scroll for more results...