Explore our collection of Agent Skills to enhance your AI workflow.
Extracts and validates structured data from LLM responses using Pydantic for reliable, type-safe outputs and automatic retries.
Implements efficient similarity search and clustering for dense vectors at scale using Facebook AI's high-performance library.
Visualizes machine learning training metrics, model architectures, and performance profiles using Google's TensorBoard toolkit.
Trains and optimizes Mixture of Experts (MoE) models to scale AI capacity with significantly reduced compute costs.
Moderates LLM inputs and outputs using Meta's specialized LlamaGuard models to ensure safety and policy compliance across six critical categories.
Drafts publication-ready ML and AI research papers for top-tier conferences like NeurIPS and ICML directly from your research codebase.
Optimizes large language models for efficient local inference using GGUF format and llama.cpp quantization techniques.
Optimizes LLM fine-tuning workflows by providing expert guidance on Unsloth implementation for faster training and reduced memory usage.
Optimizes large-scale Reinforcement Learning training for Mixture-of-Experts models using high-performance quantization and speculative rollout techniques.
Implements and manages RWKV architectures for efficient, linear-time AI inference and long-context processing.
Accelerates LLM data curation using GPU-powered deduplication, quality filtering, and PII redaction at scale.
Optimizes Transformer attention mechanisms using Flash Attention for significant speedups and 10-20x memory reduction.
Streamlines the fine-tuning of 100+ large language models using LLaMA-Factory with support for QLoRA and multimodal architectures.
Serves large language models with high throughput and low latency using PagedAttention and continuous batching.
Implements advanced PyTorch FSDP2 sharding and distributed checkpointing for efficient large-scale model training.
Trains large-scale language models using NVIDIA Megatron-Core with advanced parallelism strategies for maximum GPU efficiency.
Fine-tunes and aligns Large Language Models using Supervised Fine-Tuning and Reinforcement Learning from Human Feedback via the TRL library.
Accelerates large-scale LLM pretraining using PyTorch-native 4D parallelism and Float8 optimization.
Generates high-quality images from text and performs advanced image-to-image transformations using the HuggingFace Diffusers library.
Extends Transformer model context windows using advanced positional encoding and interpolation techniques like RoPE, YaRN, and ALiBi.
Scales LLM post-training via reinforcement learning by integrating Megatron-LM training with high-throughput SGLang inference.
Implements PyTorch-native agentic reinforcement learning workflows using Meta's torchforge library for scalable algorithm experimentation.
Provides comprehensive LLM observability, tracing, and evaluation tools to debug and monitor AI applications in real-time.
Integrates OpenAI's CLIP model to enable zero-shot image classification, semantic image search, and cross-modal retrieval without task-specific training.
Implements, fine-tunes, and deploys high-performance Large Language Models using Lightning AI's LitGPT framework.
Builds, deploys, and manages continuous AI agents through a visual workflow builder or specialized development toolkit.
Optimizes large-scale model training using DeepSpeed configurations, ZeRO optimization stages, and high-performance I/O management.
Integrates LLaVA to enable sophisticated visual instruction following and multi-turn conversational image understanding.
Deploys and manages high-performance RLHF training pipelines for large-scale language models using Ray and vLLM acceleration.
Transcribes and translates audio across 99 languages using OpenAI's robust general-purpose speech recognition models.
Scroll for more results...