发现data science & ml类别的 Claude 技能。浏览 53 个技能,找到适合您 AI 工作流程的完美功能。
Streamlines the fine-tuning process for over 100 large language models using the LLaMA-Factory framework and QLoRA techniques.
Serves Large Language Models with maximum throughput and efficiency using vLLM's PagedAttention and continuous batching.
Streamlines the fine-tuning of large language models using Axolotl through expert YAML configuration guidance and advanced training techniques.
Simplifies large language model alignment using reference-free preference optimization to improve model performance without the overhead of PPO or DPO.
Enables advanced vision-language capabilities for image understanding, multi-turn visual conversations, and document analysis.
Generates high-fidelity music and sound effects from text descriptions using Meta's AudioCraft framework.
Builds complex AI systems using Stanford's declarative programming framework to optimize prompts and create modular RAG systems automatically.
Transcribes audio, translates speech to English, and automates multilingual audio processing using OpenAI's Whisper models.
Extracts structured, type-safe data from Large Language Models using Pydantic validation and automatic retries.
Guarantees valid, type-safe JSON and structured outputs from Large Language Models using grammar-based constraints.
Enables zero-shot image classification and semantic image search by connecting visual concepts with natural language.
Accelerates LLM inference speeds by up to 3.6x using advanced decoding techniques like Medusa heads and lookahead decoding.
Optimizes AI models for efficient local inference using the GGUF format and llama.cpp quantization techniques.
Integrates Salesforce's BLIP-2 framework to enable advanced image captioning, visual question answering, and multimodal reasoning within AI workflows.
Deploys high-performance Reinforcement Learning from Human Feedback (RLHF) workflows using Ray and vLLM acceleration for large-scale model alignment.
Curates high-quality datasets for LLM training using GPU-accelerated deduplication, filtering, and PII redaction.
Interprets and manipulates neural network internals for any PyTorch model, including massive foundation models via remote execution.
Optimizes large-scale AI model training using PyTorch Fully Sharded Data Parallelism for efficient memory management and scaling.
Performs declarative causal interventions and mechanistic interpretability experiments on PyTorch models.
Accelerates LLM fine-tuning workflows with Unsloth to achieve up to 5x faster training speeds and 80% reduced memory consumption.
Manages high-performance vector search and storage for production RAG and AI applications using Pinecone's serverless infrastructure.
Connects LLMs to private data sources through advanced document ingestion, vector indexing, and retrieval-augmented generation (RAG) pipelines.
Optimizes Transformer models using Flash Attention to achieve significant speedups and memory reductions during training and inference.
Orchestrates distributed machine learning training across clusters to scale PyTorch, TensorFlow, and Hugging Face models.
Merges multiple fine-tuned AI models using mergekit to combine specialized capabilities like math and coding without expensive retraining.
Implements Group Relative Policy Optimization (GRPO) using the TRL library to enhance model reasoning and structured output capabilities.
Deploys and optimizes LLM inference on CPU, Apple Silicon, and consumer hardware using GGUF quantization.
Accelerates Large Language Model inference on NVIDIA GPUs using state-of-the-art optimization techniques for maximum throughput and minimal latency.
Manages the machine learning lifecycle by tracking experiments, versioning models, and streamlining production deployments.
Implements and optimizes RWKV architectures, a hybrid RNN-Transformer model offering linear-time inference and infinite context windows.
Scroll for more results...