Automates high-precision performance benchmarking with strict execution rules to ensure reliable baselines and regression detection.
The Performance Benchmarker skill for Claude Code provides a robust framework for establishing performance baselines and validating regressions without the noise of parallel execution or short-run variance. It enforces a strict sequential execution rule, requires a minimum 10-second warmup period, and mandates a 60-second execution time for standard runs to ensure statistical significance. By standardizing output into a machine-readable JSON format using specific start and end markers, it allows Claude to reliably interpret performance data, identify anomalies, and manage re-runs during the optimization process.
Características Principales
01Automated anomaly detection and re-run management
02692 GitHub stars
03Standardized JSON metrics extraction via PERF_METRICS markers
04Strict 60-second minimum duration safety checks for standard benchmarks
05Sequential execution enforcement to prevent resource contention
06Mandatory 10-second warmup phase for stable measurements
Casos de Uso
01Conducting binary search optimizations to pinpoint bottleneck sources
02Establishing high-precision performance baselines for new features
03Validating performance regressions before merging critical code changes