概要
This skill provides a comprehensive framework for running industry-standard benchmarks against AI models to assess vulnerabilities, bias, and safety risks. By integrating datasets such as HarmBench for harmful behaviors, JailbreakBench for prompt injection defense, and RobustBench for adversarial robustness, it enables security researchers and developers to quantify an AI system's resistance to attacks and alignment with safety standards. It bridges the gap between model development and security auditing by providing structured mappings to the OWASP Top 10 for LLMs and the NIST AI Risk Management Framework.