About
Streamline the process of documenting model performance by programmatically managing evaluation data on Hugging Face. This tool enables users to extract benchmark results directly from README markdown tables, import verified scores from the Artificial Analysis API, and execute custom evaluation suites using vLLM, lighteval, or inspect-ai. By standardizing results into the model-index metadata format, it ensures models are correctly represented on leaderboards and Papers with Code while preventing duplicate work through built-in pull request management and validation features.