Conducts rigorous, elite-level peer reviews of machine learning manuscripts following NeurIPS, ICML, and ICLR standards.
This skill transforms Claude into Dr. Ayaan Rahman, a high-level ML researcher, to provide structured, critical feedback on research papers. It systematically audits technical correctness, experimental rigor, and reproducibility while identifying common red flags like benchmark leakage or unfair comparisons. It is an essential tool for researchers preparing submissions, students learning the review process, or teams evaluating the validity of new internal or external ML methodologies before implementation.
主な機能
01Standardized review rubrics based on top-tier ML conferences like NeurIPS, ICML, and ICLR.
02Detection of common ML red flags including benchmark leakage and unfair baselines.
03Actionable feedback with prioritized suggested experiments and substantive author questions.
04Ethical impact assessment and reproducibility verification checklists.
051 GitHub stars
06Comprehensive technical auditing for correctness, novelty, and experimental integrity.
ユースケース
01Critiquing and validating new ML techniques or papers discovered on arXiv.
02Internal pre-submission review for researchers targeting major machine learning conferences.
03Standardizing the evaluation process for internal company R&D reports and technical whitepapers.