Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[Suggestion] MLPerf reproducibility/repeatability methodology from ACM/IEEE/NeurIPS? #1080

Open
Tracked by #1595 ...
gfursin opened this issue Jan 30, 2024 · 1 comment

Comments

@gfursin
Copy link
Contributor

gfursin commented Jan 30, 2024

Following many recent discussions at MLCommons about improving the repeatability and reproducibility of MLPerf inference benchmarks, we suggest to look at similar initiatives at computer systems conferences (artifact evaluation and reproducibility initiatives) and maybe adopt their methodology and badges:

Our repeatability study for MLPerf inference v3.1 highlights similar repeatability issues to what we already saw in compiler, systems and ML conferences:

  • some submissions miss critical files that makes it impossible to rebuild and rerun their experiments
  • some submissions worked fine at the time of submission but got outdated due to newer dependencies
  • manual modifications of submissions, configurations, paths, environment variables, etc is often required making repeatability very error-prone
  • containers were useful as a snapshot but they do not guarantee to be working on new hardware or with new software

A potential solution is improve repeatability of MLPerf submissions (full reproducibility is probably too costly and impossible at this stage) by introducing MLPerf reproducibility badges similar to ACM reproducibility badges:

  • "MLPerf submission available" badge is published along submission only if all artifacts are publicly available for external user to rebuild the submission (code, data, configurations, workflows, etc) .

  • "MLPerf submission functional/repeatable" badge if anyone can perform a short valid run for a given submission in a fully automated way. The MLCommons Automation and Reproducibility TaskForce can then extend MLCommons CM workflow for MLPerf to run that submission via a common interface in a unified way.

We can evaluate results after submission deadline and before the publication deadline, and assign badges to all results in the final table that is officially published. It may motivate everyone to improve the quality of their submission and get all such badges in the future instead of the community discovering such issues after MLPerf publication of results.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

No branches or pull requests

2 participants