Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Update docs: SCC24, fix broken redirect #1843

Merged
merged 212 commits into from
Sep 24, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
Show all changes
212 commits
Select commit Hold shift + click to select a range
22b063c
Support batch-size in llama2 run
arjunsuresh Feb 26, 2024
c9c5640
Merge branch 'mlcommons:master' into master
arjunsuresh Feb 26, 2024
18f521c
Merge branch 'mlcommons:master' into master
arjunsuresh Apr 11, 2024
5773906
Merge branch 'mlcommons:master' into master
arjunsuresh May 2, 2024
5fe4dfd
Add Rclone-Cloudflare download instructions to README.md
nathanw-mlc Feb 21, 2024
8a00168
Add Rclone-Cloudflare download instructiosn to README.md
nathanw-mlc Feb 21, 2024
1b69968
Minor wording edit to README.md
nathanw-mlc Feb 21, 2024
d30a0ca
Add Rclone-Cloudflare download instructions to README.md
nathanw-mlc Feb 21, 2024
48f8bbb
Add Rclone-GDrive download instructions to README.md
nathanw-mlc Feb 21, 2024
0e70449
Add new and old instructions to README.md
nathanw-mlc Feb 21, 2024
ef482c3
Tweak language in README.md
nathanw-mlc Feb 21, 2024
faa0134
Language tweak in README.md
nathanw-mlc Feb 21, 2024
c7945ac
Minor language tweak in README.md
nathanw-mlc Feb 21, 2024
949ff6b
Fix typo in README.md
nathanw-mlc Feb 23, 2024
22d7072
Count error when logging errors: submission_checker.py
arjunsuresh Mar 14, 2024
6f2f14e
Fixes #1648, restrict loadgen uncommitted error message to within the…
arjunsuresh Feb 28, 2024
3361249
Update test-rnnt.yml (#1688)
arjunsuresh May 2, 2024
b747899
Added docs init
arjunsuresh May 2, 2024
c023fc9
Merge branch 'mlcommons:master' into docs
arjunsuresh May 2, 2024
6fe12df
Fix benchmark URLs
arjunsuresh May 3, 2024
949f8f7
Fix links
arjunsuresh May 8, 2024
120aced
Add _full variation to run commands
arjunsuresh May 8, 2024
dccbe1e
Added script flow diagram
arjunsuresh May 13, 2024
6b36d66
Merge branch 'mlcommons:master' into master
arjunsuresh May 13, 2024
6436dea
Merge branch 'mlcommons:master' into docs
arjunsuresh May 13, 2024
c843a08
Added docker setup command for CM, extra run options
arjunsuresh May 17, 2024
664a37a
Added support for docker options in the docs
arjunsuresh May 17, 2024
a97fc5f
Added --quiet to the CM run_cmds in docs
arjunsuresh May 18, 2024
b150d6a
Merge branch 'master' into master
arjunsuresh May 21, 2024
18ff1a8
Fix the test query count for cm commands
arjunsuresh May 21, 2024
2e3dd93
Merge branch 'mlcommons:master' into master
arjunsuresh May 22, 2024
204dbbf
Merge branch 'mlcommons:master' into master
arjunsuresh May 28, 2024
ac6d20a
Support ctuning-cpp implementation
arjunsuresh May 31, 2024
6b2264b
Added commands for mobilenet models
arjunsuresh May 31, 2024
9b896a6
Docs cleanup
arjunsuresh May 31, 2024
ce5a0b0
Docs cleanup
arjunsuresh May 31, 2024
9ff02d7
Merge branch 'master' into master
arjunsuresh Jun 4, 2024
d58ba74
Merge branch 'mlcommons:master' into master
arjunsuresh Jun 5, 2024
69661bb
Fix merge conflicts
arjunsuresh Jun 5, 2024
1f3eacd
Added separate files for dataset and models in the docs
arjunsuresh Jun 5, 2024
63fcc60
Remove redundant tab in the docs
arjunsuresh Jun 5, 2024
5e49864
Fixes some WIP models in the docs
arjunsuresh Jun 5, 2024
a33c2a7
Use the official docs page for CM installation
arjunsuresh Jun 5, 2024
16a8009
Fix the deadlink in docs
arjunsuresh Jun 5, 2024
6ad8a0e
Fix indendation issue in docs
arjunsuresh Jun 6, 2024
cf0ca4f
Added dockerinfo for nvidia implementation
arjunsuresh Jun 6, 2024
4901320
Added run options for gptj
arjunsuresh Jun 6, 2024
c007322
Added execution environment tabs
anandhu-eng Jun 6, 2024
fbbc894
Merge pull request #5 from anandhu-eng/docs
arjunsuresh Jun 6, 2024
1fba83e
Cleanup of the docs
arjunsuresh Jun 6, 2024
ee35e73
Cleanup of the docs
arjunsuresh Jun 6, 2024
2d47e06
Reordered the sections of the docs page
arjunsuresh Jun 7, 2024
e43bb87
Removed an unnecessary heading in the docs
arjunsuresh Jun 7, 2024
87471ee
Fixes the commands for datacenter
arjunsuresh Jun 7, 2024
f40bc9c
Fix the build --sdist for loadgen
arjunsuresh Jun 16, 2024
36af6b4
Merge branch 'mlcommons:master' into docs
arjunsuresh Jun 16, 2024
9dc997f
Fixes #1761, llama2 and mixtral runtime error on CPU systems
arjunsuresh Jul 2, 2024
165f5f0
Merge branch 'master' into master
arjunsuresh Jul 2, 2024
0718769
Added mixtral to the benchmark list, improved benchmark docs
arjunsuresh Jul 3, 2024
387013c
Merge branch 'mlcommons:master' into master
arjunsuresh Jul 8, 2024
bc19ba1
Update docs for MLPerf inference v4.1
arjunsuresh Jul 8, 2024
1f9bc3b
Update docs for MLPerf inference v4.1
arjunsuresh Jul 8, 2024
8ff59e1
Fix typo
arjunsuresh Jul 8, 2024
7faa762
Gave direct link to implementation readmes
arjunsuresh Jul 9, 2024
b8573fe
Added tables detailing implementations
anandhu-eng Jul 9, 2024
d3cbc48
Merge pull request #6 from anandhu-eng/candd_readme_change
arjunsuresh Jul 9, 2024
35e3bbd
Update vision README.md, split the frameworks into separate rows
arjunsuresh Jul 9, 2024
3e7f86c
Update README.md
arjunsuresh Jul 9, 2024
d43165d
pointed links to specific frameworks
anandhu-eng Jul 9, 2024
1e3cc6c
pointed links to specific frameworks
anandhu-eng Jul 9, 2024
a35649e
Merge pull request #7 from anandhu-eng/docsUpdate
arjunsuresh Jul 9, 2024
291537e
Update Submission_Guidelines.md
arjunsuresh Jul 9, 2024
3ea811e
Update Submission_Guidelines.md
arjunsuresh Jul 9, 2024
ef51ae3
Update Submission_Guidelines.md
arjunsuresh Jul 9, 2024
4d36503
api support llama2
anandhu-eng Jul 16, 2024
641df17
Added request module and reduced max token len
anandhu-eng Jul 16, 2024
751b9fc
Merge branch 'mlcommons:master' into master
arjunsuresh Jul 16, 2024
6b1a6a9
Merge branch 'master' into llama2_api
arjunsuresh Jul 16, 2024
b1ef8f1
Merge pull request #8 from anandhu-eng/llama2_api
arjunsuresh Jul 16, 2024
99ee8b6
Fix for llama2 api server
arjunsuresh Jul 16, 2024
c736d33
Update SUT_API offline to work for OpenAI
mgoin Jul 16, 2024
280a294
Update SUT_API.py
mgoin Jul 16, 2024
4e4aff3
Merge pull request #9 from mgoin/patch-2
arjunsuresh Jul 16, 2024
8b4c88f
Minor fixes
arjunsuresh Jul 16, 2024
8464902
Fix json import in SUT_API.py
arjunsuresh Jul 16, 2024
b00755d
Fix llama2 token length
arjunsuresh Jul 16, 2024
802374b
Added model name verification with server
anandhu-eng Jul 17, 2024
44ae1d9
clean temp files
anandhu-eng Jul 17, 2024
fe3644e
support num_workers in LLAMA2 SUTs
arjunsuresh Jul 17, 2024
1ef8072
Remove batching from Offline SUT_API.py
mgoin Jul 17, 2024
c0dc52e
Update SUT_API.py
mgoin Jul 17, 2024
ce2e686
Merge pull request #11 from mgoin/patch-3
arjunsuresh Jul 17, 2024
7517a90
Minor fixes for llama2 API
arjunsuresh Jul 17, 2024
d3db567
Fix for llama2 API
arjunsuresh Jul 17, 2024
7607097
Merge branch 'mlcommons:master' into master
arjunsuresh Jul 17, 2024
93b5d64
Merge pull request #10 from anandhu-eng/vllm_enhancement
arjunsuresh Jul 17, 2024
b9ba3d7
Merge branch 'mlcommons:master' into master
arjunsuresh Jul 22, 2024
6d9f638
Merge branch 'mlcommons:master' into master
arjunsuresh Aug 1, 2024
bd60060
removed table of contents
anandhu-eng Aug 12, 2024
32c4702
enabled llama2-nvidia + vllm-NM : WIP
anandhu-eng Aug 12, 2024
cd823cc
enabled dlrm for intel
anandhu-eng Aug 12, 2024
0c94ea9
lower cased implementation
anandhu-eng Aug 12, 2024
5b0df15
added raw data input
anandhu-eng Aug 12, 2024
f89295c
corrected data download commands
anandhu-eng Aug 12, 2024
8e1eb75
renamed filename
anandhu-eng Aug 12, 2024
087dad9
changes for bert and vllm
anandhu-eng Aug 13, 2024
c1032c2
documentation to work on custom repo and branch
anandhu-eng Aug 13, 2024
e8cb2a8
benchmark index page update
anandhu-eng Aug 13, 2024
7e37072
enabled sdxl for nvidia and intel
anandhu-eng Aug 13, 2024
73ce4fd
Merge branch 'mlcommons:master' into docs
arjunsuresh Aug 13, 2024
0f816ee
Merge branch 'master' into cm_readme_inference_update
arjunsuresh Aug 13, 2024
509c2c5
Merge pull request #12 from anandhu-eng/cm_readme_inference_update
arjunsuresh Aug 13, 2024
98b945c
Merge pull request #13 from GATEOverflow/master
arjunsuresh Aug 13, 2024
283b39c
updated vllm server run cmd
anandhu-eng Aug 13, 2024
2c9b859
Merge changes from master branch of https://github.com/GATEOverflow/i…
anandhu-eng Aug 13, 2024
4f5cbcd
benchmark page information addition
anandhu-eng Aug 14, 2024
8e71518
fix indendation issue
anandhu-eng Aug 14, 2024
e9dcf17
Added submission categories
anandhu-eng Aug 14, 2024
4f56494
update submission page - generate submission with or w/o using CM for…
anandhu-eng Aug 14, 2024
f1135ea
Updated kits dataset documentation
anandhu-eng Aug 16, 2024
cb71cd1
Updated model parameters
anandhu-eng Aug 16, 2024
2016369
Merge branch 'mlcommons:master' into master
arjunsuresh Aug 16, 2024
608ad33
Merge branch 'master' into cm_readme_inference_update
arjunsuresh Aug 16, 2024
1579967
updation of information
anandhu-eng Aug 19, 2024
5ba36a9
updated non cm based benchmark
anandhu-eng Aug 19, 2024
4805612
Merge changes from GateOverflow
anandhu-eng Aug 19, 2024
38f8067
Merge pull request #14 from anandhu-eng/cm_readme_inference_update
arjunsuresh Aug 19, 2024
a461646
added info about hf password
anandhu-eng Aug 20, 2024
dd47e45
added links to model and access tokens
anandhu-eng Aug 20, 2024
a1d66d4
Updated reference results structuree tree
anandhu-eng Aug 20, 2024
c5ae6ed
submission docs cleanup
anandhu-eng Aug 20, 2024
9dc81e8
Merge branch 'master' into cm_readme_inference_update
arjunsuresh Aug 20, 2024
66a9f10
Merge pull request #15 from anandhu-eng/cm_readme_inference_update
arjunsuresh Aug 20, 2024
4940585
Merge branch 'master' into docs
arjunsuresh Aug 20, 2024
f4ba37d
Merge branch 'mlcommons:master' into master
arjunsuresh Aug 22, 2024
bc80f65
Some cleanups for benchmark info
arjunsuresh Aug 22, 2024
65e63db
Some cleanups for benchmark info
arjunsuresh Aug 22, 2024
ba9820d
Some cleanups for benchmark info
arjunsuresh Aug 22, 2024
9ea1d14
added generic stubs deepsparse
anandhu-eng Aug 22, 2024
5b6eb52
Merge branch 'master' into cm_readme_inference_update
anandhu-eng Aug 22, 2024
63888bc
Some cleanups for benchmark info
arjunsuresh Aug 22, 2024
b956c6d
Some cleanups for benchmark info
arjunsuresh Aug 22, 2024
2c49334
Some cleanups for benchmark info
arjunsuresh Aug 22, 2024
13db0f8
Some cleanups for benchmark info (FID and CLIP data added)
arjunsuresh Aug 22, 2024
4eefc94
typo fix for bert deepsparse framework
anandhu-eng Aug 23, 2024
f0dbe10
Merge branch 'master' into cm_readme_inference_update
arjunsuresh Aug 23, 2024
6017bcc
Merge pull request #16 from anandhu-eng/cm_readme_inference_update
arjunsuresh Aug 23, 2024
e6abadd
added min system requirements for models
anandhu-eng Aug 23, 2024
8db76b4
Merge branch 'master' into cm_readme_inference_update
anandhu-eng Aug 23, 2024
37674fa
Merge pull request #17 from anandhu-eng/cm_readme_inference_update
arjunsuresh Aug 23, 2024
d994a86
fixed code version
anandhu-eng Sep 3, 2024
fd8945c
changes for displaying reference and intel implementation tip
anandhu-eng Sep 3, 2024
8815065
added reference to installation page
anandhu-eng Sep 3, 2024
5c73d16
Merge pull request #18 from anandhu-eng/cm_readme_inference_update
arjunsuresh Sep 3, 2024
8085e8b
updated neural magic documentation
anandhu-eng Sep 3, 2024
d078534
Merge branch 'master' into cm_readme_inference_update
arjunsuresh Sep 3, 2024
edbaf90
Merge pull request #19 from anandhu-eng/cm_readme_inference_update
arjunsuresh Sep 3, 2024
99285f6
Merge pull request #20 from GATEOverflow/docs
arjunsuresh Sep 3, 2024
9dbd46f
Merge pull request #21 from GATEOverflow/master
arjunsuresh Sep 3, 2024
32cdf40
Merge branch 'master' into docs
arjunsuresh Sep 3, 2024
9a27105
Added links to the install page, redirect benchmarks page
arjunsuresh Sep 4, 2024
294f85e
Merge branch 'mlcommons:master' into docs
arjunsuresh Sep 5, 2024
069c2dd
Merge branch 'master' into docs
arjunsuresh Sep 5, 2024
b30e51a
Merge pull request #22 from GATEOverflow/docs
arjunsuresh Sep 5, 2024
8d76337
added tips about batch size and dataset for nvidia llama2
anandhu-eng Sep 10, 2024
4ac509f
Merge branch 'cm_readme_inference_update' of https://github.com/anand…
anandhu-eng Sep 10, 2024
fd58737
Merge branch 'master' into cm_readme_inference_update
anandhu-eng Sep 10, 2024
e199221
fix conditions logic
anandhu-eng Sep 10, 2024
b144936
Merge branch 'cm_readme_inference_update' of https://github.com/anand…
anandhu-eng Sep 10, 2024
439b150
modified tips and additional run cmds
anandhu-eng Sep 10, 2024
b8188c2
sentence corrections
anandhu-eng Sep 10, 2024
3369b3c
Merge pull request #23 from anandhu-eng/cm_readme_inference_update
arjunsuresh Sep 10, 2024
f0b9e7f
Minor fix for the documentation
arjunsuresh Sep 10, 2024
aba2ce8
fixed bug in deepsparse generic model stubs + styling
anandhu-eng Sep 17, 2024
67bf51a
Merge branch 'master' into cm_readme_inference_update
anandhu-eng Sep 17, 2024
33cad44
added more information to stubs
anandhu-eng Sep 17, 2024
8cea28a
Added SCC24 readme, support reproducibility in the docs
arjunsuresh Sep 18, 2024
ae8f9e6
Made clear the custom CM repo URL format
arjunsuresh Sep 18, 2024
a5c1627
Support conditional implementation, setup and run tips
arjunsuresh Sep 18, 2024
3e24bb9
Support rocm for sdxl
arjunsuresh Sep 19, 2024
8d6392d
Fix _short tag support
arjunsuresh Sep 19, 2024
0511c95
Fix install URL
arjunsuresh Sep 19, 2024
e8b2adc
Expose bfloat16 and float16 options for sdxl
arjunsuresh Sep 19, 2024
d7080cd
Expose download model to host option for sdxl
arjunsuresh Sep 19, 2024
c588fa4
Merge branch 'master' into cm_readme_inference_update
anandhu-eng Sep 20, 2024
c454ac0
Merge pull request #24 from anandhu-eng/cm_readme_inference_update
arjunsuresh Sep 20, 2024
9009382
IndySCC24 documentation added
arjunsuresh Sep 20, 2024
00c650f
Improve the SCC24 docs
arjunsuresh Sep 20, 2024
4c92e2a
Improve the support of short variation
arjunsuresh Sep 20, 2024
60d3a8a
Improved the indyscc24 documentation
arjunsuresh Sep 20, 2024
b2f95b2
Updated scc run commands
anandhu-eng Sep 23, 2024
84ba650
removed test_query_count option for scc
anandhu-eng Sep 23, 2024
213c605
Merge pull request #25 from anandhu-eng/scc
arjunsuresh Sep 23, 2024
6c23816
Remove scc24 in the main docs
arjunsuresh Sep 23, 2024
469b091
Remove scc24 in the main docs
arjunsuresh Sep 23, 2024
21d16ed
Fix docs: indendation issue on the submission page
arjunsuresh Sep 23, 2024
5d4a302
generalised code for skipping test query count
anandhu-eng Sep 24, 2024
fb152b6
Merge pull request #26 from anandhu-eng/branch_from+go
arjunsuresh Sep 24, 2024
21e7259
Fixes for SCC24 docs
arjunsuresh Sep 24, 2024
93649dd
Fix scenario text in main.py
arjunsuresh Sep 24, 2024
0cc5d7b
Fix links for scc24
arjunsuresh Sep 24, 2024
70f9a81
Fix links for scc24
arjunsuresh Sep 24, 2024
6f56438
Improve the general docs
arjunsuresh Sep 24, 2024
a46ebee
Fix links for scc24
arjunsuresh Sep 24, 2024
913ffd4
Use float16 in scc24 doc
arjunsuresh Sep 24, 2024
b21cf39
Improve scc24 docs
arjunsuresh Sep 24, 2024
2271866
Improve scc24 docs
arjunsuresh Sep 24, 2024
3c072e0
Use float16 in scc24 doc
arjunsuresh Sep 24, 2024
7b776b7
fixed command bug
anandhu-eng Sep 24, 2024
7b62b53
Merge pull request #27 from anandhu-eng/bugfix
arjunsuresh Sep 24, 2024
a2f6125
Merge branch 'master' into master
arjunsuresh Sep 24, 2024
594ab62
Merge branch 'master' into master
arjunsuresh Sep 24, 2024
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
2 changes: 2 additions & 0 deletions docs/benchmarks/image_classification/mobilenets.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,6 +5,8 @@ hide:

# Image Classification using Mobilenet models

Install CM following the [installation page](site:install).

Mobilenet models are not official MLPerf models and so cannot be used for a Closed division MLPerf inference submission. But since they can be run with Imagenet dataset, we are allowed to use them for Open division submission. Only CPU runs are supported now.

## TFLite Backend
Expand Down
2 changes: 2 additions & 0 deletions docs/benchmarks/image_classification/resnet50.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,10 @@ hide:
- toc
---


# Image Classification using ResNet50


=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

Expand Down
1 change: 0 additions & 1 deletion docs/benchmarks/language/gpt-j.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@ hide:

# Text Summarization using GPT-J


=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

Expand Down
3 changes: 1 addition & 2 deletions docs/benchmarks/language/llama2-70b.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@ hide:

# Text Summarization using LLAMA2-70b


=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

Expand All @@ -25,4 +24,4 @@ hide:

{{ mlperf_inference_implementation_readme (4, "llama2-70b-99", "neuralmagic") }}

{{ mlperf_inference_implementation_readme (4, "llama2-70b-99.9", "neuralmagic") }}
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99.9", "neuralmagic") }}
4 changes: 3 additions & 1 deletion docs/benchmarks/language/mixtral-8x7b.md
Original file line number Diff line number Diff line change
Expand Up @@ -3,7 +3,9 @@ hide:
- toc
---

# Question Answering, Math, and Code Generation using Mixtral-8x7B

=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "mixtral-8x7b", "reference") }}
{{ mlperf_inference_implementation_readme (4, "mixtral-8x7b", "reference") }}
48 changes: 48 additions & 0 deletions docs/benchmarks/language/reproducibility/indyscc24-bert.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,48 @@
---
hide:
- toc
---

# Question and Answering using Bert Large for IndySCC 2024

## Introduction

This guide is designed for the [IndySCC 2024](https://sc24.supercomputing.org/students/indyscc/) to walk participants through running and optimizing the [MLPerf Inference Benchmark](https://arxiv.org/abs/1911.02549) using [Bert Large](https://github.com/mlcommons/inference/tree/master/language/bert#supported-models) across various software and hardware configurations. The goal is to maximize system throughput (measured in samples per second) without compromising accuracy.

For a valid MLPerf inference submission, two types of runs are required: a performance run and an accuracy run. In this competition, we focus on the `Offline` scenario, where throughput is the key metric—higher values are better. The official MLPerf inference benchmark for Bert Large requires processing a minimum of 10833 samples in both performance and accuracy modes using the Squad v1.1 dataset. Setting up for Nvidia GPUs may take 2-3 hours but can be done offline. Your final output will be a tarball (`mlperf_submission.tar.gz`) containing MLPerf-compatible results, which you will submit to the SCC organizers for scoring.

## Scoring

In the SCC, your first objective will be to run a reference (unoptimized) Python implementation or a vendor-provided version (such as Nvidia's) of the MLPerf inference benchmark to secure a baseline score.

Once the initial run is successful, you'll have the opportunity to optimize the benchmark further by maximizing system utilization, applying quantization techniques, adjusting ML frameworks, experimenting with batch sizes, and more, all of which can earn you additional points.

Since vendor implementations of the MLPerf inference benchmark vary and are often limited to single-node benchmarking, teams will compete within their respective hardware categories (e.g., Nvidia GPUs, AMD GPUs). Points will be awarded based on the throughput achieved on your system.


!!! info
Both MLPerf and CM automation are evolving projects.
If you encounter issues or have questions, please submit them [here](https://github.com/mlcommons/cm4mlops/issues)

## Artifacts to submit to the SCC committee

You will need to submit the following files:

* `mlperf_submission_short.tar.gz` - automatically generated file with validated MLPerf results.
* `mlperf_submission_short_summary.json` - automatically generated summary of MLPerf results.
* `mlperf_submission_short.run` - CM commands to run MLPerf BERT inference benchmark saved to this file.
* `mlperf_submission_short.tstamps` - execution timestamps before and after CM command saved to this file.
* `mlperf_submission_short.md` - description of your platform and some highlights of the MLPerf benchmark execution.



=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "bert-99", "reference", extra_variation_tags=",_short", scenarios=["Offline"],categories=["Edge"], setup_tips=False) }}

=== "Nvidia"
## Nvidia MLPerf Implementation
{{ mlperf_inference_implementation_readme (4, "bert-99", "nvidia", extra_variation_tags=",_short", scenarios=["Offline"],categories=["Edge"], setup_tips=False, implementation_tips=False) }}


1 change: 0 additions & 1 deletion docs/benchmarks/medical_imaging/3d-unet.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,7 +5,6 @@ hide:

# Medical Imaging using 3d-unet (KiTS 2019 kidney tumor segmentation task)


=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

Expand Down
4 changes: 1 addition & 3 deletions docs/benchmarks/recommendation/dlrm-v2.md
Original file line number Diff line number Diff line change
Expand Up @@ -5,8 +5,6 @@ hide:

# Recommendation using DLRM v2


## Benchmark Implementations
=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

Expand All @@ -26,4 +24,4 @@ hide:

{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99", "intel") }}

{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "intel") }}
{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "intel") }}
96 changes: 96 additions & 0 deletions docs/benchmarks/text_to_image/reproducibility/scc24.md
Original file line number Diff line number Diff line change
@@ -0,0 +1,96 @@
---
hide:
- toc
---

# Text-to-Image with Stable Diffusion for Student Cluster Competition 2024

## Introduction

This guide is designed for the [Student Cluster Competition 2024](https://sc24.supercomputing.org/students/student-cluster-competition/) to walk participants through running and optimizing the [MLPerf Inference Benchmark](https://arxiv.org/abs/1911.02549) using [Stable Diffusion XL 1.0](https://github.com/mlcommons/inference/tree/master/text_to_image#supported-models) across various software and hardware configurations. The goal is to maximize system throughput (measured in samples per second) without compromising accuracy. Since the model performs poorly on CPUs, it is essential to run it on GPUs.

For a valid MLPerf inference submission, two types of runs are required: a performance run and an accuracy run. In this competition, we focus on the `Offline` scenario, where throughput is the key metric—higher values are better. The official MLPerf inference benchmark for Stable Diffusion XL requires processing a minimum of 5,000 samples in both performance and accuracy modes using the COCO 2014 dataset. However, for SCC, we have reduced this and we also have two variants. `scc-base` variant has dataset size reduced to 50 samples, making it possible to complete both performance and accuracy runs in approximately 5-10 minutes. `scc-main` variant has dataset size of 500 and running it will fetch extra points as compared to running just the base variant. Setting up for Nvidia GPUs may take 2-3 hours but can be done offline. Your final output will be a tarball (`mlperf_submission.tar.gz`) containing MLPerf-compatible results, which you will submit to the SCC organizers for scoring.

## Scoring

In the SCC, your first objective will be to run `scc-base` variant for reference (unoptimized) Python implementation or a vendor-provided version (such as Nvidia's) of the MLPerf inference benchmark to secure a baseline score.

Once the initial run is successful, you'll have the opportunity to optimize the benchmark further by maximizing system utilization, applying quantization techniques, adjusting ML frameworks, experimenting with batch sizes, and more, all of which can earn you additional points.

Since vendor implementations of the MLPerf inference benchmark vary and are often limited to single-node benchmarking, teams will compete within their respective hardware categories (e.g., Nvidia GPUs, AMD GPUs). Points will be awarded based on the throughput achieved on your system.

Additionally, significant bonus points will be awarded if your team enhances an existing implementation, adds support for new hardware (such as an unsupported GPU), enables multi-node execution, or adds/extends scripts to [cm4mlops repository](https://github.com/mlcommons/cm4mlops/tree/main/script) supporting new devices, frameworks, implementations etc. All improvements must be made publicly available under the Apache 2.0 license and submitted alongside your results to the SCC committee to earn these bonus points, contributing to the MLPerf community.


!!! info
Both MLPerf and CM automation are evolving projects.
If you encounter issues or have questions, please submit them [here](https://github.com/mlcommons/cm4mlops/issues)

## Artifacts to submit to the SCC committee

You will need to submit the following files:

* `mlperf_submission.run` - CM commands to run MLPerf inference benchmark saved to this file.
* `mlperf_submission.md` - description of your platform and some highlights of the MLPerf benchmark execution.
* `<Team Name>` under which results are pushed to the github repository.


## SCC interview

You are encouraged to highlight and explain the obtained MLPerf inference throughput on your system
and describe any improvements and extensions to this benchmark (such as adding new hardware backend
or supporting multi-node execution) useful for the community and [MLCommons](https://mlcommons.org).

## Run Commands

=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

{{ mlperf_inference_implementation_readme (4, "sdxl", "reference", extra_variation_tags=",_short,_scc24-base", devices=["ROCm", "CUDA"],scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, skip_test_query_count=True, extra_input_string="--precision=float16") }}

=== "Nvidia"
## Nvidia MLPerf Implementation
{{ mlperf_inference_implementation_readme (4, "sdxl", "nvidia", extra_variation_tags=",_short,_scc24-base", scenarios=["Offline"],categories=["Datacenter"], setup_tips=False, implementation_tips=False, skip_test_query_count=True) }}

!!! info
Once the above run is successful, you can change `_scc24-base` to `_scc24-main` to run the main variant.

## Submission Commands

### Generate actual submission tree

```bash
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=open \
--category=datacenter \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--run_style=test \
--adr.submission-checker.tags=_short-run \
--quiet \
--submitter=<Team Name>
```

* Use `--hw_name="My system name"` to give a meaningful system name.


### Push Results to GitHub

Fork the repository URL at [https://github.com/gateoverflow/cm4mlperf-inference](https://github.com/gateoverflow/cm4mlperf-inference).

Run the following command after **replacing `--repo_url` with your GitHub fork URL**.

```bash
cm run script --tags=push,github,mlperf,inference,submission \
--repo_url=https://github.com/gateoverflow/cm4mlperf-inference \
--repo_branch=mlperf-inference-results-scc24 \
--commit_message="Results on system <HW Name>" \
--quiet
```

Once uploaded give a Pull Request to the origin repository. Github action will be running there and once
finished you can see your submitted results at [https://gateoverflow.github.io/cm4mlperf-inference](https://gateoverflow.github.io/cm4mlperf-inference).
26 changes: 13 additions & 13 deletions docs/install/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -8,24 +8,24 @@ We use MLCommons CM Automation framework to run MLPerf inference benchmarks.

CM needs `git`, `python3-pip` and `python3-venv` installed on your system. If any of these are absent, please follow the [official CM installation page](https://docs.mlcommons.org/ck/install) to install them. Once the dependencies are installed, do the following

## Activate a VENV for CM
## Activate a Virtual ENV for CM
This step is not mandatory as CM can use separate virtual environment for MLPerf inference. But the latest `pip` install requires this or else will need the `--break-system-packages` flag while installing `cm4mlops`.

```bash
python3 -m venv cm
source cm/bin/activate
```

## Install CM and pulls any needed repositories

```bash
pip install cm4mlops
```

## To work on custom GitHub repo and branch

```bash
pip install cmind && cm init --quiet --repo=mlcommons@cm4mlops --branch=mlperf-inference
```

Here, repo is in the format `githubUsername@githubRepo`.
=== "Use the default fork of CM MLOps repository"
```bash
pip install cm4mlops
```

=== "Use custom fork/branch of the CM MLOps repository"
```bash
pip install cmind && cm init --quiet --repo=mlcommons@cm4mlops --branch=mlperf-inference
```
Here, `repo` is in the format `githubUsername@githubRepo`.

Now, you are ready to use the `cm` commands to run MLPerf inference as given in the [benchmarks](../index.md) page
2 changes: 2 additions & 0 deletions docs/requirements.txt
Original file line number Diff line number Diff line change
Expand Up @@ -2,3 +2,5 @@ mkdocs-material
swagger-markdown
mkdocs-macros-plugin
ruamel.yaml
mkdocs-redirects
mkdocs-site-urls
90 changes: 45 additions & 45 deletions docs/submission/index.md
Original file line number Diff line number Diff line change
Expand Up @@ -60,63 +60,63 @@ Once all the results across all the models are ready you can use the following c
=== "Closed Edge"
### Closed Edge Submission
```bash
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=closed \
--category=edge \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=closed \
--category=edge \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
```

=== "Closed Datacenter"
### Closed Datacenter Submission
```bash
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=closed \
--category=datacenter \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=closed \
--category=datacenter \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
```
=== "Open Edge"
### Open Edge Submission
```bash
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=open \
--category=edge \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=open \
--category=edge \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
```
=== "Open Datacenter"
### Closed Datacenter Submission
```bash
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=open \
--category=datacenter \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
cm run script --tags=generate,inference,submission \
--clean \
--preprocess_submission=yes \
--run-checker \
--submitter=MLCommons \
--tar=yes \
--env.CM_TAR_OUTFILE=submission.tar.gz \
--division=open \
--category=datacenter \
--env.CM_DETERMINE_MEMORY_CONFIGURATION=yes \
--quiet
```

* Use `--hw_name="My system name"` to give a meaningful system name. Examples can be seen [here](https://github.com/mlcommons/inference_results_v3.0/tree/main/open/cTuning/systems)
Expand All @@ -134,7 +134,7 @@ If you are collecting results across multiple systems you can generate different
Run the following command after **replacing `--repo_url` with your GitHub repository URL**.

```bash
cm run script --tags=push,github,mlperf,inference,submission \
cm run script --tags=push,github,mlperf,inference,submission \
--repo_url=https://github.com/GATEOverflow/mlperf_inference_submissions_v4.1 \
--commit_message="Results on <HW name> added by <Name>" \
--quiet
Expand Down
Loading
Loading