Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Dev #43

Merged
merged 18 commits into from
May 4, 2024
Merged

Dev #43

Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
5 changes: 5 additions & 0 deletions automation/script/module_misc.py
Original file line number Diff line number Diff line change
Expand Up @@ -1889,6 +1889,8 @@ def docker(i):

all_gpus = i.get('docker_all_gpus', docker_settings.get('all_gpus'))

num_gpus = i.get('docker_num_gpus', docker_settings.get('num_gpus'))

device = i.get('docker_device', docker_settings.get('device'))

r = check_gh_token(i, docker_settings, quiet)
Expand Down Expand Up @@ -1983,6 +1985,9 @@ def docker(i):
if all_gpus:
cm_docker_input['all_gpus'] = True

if num_gpus:
cm_docker_input['num_gpus'] = str(num_gpus)

if device:
cm_docker_input['device'] = device

Expand Down
7 changes: 6 additions & 1 deletion automation/utils/module_cfg.py
Original file line number Diff line number Diff line change
Expand Up @@ -282,14 +282,19 @@ def select_cfg(i):
meta = r['meta']
ss['meta'] = meta

selector = sorted(selector, key = lambda x: x['meta'].get('name',''))
s = 0
for ss in selector:
alias = ss['alias']
name = meta.get('name','')
name = ss['meta'].get('name','')

x = name
if x!='': x+=' '
x += '('+alias+')'

print ('{}) {}'.format(s, x))

s+=1

print ('')
select = input ('Enter configuration number of press Enter for 0: ')
Expand Down
Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
uid: 854e65fb31584d63

name: "Nvidia Ubuntu 20.04 CUDA 11.8 cuDNN 8.6.0 PyTorch 1.13.0"

input:
docker_base_image: 'nvcr.io/nvidia/pytorch:22.10-py3'
docker_os: ubuntu
docker_os_version: '20.04'

Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
uid: e0e7167139a74e36

name: "Nvidia Ubuntu 22.04 CUDA 12.1 cuDNN 8.9.1 PyTorch 2.0.0"

input:
docker_base_image: 'nvcr.io/nvidia/pytorch:23.05-py3'
docker_os: ubuntu
docker_os_version: '22.04'

Original file line number Diff line number Diff line change
@@ -0,0 +1,9 @@
uid: 49fc51f2999b4545

name: "Nvidia Ubuntu 22.04 CUDA 12.4 cuDNN 9.0.0 PyTorch 2.3.0"

input:
docker_base_image: 'nvcr.io/nvidia/pytorch:24.03-py3'
docker_os: ubuntu
docker_os_version: '22.04'

2 changes: 0 additions & 2 deletions script/app-image-classification-onnx-py/_cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -117,5 +117,3 @@ docker:
- env.CM_IMAGE
- output
- j
pre_run_cmds:
- echo \"CM pre run commands\"
2 changes: 1 addition & 1 deletion script/app-loadgen-generic-python/README-extra.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,7 +18,7 @@ and pull CM repository with portable automation scripts to benchmark ML Systems:

```bash
pip install cmind
cm pull repo mlcommons@ck
cm pull repo mlcommons@cm4mlops --checkout=dev
```

### Clean CM cache
Expand Down
4 changes: 2 additions & 2 deletions script/app-mlperf-inference-nvidia/README-about.md
Original file line number Diff line number Diff line change
Expand Up @@ -52,13 +52,13 @@ Assuming all the downloaded files are to the user home directory please do the f
--cudnn_tar_file_path=$HOME/cudnn-linux-x86_64-8.9.2.26_cuda11-archive.tar.xz \
--imagenet_path=$HOME/imagenet-2012-val \
--scratch_path=$HOME/mlperf_scratch \
--docker_cm_repo=mlcommons@ck \
--docker_cm_repo=mlcommons@cm4mlops \
--results_dir=$HOME/results_dir \
--submission_dir=$HOME/submission_dir \
--adr.compiler.tags=gcc
```
* Use `--docker_cache=no` to turn off docker caching
* Use `--docker_run_cmd_prefix="cm pull repo mlcommons@ck"` to update the CK repository when docker caching is used
* Use `--docker_run_cmd_prefix="cm pull repo mlcommons@cm4mlops --checkout=dev"` to update the CK repository when docker caching is used
* Use `--custom_system=no` if you are using a similar system to the [Nvidia submission systems for MLPerf inference 3.0](https://github.com/mlcommons/inference_results_v3.0/tree/main/closed/NVIDIA/systems).

6. At the end of the build you'll get the following prompt unless you have chosen `--custom_system=no`. Please give a system name and say yes to generating the configuration files
Expand Down
4 changes: 2 additions & 2 deletions script/app-mlperf-inference-nvidia/README.md
Original file line number Diff line number Diff line change
Expand Up @@ -65,13 +65,13 @@ Assuming all the downloaded files are to the user home directory please do the f
--cudnn_tar_file_path=$HOME/cudnn-linux-x86_64-8.9.2.26_cuda11-archive.tar.xz \
--imagenet_path=$HOME/imagenet-2012-val \
--scratch_path=$HOME/mlperf_scratch \
--docker_cm_repo=mlcommons@ck \
--docker_cm_repo=mlcommons@cm4mlops \
--results_dir=$HOME/results_dir \
--submission_dir=$HOME/submission_dir \
--adr.compiler.tags=gcc
```
* Use `--docker_cache=no` to turn off docker caching
* Use `--docker_run_cmd_prefix="cm pull repo mlcommons@ck"` to update the CK repository when docker caching is used
* Use `--docker_run_cmd_prefix="cm pull repo mlcommons@cm4mlops"` to update the CK repository when docker caching is used
* Use `--custom_system=no` if you are using a similar system to the [Nvidia submission systems for MLPerf inference 3.0](https://github.com/mlcommons/inference_results_v3.0/tree/main/closed/NVIDIA/systems).

6. At the end of the build you'll get the following prompt unless you have chosen `--custom_system=no`. Please give a system name and say yes to generating the configuration files
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -12,7 +12,7 @@ image from the Community AMIs is the recommended OS image as it comes with the Q
```
sudo yum install -y python38-devel git
python3.8 -m pip install cmind
cm pull repo mlcommons@ck
cm pull repo mlcommons@cm4mlops
cm run script --tags=get,python --version_min=3.8.1
```

Expand Down
8 changes: 4 additions & 4 deletions script/app-mlperf-inference/README-extra.md
Original file line number Diff line number Diff line change
Expand Up @@ -18,16 +18,16 @@ source .profile
Next you need to install a CM repository with [cross-platform CM scripts](https://github.com/mlcommons/cm4mlops/tree/main/script) for ML Systems:

```bash
cm pull repo mlcommons@ck
cm pull repo mlcommons@cm4mlops --checkout=dev
```

Note that you can fork [this repository](https://github.com/mlcommons/ck) and use it instead of mlcommons@ck
Note that you can fork [this repository](https://github.com/mlcommons/cm4mlops) and use it instead of mlcommons@cm4mlops
to add CM scripts for your own public and private ML models, data sets, software and hardware.
In such case, just change mlcommons@ck to your own fork in the above command.
In such case, just change mlcommons@cm4mlops to your own fork in the above command.

You can find the location of this repository on your system as follows:
```bash
cm find repo mlcommons@ck
cm find repo mlcommons@cm4mlops
```

Now we suggest you to set up a virtual python via CM to avoid mixing up your native Python installation:
Expand Down
4 changes: 4 additions & 0 deletions script/app-mlperf-inference/_cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -244,6 +244,10 @@ variations:
deps:
- tags: get,mlperf,inference,nvidia,scratch,space
- tags: get,nvidia-docker
skip_if_env:
CM_SKIP_GET_NVIDIA_DOCKER:
- yes

mounts:
- "${{ CM_CUDNN_TAR_FILE_PATH }}:${{ CM_CUDNN_TAR_FILE_PATH }}"
- "${{ CM_TENSORRT_TAR_FILE_PATH }}:${{ CM_TENSORRT_TAR_FILE_PATH }}"
Expand Down
48 changes: 0 additions & 48 deletions script/build-docker-image/_cm.json

This file was deleted.

69 changes: 0 additions & 69 deletions script/build-dockerfile/_cm.json

This file was deleted.

2 changes: 1 addition & 1 deletion script/build-dockerfile/customize.py
Original file line number Diff line number Diff line change
Expand Up @@ -41,7 +41,7 @@ def preprocess(i):
continue
arg=env_
if env_ in default_env: #other inputs to be done later
arg=arg+"="+default_env[env_]
arg=arg+"="+str(default_env[env_])
#build_args.append(arg)
#input_args.append("--"+input_+"="+"$"+env_)

Expand Down
4 changes: 4 additions & 0 deletions script/build-mlperf-inference-server-nvidia/_cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -240,6 +240,10 @@ docker:
- tags: get,mlperf,inference,results,dir
- tags: get,mlperf,inference,submission,dir
- tags: get,nvidia-docker
skip_if_env:
CM_SKIP_GET_NVIDIA_DOCKER:
- yes

pre_run_cmds:
- cm pull repo
run_cmd_prefix: sudo apt remove -y cmake
Expand Down
2 changes: 1 addition & 1 deletion script/generate-mlperf-tiny-report/README-extra.md
Original file line number Diff line number Diff line change
Expand Up @@ -11,7 +11,7 @@ Install [MLCommons CM framework](https://github.com/mlcommons/ck/blob/master/doc

Pull the MLCommons CK repository with automation recipes for interoperable MLOps:
```bash
cm pull repo mlcommons@ck
cm pull repo mlcommons@cm4mlops --checkout=dev
```

Install repositories with raw MLPerf inference benchmark results:
Expand Down
36 changes: 0 additions & 36 deletions script/get-cuda-devices/_cm.json

This file was deleted.

Loading
Loading