diff --git a/docs/install/index.md b/docs/install/index.md index 195521c7e..1750d86e4 100644 --- a/docs/install/index.md +++ b/docs/install/index.md @@ -12,8 +12,8 @@ CM needs `git`, `python3-pip` and `python3-venv` installed on your system. If an This step is not mandatory as CM can use separate virtual environment for MLPerf inference. But the latest `pip` install requires this or else will need the `--break-system-packages` flag while installing `cm4mlops`. ```bash - python3 -m venv cm - source cm/bin/activate +python3 -m venv cm +source cm/bin/activate ``` ## Install CM and pulls any needed repositories diff --git a/main.py b/main.py index aa8dd769e..6a607cc10 100755 --- a/main.py +++ b/main.py @@ -140,6 +140,9 @@ def mlperf_inference_implementation_readme(spaces, model, implementation, *, imp # ref to cm installation content += f"{cur_space3}Please refer to the [installation page](site:inference/install/) to install CM for running the automated benchmark commands.\n\n" test_query_count=get_test_query_count(model, implementation, device.lower()) + if device.lower() == "cuda" and execution_env.lower() == "native": + content += f"\n{cur_space3}!!! tip\n\n" + content += f"{cur_space3} - It is advisable to use the commands in the Docker tab for CUDA. Run the below native command only if you are already on a CUDA setup with cuDNN and TensorRT installed.\n\n" if "99.9" not in model: #not showing docker command as it is already done for the 99% variant if implementation == "neuralmagic": @@ -442,7 +445,8 @@ def mlperf_inference_run_command(spaces, model, implementation, framework, categ if "short" in extra_variation_tags: full_ds_needed_tag = "" else: - full_ds_needed_tag = ",_full" + full_ds_needed_tag = "_full," + docker_setup_cmd = f"""\n {f_pre_space}```bash