From 9e43557adc90197e7c2beeeb2f1b083469ddf8d1 Mon Sep 17 00:00:00 2001 From: ANANDHU S <71482562+anandhu-eng@users.noreply.github.com> Date: Wed, 15 May 2024 19:29:43 +0530 Subject: [PATCH] cleaned temp files --- .../offline/performance/run_1/console.out | 16 -- .../tmp-run.sh | 254 ------------------ .../tmp-state.json | 207 -------------- 3 files changed, 477 deletions(-) delete mode 100644 script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1/console.out delete mode 100755 script/app-mlperf-inference-mlcommons-python/tmp-run.sh delete mode 100644 script/app-mlperf-inference-mlcommons-python/tmp-state.json diff --git a/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1/console.out b/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1/console.out deleted file mode 100644 index 8e52c63ef..000000000 --- a/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1/console.out +++ /dev/null @@ -1,16 +0,0 @@ -Loading BERT configs... -Loading PyTorch model... -Constructing SUT... -Finished constructing SUT. -Constructing QSL... -No cached features at 'eval_features.pickle'... converting from examples... -Creating tokenizer... -Reading examples... -Converting examples to features... -Caching features at 'eval_features.pickle'... -Finished constructing QSL. - * Serving Flask app 'network_SUT' - * Debug mode: off -WARNING: This is a development server. Do not use it in a production deployment. Use a production WSGI server instead. - * Running on http://127.0.0.1:8000 -Press CTRL+C to quit diff --git a/script/app-mlperf-inference-mlcommons-python/tmp-run.sh b/script/app-mlperf-inference-mlcommons-python/tmp-run.sh deleted file mode 100755 index 880f35242..000000000 --- a/script/app-mlperf-inference-mlcommons-python/tmp-run.sh +++ /dev/null @@ -1,254 +0,0 @@ -#!/bin/bash - -export LDFLAGS="-L/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/lib64 ${LDFLAGS}" -export CM_HOST_OS_DEFAULT_LIBRARY_PATH="/usr/local/lib/x86_64-linux-gnu:/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu:/usr/lib/x86_64-linux-gnu64:/usr/local/lib64:/lib64:/usr/lib64:/usr/local/lib:/lib:/usr/lib:/usr/x86_64-linux-gnu/lib64:/usr/x86_64-linux-gnu/lib:${CM_HOST_OS_DEFAULT_LIBRARY_PATH}" -export CPLUS_INCLUDE_PATH="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/include:/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/include:${CPLUS_INCLUDE_PATH}" -export C_INCLUDE_PATH="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/include:/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/include:${C_INCLUDE_PATH}" -export DYLD_FALLBACK_LIBRARY_PATH="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/lib64:/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/lib:${DYLD_FALLBACK_LIBRARY_PATH}" -export LD_LIBRARY_PATH="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/lib64:/lib/x86_64-linux-gnu:/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/lib:${LD_LIBRARY_PATH}" -export PYTHONPATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/vision/classification_and_detection/python:/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/tools/submission:/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/get-mlperf-inference-utils:/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/python:${PYTHONPATH}" -export CK_PROGRAM_TMP_DIR="None" -export CM_BOTO3_VERSION="1.34.105" -export CM_CUDA_CACHE_TAGS="version-12.4" -export CM_CUDA_FULL_TOOLKIT_INSTALL="yes" -export CM_CUDA_INSTALLED_PATH="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install" -export CM_CUDA_INSTALL_DRIVER="no" -export CM_CUDA_LINUX_FILENAME="cuda_12.4.1_550.54.15_linux.run" -export CM_CUDA_NEEDS_CUDNN="yes" -export CM_CUDA_PATH_BIN="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/bin" -export CM_CUDA_PATH_INCLUDE="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/include" -export CM_CUDA_PATH_INCLUDE_CUDNN="/usr/include" -export CM_CUDA_PATH_LIB="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/lib64" -export CM_CUDA_PATH_LIB_CUDNN="/lib/x86_64-linux-gnu" -export CM_CUDA_PATH_LIB_CUDNN_EXISTS="yes" -export CM_CUDA_RUN_FILE_PATH="/home/anandhu/CM/repos/local/cache/49f9bac43fd4420c/cuda_12.4.1_550.54.15_linux.run" -export CM_CUDA_VERSION="12.4" -export CM_CUDA_VERSION_STRING="cu124" -export CM_DATASET_PATH="/home/anandhu/CM/repos/local/cache/088af364281445f2" -export CM_DATASET_SQUAD_PATH="/home/anandhu/CM/repos/local/cache/088af364281445f2" -export CM_DATASET_SQUAD_VAL_PATH="/home/anandhu/CM/repos/local/cache/088af364281445f2/dev-v1.1.json" -export CM_DATASET_SQUAD_VOCAB_PATH="/home/anandhu/CM/repos/local/cache/48d205c056224895/vocab.txt" -export CM_ENABLE_NUMACTL="0" -export CM_ENABLE_PROFILING="0" -export CM_HOST_CPU_ARCHITECTURE="x86_64" -export CM_HOST_CPU_FAMILY="6" -export CM_HOST_CPU_L1D_CACHE_SIZE="896 KiB (24 instances)" -export CM_HOST_CPU_L1I_CACHE_SIZE="1.3 MiB (24 instances)" -export CM_HOST_CPU_L2_CACHE_SIZE="32 MiB (12 instances)" -export CM_HOST_CPU_L3_CACHE_SIZE="36 MiB (1 instance)" -export CM_HOST_CPU_MAX_MHZ="5800.0000" -export CM_HOST_CPU_MODEL_NAME="13th Gen Intel(R) Core(TM) i9-13900K" -export CM_HOST_CPU_NUMA_NODES="1" -export CM_HOST_CPU_ON_LINE_CPUS_LIST="0-31" -export CM_HOST_CPU_PHYSICAL_CORES_PER_SOCKET="24" -export CM_HOST_CPU_SOCKETS="1" -export CM_HOST_CPU_THREADS_PER_CORE="2" -export CM_HOST_CPU_TOTAL_CORES="32" -export CM_HOST_CPU_TOTAL_LOGICAL_CORES="32" -export CM_HOST_CPU_VENDOR_ID="GenuineIntel" -export CM_HOST_DISK_CAPACITY="6.4T" -export CM_HOST_MEMORY_CAPACITY="134G" -export CM_HOST_OS_BITS="64" -export CM_HOST_OS_FLAVOR="ubuntu" -export CM_HOST_OS_FLAVOR_LIKE="debian" -export CM_HOST_OS_GLIBC_VERSION="2.39" -export CM_HOST_OS_KERNEL_VERSION="6.8.0-31-generic" -export CM_HOST_OS_MACHINE="x86_64" -export CM_HOST_OS_PACKAGE_MANAGER="apt" -export CM_HOST_OS_PACKAGE_MANAGER_INSTALL_CMD="DEBIAN_FRONTEND=noninteractive apt-get install -y" -export CM_HOST_OS_PACKAGE_MANAGER_UPDATE_CMD="apt-get update -y" -export CM_HOST_OS_TYPE="linux" -export CM_HOST_OS_VERSION="24.04" -export CM_HOST_PLATFORM_FLAVOR="x86_64" -export CM_HOST_PYTHON_BITS="64" -export CM_HOST_SYSTEM_NAME="intel-spr-i9" -export CM_HW_NAME="intel_spr_i9" -export CM_IMAGENET_ACCURACY_DTYPE="float32" -export CM_LIBRISPEECH_ACCURACY_DTYPE="float32" -export CM_LOGS_DIR="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1" -export CM_MAX_EXAMPLES="1000" -export CM_MLPERF_ACCURACY_RESULTS_DIR="" -export CM_MLPERF_BACKEND="pytorch" -export CM_MLPERF_BACKEND_VERSION="2.3.0" -export CM_MLPERF_CONF="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/mlperf.conf" -export CM_MLPERF_DEVICE="gpu" -export CM_MLPERF_FIND_PERFORMANCE_MODE="yes" -export CM_MLPERF_IMPLEMENTATION="reference" -export CM_MLPERF_INFERENCE_3DUNET_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/vision/medical_imaging/3d-unet-kits19" -export CM_MLPERF_INFERENCE_BERT_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/language/bert" -export CM_MLPERF_INFERENCE_CLASSIFICATION_AND_DETECTION_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/vision/classification_and_detection" -export CM_MLPERF_INFERENCE_CONF_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/mlperf.conf" -export CM_MLPERF_INFERENCE_DLRM_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/recommendation/dlrm" -export CM_MLPERF_INFERENCE_DLRM_V2_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/recommendation/dlrm_v2" -export CM_MLPERF_INFERENCE_FINAL_RESULTS_DIR="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python/test_results" -export CM_MLPERF_INFERENCE_GPTJ_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/language/gpt-j" -export CM_MLPERF_INFERENCE_LOADGEN_INCLUDE_PATH="/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/include" -export CM_MLPERF_INFERENCE_LOADGEN_INSTALL_PATH="/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install" -export CM_MLPERF_INFERENCE_LOADGEN_LIBRARY_PATH="/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/lib" -export CM_MLPERF_INFERENCE_LOADGEN_PYTHON_PATH="/home/anandhu/CM/repos/local/cache/c2666880367f41e8/install/python" -export CM_MLPERF_INFERENCE_RNNT_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/speech_recognition/rnnt" -export CM_MLPERF_INFERENCE_SOURCE="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference" -export CM_MLPERF_INFERENCE_VISION_PATH="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/inference/vision" -export CM_MLPERF_LAST_RELEASE="v4.0" -export CM_MLPERF_LOADGEN_ALL_MODES="no" -export CM_MLPERF_LOADGEN_COMPLIANCE="no" -export CM_MLPERF_LOADGEN_EXTRA_OPTIONS=" --mlperf_conf '/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/mlperf.conf'" -export CM_MLPERF_LOADGEN_LOGS_DIR="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1" -export CM_MLPERF_LOADGEN_MODE="performance" -export CM_MLPERF_LOADGEN_MODES="['performance']" -export CM_MLPERF_LOADGEN_QPS_OPT="" -export CM_MLPERF_LOADGEN_QUERY_COUNT="1000" -export CM_MLPERF_LOADGEN_SCENARIO="Offline" -export CM_MLPERF_LOADGEN_SCENARIOS="['Offline']" -export CM_MLPERF_LOADGEN_TARGET_QPS="1.0" -export CM_MLPERF_MODEL="bert-99" -export CM_MLPERF_MODEL_EQUAL_ISSUE_MODE="no" -export CM_MLPERF_MODEL_PRECISION="float32" -export CM_MLPERF_MODEL_SKIP_BATCHING="True" -export CM_MLPERF_OUTPUT_DIR="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1" -export CM_MLPERF_PRINT_SUMMARY="no" -export CM_MLPERF_PYTHON="yes" -export CM_MLPERF_QUANTIZATION="False" -export CM_MLPERF_RANGING_USER_CONF="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/generate-mlperf-inference-user-conf/tmp/ranging_49eb9d34406549189294b08daa9edeb2.conf" -export CM_MLPERF_RESULT_PUSH_TO_GITHUB="False" -export CM_MLPERF_RUN_CMD="/usr/bin/python3 run.py --backend=pytorch --scenario=Offline --mlperf_conf '/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/mlperf.conf' --user_conf '/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/generate-mlperf-inference-user-conf/tmp/49eb9d34406549189294b08daa9edeb2.conf' --network sut" -export CM_MLPERF_RUN_STYLE="test" -export CM_MLPERF_SKIP_RUN="no" -export CM_MLPERF_SUBMISSION_GENERATION_STYLE="short" -export CM_MLPERF_SUT_NAME_IMPLEMENTATION_PREFIX="reference" -export CM_MLPERF_SUT_NAME_RUN_CONFIG_SUFFIX="" -export CM_MLPERF_SUT_NAME_RUN_CONFIG_SUFFIX1="network_sut" -export CM_MLPERF_TESTING_USER_CONF="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/generate-mlperf-inference-user-conf/tmp/49eb9d34406549189294b08daa9edeb2.conf" -export CM_MLPERF_USER_CONF="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/generate-mlperf-inference-user-conf/tmp/49eb9d34406549189294b08daa9edeb2.conf" -export CM_ML_MODEL="BERT" -export CM_ML_MODEL_BERT_LARGE_FP32_PATH="/home/anandhu/CM/repos/local/cache/b0f750a1b4b44b21/model.pytorch" -export CM_ML_MODEL_BERT_PACKED="no" -export CM_ML_MODEL_BERT_VOCAB_FILE_WITH_PATH="/home/anandhu/CM/repos/local/cache/48d205c056224895/vocab.txt" -export CM_ML_MODEL_DATASET="squad-1.1" -export CM_ML_MODEL_F1="90.874" -export CM_ML_MODEL_FILE="model.pytorch" -export CM_ML_MODEL_FILE_WITH_PATH="/home/anandhu/CM/repos/local/cache/b0f750a1b4b44b21/model.pytorch" -export CM_ML_MODEL_FRAMEWORK="pytorch" -export CM_ML_MODEL_INPUT_IDS_NAME="input_ids" -export CM_ML_MODEL_INPUT_MASK_NAME="input_mask" -export CM_ML_MODEL_INPUT_SEGMENTS_NAME="segment_ids" -export CM_ML_MODEL_MAX_SEQ_LENGTH="384" -export CM_ML_MODEL_NAME="MLPERF BERT Large on SQuAD v1.1" -export CM_ML_MODEL_OUTPUT_END_LOGITS_NAME="output_end_logits" -export CM_ML_MODEL_OUTPUT_START_LOGITS_NAME="output_start_logits" -export CM_ML_MODEL_PRECISION="fp32" -export CM_ML_MODEL_STARTING_WEIGHTS_FILENAME="https://armi.in/files/fp32/model.pytorch" -export CM_MODEL="bert-99" -export CM_NETWORK_LOADGEN="sut" -export CM_NUM_THREADS="32" -export CM_NVCC_BIN="nvcc" -export CM_NVCC_BIN_WITH_PATH="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install/bin/nvcc" -export CM_OPENIMAGES_ACCURACY_DTYPE="float32" -export CM_OUTPUT_FOLDER_NAME="test_results" -export CM_PYTHONLIB_ABSL_PY_CACHE_TAGS="version-2.1.0" -export CM_PYTHONLIB_BOTO3_CACHE_TAGS="version-1.34.105" -export CM_PYTHONLIB_FLASK_CACHE_TAGS="version-3.0.3" -export CM_PYTHONLIB_PIP_CACHE_TAGS="version-24.0" -export CM_PYTHONLIB_PSUTIL_CACHE_TAGS="version-5.9.8" -export CM_PYTHONLIB_PYDANTIC_CACHE_TAGS="version-1.10.9" -export CM_PYTHONLIB_SIX_CACHE_TAGS="version-1.16.0" -export CM_PYTHONLIB_TOKENIZATION_CACHE_TAGS="version-1.0.7" -export CM_PYTHONLIB_TORCHVISION_CACHE_TAGS="version-0.18.0" -export CM_PYTHONLIB_TORCH_CACHE_TAGS="version-2.3.0" -export CM_PYTHONLIB_TRANSFORMERS_CACHE_TAGS="version-4.40.2" -export CM_PYTHON_BIN="python3" -export CM_PYTHON_BIN_PATH="/usr/bin" -export CM_PYTHON_BIN_WITH_PATH="/usr/bin/python3" -export CM_PYTHON_CACHE_TAGS="version-3.12.3,non-virtual" -export CM_PYTHON_MAJOR_VERSION="3" -export CM_PYTHON_MINOR_VERSION="12" -export CM_PYTHON_PATCH_VERSION="3" -export CM_PYTHON_VERSION="3.12.3" -export CM_QUIET="yes" -export CM_RUN_CMD="/usr/bin/python3 run.py --backend=pytorch --scenario=Offline --mlperf_conf '/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/mlperf.conf' --user_conf '/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/generate-mlperf-inference-user-conf/tmp/49eb9d34406549189294b08daa9edeb2.conf' --network sut 2>&1 | tee /home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1/console.out" -export CM_RUN_DIR="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/language/bert" -export CM_SIX_VERSION="1.16.0" -export CM_SQUAD_ACCURACY_DTYPE="float32" -export CM_SUT_CONFIGS_PATH="" -export CM_SUT_DESC_CACHE="no" -export CM_SUT_META_EXISTS="yes" -export CM_SUT_NAME="intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut" -export CM_TEST_QUERY_COUNT="1000" -export CM_TMP_CURRENT_PATH="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python" -export CM_TMP_CURRENT_SCRIPT_PATH="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/benchmark-program" -export CM_TMP_CURRENT_SCRIPT_REPO_PATH="/home/anandhu/CM/repos/anandhu-eng@cm4mlops" -export CM_TMP_CURRENT_SCRIPT_REPO_PATH_WITH_PREFIX="/home/anandhu/CM/repos/anandhu-eng@cm4mlops" -export CM_TMP_CURRENT_SCRIPT_WORK_PATH="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python" -export CM_TMP_PIP_VERSION_STRING="" -export CM_TOKENIZATION_VERSION="1.0.7" -export CM_TORCHVISION_VERSION="0.18.0" -export CM_TORCHVISION_VERSION_EXTRA="CUDA" -export CM_TORCH_VERSION="2.3.0" -export CM_TORCH_VERSION_EXTRA="CUDA" -export CM_TRANSFORMERS_VERSION="4.40.2" -export CUDA_HOME="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install" -export CUDA_PATH="/home/anandhu/CM/repos/local/cache/464985797dc34b1e/install" -export DATASET_FILE="/home/anandhu/CM/repos/local/cache/088af364281445f2/dev-v1.1.json" -export DATA_DIR="/home/anandhu/CM/repos/local/cache/088af364281445f2" -export HOST_CPU_ARCHITECTURE="x86_64" -export HOST_CPU_FAMILY="6" -export HOST_CPU_L1D_CACHE_SIZE="896 KiB (24 instances)" -export HOST_CPU_L1I_CACHE_SIZE="1.3 MiB (24 instances)" -export HOST_CPU_L2_CACHE_SIZE="32 MiB (12 instances)" -export HOST_CPU_L3_CACHE_SIZE="36 MiB (1 instance)" -export HOST_CPU_MAX_MHZ="5800.0000" -export HOST_CPU_MODEL_NAME="13th Gen Intel(R) Core(TM) i9-13900K" -export HOST_CPU_NUMA_NODES="1" -export HOST_CPU_ON_LINE_CPUS_LIST="0-31" -export HOST_CPU_PHYSICAL_CORES_PER_SOCKET="24" -export HOST_CPU_SOCKETS="1" -export HOST_CPU_THREADS_PER_CORE="2" -export HOST_CPU_TOTAL_CORES="32" -export HOST_CPU_TOTAL_LOGICAL_CORES="32" -export HOST_CPU_VENDOR_ID="GenuineIntel" -export HOST_DISK_CAPACITY="6.4T" -export HOST_MEMORY_CAPACITY="134G" -export HOST_OS_BITS="64" -export HOST_OS_FLAVOR="ubuntu" -export HOST_OS_FLAVOR_LIKE="debian" -export HOST_OS_GLIBC_VERSION="2.39" -export HOST_OS_KERNEL_VERSION="6.8.0-31-generic" -export HOST_OS_MACHINE="x86_64" -export HOST_OS_PACKAGE_MANAGER="apt" -export HOST_OS_PACKAGE_MANAGER_INSTALL_CMD="DEBIAN_FRONTEND=noninteractive apt-get install -y" -export HOST_OS_PACKAGE_MANAGER_UPDATE_CMD="apt-get update -y" -export HOST_OS_TYPE="linux" -export HOST_OS_VERSION="24.04" -export HOST_PLATFORM_FLAVOR="x86_64" -export HOST_PYTHON_BITS="64" -export HOST_SYSTEM_NAME="intel-spr-i9" -export LOG_PATH="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python/test_results/intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut/bert-99/offline/performance/run_1" -export ML_MODEL="BERT" -export ML_MODEL_BERT_LARGE_FP32_PATH="/home/anandhu/CM/repos/local/cache/b0f750a1b4b44b21/model.pytorch" -export ML_MODEL_BERT_PACKED="no" -export ML_MODEL_BERT_VOCAB_FILE_WITH_PATH="/home/anandhu/CM/repos/local/cache/48d205c056224895/vocab.txt" -export ML_MODEL_DATASET="squad-1.1" -export ML_MODEL_F1="90.874" -export ML_MODEL_FILE="model.pytorch" -export ML_MODEL_FILE_WITH_PATH="/home/anandhu/CM/repos/local/cache/b0f750a1b4b44b21/model.pytorch" -export ML_MODEL_FRAMEWORK="pytorch" -export ML_MODEL_INPUT_IDS_NAME="input_ids" -export ML_MODEL_INPUT_MASK_NAME="input_mask" -export ML_MODEL_INPUT_SEGMENTS_NAME="segment_ids" -export ML_MODEL_MAX_SEQ_LENGTH="384" -export ML_MODEL_NAME="MLPERF BERT Large on SQuAD v1.1" -export ML_MODEL_OUTPUT_END_LOGITS_NAME="output_end_logits" -export ML_MODEL_OUTPUT_START_LOGITS_NAME="output_start_logits" -export ML_MODEL_PRECISION="fp32" -export ML_MODEL_STARTING_WEIGHTS_FILENAME="https://armi.in/files/fp32/model.pytorch" -export MODEL_DIR="/home/anandhu/CM/repos/local/cache/b0f750a1b4b44b21" -export MODEL_FILE="/home/anandhu/CM/repos/local/cache/b0f750a1b4b44b21/model.pytorch" -export OUTPUT_BASE_DIR="/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/app-mlperf-inference-mlcommons-python" -export RUN_DIR="/home/anandhu/CM/repos/local/cache/2375d79dbb604f09/inference/language/bert" -export SKIP_VERIFY_ACCURACY="True" -export USE_CUDA="True" -export USE_GPU="True" -export VOCAB_FILE="/home/anandhu/CM/repos/local/cache/48d205c056224895/vocab.txt" - - -. "/home/anandhu/CM/repos/anandhu-eng@cm4mlops/script/benchmark-program/run-ubuntu.sh" diff --git a/script/app-mlperf-inference-mlcommons-python/tmp-state.json b/script/app-mlperf-inference-mlcommons-python/tmp-state.json deleted file mode 100644 index ece61df58..000000000 --- a/script/app-mlperf-inference-mlcommons-python/tmp-state.json +++ /dev/null @@ -1,207 +0,0 @@ -{ - "CM_HW_META": { - "accelerator_frequency": "2520.000000 MHz", - "accelerator_host_interconnect": "N/A", - "accelerator_interconnect": "N/A", - "accelerator_interconnect_topology": "", - "accelerator_memory_capacity": "23.64288330078125 GB", - "accelerator_memory_configuration": "N/A", - "accelerator_model_name": "NVIDIA GeForce RTX 4090", - "accelerator_on-chip_memories": "", - "accelerators_per_node": "1", - "cooling": "air", - "division": "open", - "framework": "pytorch2.2", - "host_memory_capacity": "134G", - "host_memory_configuration": "undefined", - "host_network_card_count": "1", - "host_networking": "Gig Ethernet", - "host_networking_topology": "N/A", - "host_processor_caches": "L1d cache: 896 KiB (24 instances), L1i cache: 1.3 MiB (24 instances), L2 cache: 32 MiB (12 instances), L3 cache: 36 MiB (1 instance)", - "host_processor_core_count": "24", - "host_processor_frequency": "5800.0000", - "host_processor_interconnect": "", - "host_processor_model_name": "13th Gen Intel(R) Core(TM) i9-13900K", - "host_processors_per_node": "1", - "host_storage_capacity": "6.4T", - "host_storage_type": "SSD", - "hw_notes": "", - "number_of_nodes": "1", - "operating_system": "Ubuntu 24.04 (linux-6.8.0-31-generic-glibc2.39)", - "other_software_stack": "Python: 3.12.3, LLVM-15.0.6", - "status": "available", - "submitter": "CTuning", - "sw_notes": "", - "system_name": "intel_spr_i9 (auto detected)", - "system_type": "edge", - "system_type_detail": "edge server" - }, - "CM_SUT_CONFIG": { - "intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut": { - "3d-unet-99": { - "Offline": { - "target_qps": 1.0 - }, - "Server": { - "target_qps": 1.0 - }, - "SingleStream": { - "target_latency": 500 - } - }, - "3d-unet-99.9": { - "Offline": { - "target_qps": 1.0 - }, - "Server": { - "target_qps": 1.0 - }, - "SingleStream": { - "target_latency": 500 - } - }, - "bert-99": { - "Offline": { - "target_qps": 1 - }, - "Server": { - "target_qps": 1.0 - }, - "SingleStream": { - "target_latency": 1 - } - }, - "bert-99.9": { - "Offline": { - "target_qps": 1.0 - }, - "Server": { - "target_qps": 1.0 - } - }, - "gpt-j": { - "Offline": { - "target_qps": 1.0 - }, - "Server": { - "target_qps": 1.0 - }, - "SingleStream": { - "target_latency": 500 - } - }, - "llama2-70b-99": { - "Offline": { - "target_qps": 0.1 - }, - "Server": { - "target_qps": 0.1 - }, - "SingleStream": { - "target_latency": 2000 - } - }, - "llama2-70b-99.9": { - "Offline": { - "target_qps": 0.1 - }, - "Server": { - "target_qps": 0.1 - }, - "SingleStream": { - "target_latency": 2000 - } - }, - "resnet50": { - "MultiStream": { - "target_latency": 0.1 - }, - "Offline": { - "target_qps": 1.0 - }, - "Server": { - "target_qps": 1.0 - }, - "SingleStream": { - "target_latency": 0.1 - } - }, - "retinanet": { - "MultiStream": { - "target_latency": 1 - }, - "Offline": { - "target_qps": 1.0 - }, - "Server": { - "target_qps": 1.0 - }, - "SingleStream": { - "target_latency": 1 - } - }, - "sdxl": { - "Offline": { - "target_qps": 1.0 - }, - "Server": { - "target_qps": 1.0 - }, - "SingleStream": { - "target_latency": 200 - } - } - } - }, - "CM_SUT_CONFIG_NAME": "intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut", - "CM_SUT_CONFIG_PATH": { - "intel_spr_i9-reference-gpu-pytorch-v2.3.0-network_sut": "/home/anandhu/CM/repos/local/cache/5043351ffa174ecf/configs/intel_spr_i9/reference-implementation/gpu-device/pytorch-framework/framework-version-v2.3.0/network_sut-config.yaml" - }, - "CM_SUT_META": { - "accelerator_frequency": "2520.000000 MHz", - "accelerator_host_interconnect": "N/A", - "accelerator_interconnect": "N/A", - "accelerator_interconnect_topology": "", - "accelerator_memory_capacity": "23.64288330078125 GB", - "accelerator_memory_configuration": "N/A", - "accelerator_model_name": "NVIDIA GeForce RTX 4090", - "accelerator_on-chip_memories": "", - "accelerators_per_node": "1", - "cooling": "air", - "division": "open", - "framework": "pytorch2.2", - "host_memory_capacity": "134G", - "host_memory_configuration": "undefined", - "host_network_card_count": "1", - "host_networking": "Gig Ethernet", - "host_networking_topology": "N/A", - "host_processor_caches": "L1d cache: 896 KiB (24 instances), L1i cache: 1.3 MiB (24 instances), L2 cache: 32 MiB (12 instances), L3 cache: 36 MiB (1 instance)", - "host_processor_core_count": "24", - "host_processor_frequency": "5800.0000", - "host_processor_interconnect": "", - "host_processor_model_name": "13th Gen Intel(R) Core(TM) i9-13900K", - "host_processors_per_node": "1", - "host_storage_capacity": "6.4T", - "host_storage_type": "SSD", - "hw_notes": "", - "number_of_nodes": "1", - "operating_system": "Ubuntu 24.04 (linux-6.8.0-31-generic-glibc2.39)", - "other_software_stack": "Python: 3.12.3, LLVM-15.0.6", - "status": "available", - "submitter": "CTuning", - "sw_notes": "", - "system_name": "intel_spr_i9 (auto detected)", - "system_type": "edge", - "system_type_detail": "edge server" - }, - "RUN": { - "Offline": {} - }, - "docker": {}, - "mlperf-inference-implementation": { - "script_id": "app-mlperf-inference,d775cac873ee4231:reference,bert-99,pytorch,cuda,test,offline" - }, - "mlperf_inference_run_cmd": "cm run script --tags=generate-run-cmds,inference --model=bert-99 --backend=pytorch --mode=performance --device=cuda --quiet --test_query_count=1000 --network=sut --adr.cuda.version=12.4.1", - "os_uname_all": "Linux intel-spr-i9 6.8.0-31-generic #31-Ubuntu SMP PREEMPT_DYNAMIC Sat Apr 20 00:40:06 UTC 2024 x86_64 x86_64 x86_64 GNU/Linux", - "os_uname_machine": "x86_64" -}