Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Fix a typo in benchmark-program #583

Merged
merged 5 commits into from
Nov 22, 2024
Merged
Show file tree
Hide file tree
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension


Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
Original file line number Diff line number Diff line change
Expand Up @@ -39,6 +39,6 @@ jobs:
pip install --upgrade cm4mlops
cm pull repo

cm run script --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r4.1-dev --preprocess_submission=yes --adr.submission-checker-src.tags=_branch.dev --execution_mode=valid --gpu_name=rtx_4090 --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=$hw_name --implementation=nvidia --backend=tensorrt --category=datacenter,edge --division=closed --docker_dt=yes --docker_it=no --docker_cm_repo=gateoverflow@cm4mlops --adr.compiler.tags=gcc --device=cuda --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean --docker --quiet
cm run script --tags=run-mlperf,inference,_all-scenarios,_submission,_full,_r4.1-dev --preprocess_submission=yes --adr.submission-checker-src.tags=_branch.dev --execution_mode=valid --gpu_name=rtx_4090 --pull_changes=yes --pull_inference_changes=yes --model=${{ matrix.model }} --submitter="MLCommons" --hw_name=$hw_name --implementation=nvidia --backend=tensorrt --category=datacenter,edge --division=closed --docker_dt=yes --docker_it=no --docker_cm_repo=gateoverflow@cm4mlops --adr.compiler.tags=gcc --device=cuda --use_model_from_host=yes --use_dataset_from_host=yes --results_dir=$HOME/gh_action_results --submission_dir=$HOME/gh_action_submissions --clean --docker --quiet

cm run script --tags=push,github,mlperf,inference,submission --repo_url=https://github.com/gateoverflow/mlperf_inference_unofficial_submissions_v5.0 --repo_branch=main --commit_message="Results from GH action on NVIDIA_$hw_name" --quiet --submission_dir=$HOME/gh_action_submissions --hw_name=$hw_name
28 changes: 21 additions & 7 deletions script/app-mlperf-inference/_cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -769,36 +769,44 @@ variations:
sdxl,nvidia-original:
docker:
deps:
- enable_if_env:
- enable_if_any_env:
CM_MLPERF_MODEL_SDXL_DOWNLOAD_TO_HOST:
- 'yes'
CM_USE_MODEL_FROM_HOST:
- 'yes'
tags: get,ml-model,sdxl,_fp16,_rclone

sdxl,reference,float16:
docker:
image_name: mlperf-inference-mlcommons-python-implementation-sdxl-float16
deps:
- enable_if_env:
- enable_if_any_env:
CM_MLPERF_MODEL_SDXL_DOWNLOAD_TO_HOST:
- 'yes'
CM_USE_MODEL_FROM_HOST:
- 'yes'
tags: get,ml-model,sdxl,_fp16,_rclone

sdxl,reference,bfloat16:
docker:
image_name: mlperf-inference-mlcommons-python-implementation-sdxl-bfloat16
deps:
- enable_if_env:
- enable_if_any_env:
CM_MLPERF_MODEL_SDXL_DOWNLOAD_TO_HOST:
- 'yes'
CM_USE_MODEL_FROM_HOST:
- 'yes'
tags: get,ml-model,sdxl,_fp16,_rclone

sdxl,reference,float32:
docker:
image_name: mlperf-inference-mlcommons-python-implementation-sdxl-float32
deps:
- enable_if_env:
- enable_if_any_env:
CM_MLPERF_MODEL_SDXL_DOWNLOAD_TO_HOST:
- 'yes'
CM_USE_MODEL_FROM_HOST:
- 'yes'
tags: get,ml-model,sdxl,_fp32,_rclone

llama2-70b_:
Expand Down Expand Up @@ -849,9 +857,11 @@ variations:
docker:
image_name: mlperf-inference-mlcommons-python-implementation-llama2-70b
deps:
- enable_if_env:
- enable_if_any_env:
CM_MLPERF_MODEL_LLAMA2_70B_DOWNLOAD_TO_HOST:
- 'yes'
CM_USE_MODEL_FROM_HOST:
- 'yes'
tags: get,ml-model,llama2

llama2-70b_,amd:
Expand All @@ -860,9 +870,11 @@ variations:
mounts:
- "${{ CM_LLAMA2_FINAL_SAFE_TENSORS_PATH }}:${{ CM_LLAMA2_FINAL_SAFE_TENSORS_PATH }"
deps:
- enable_if_env:
- enable_if_any_env:
CM_MLPERF_MODEL_LLAMA2_70B_DOWNLOAD_TO_HOST:
- 'yes'
CM_USE_MODEL_FROM_HOST:
- 'yes'
tags: get,ml-model,llama2,_amd,_pytorch

mixtral-8x7b:
Expand Down Expand Up @@ -901,9 +913,11 @@ variations:
names:
- ml-model
- mixtral-model
enable_if_env:
enable_if_any_env:
CM_MLPERF_MODEL_MIXTRAL_8X7B_DOWNLOAD_TO_HOST:
- 'yes'
CM_USE_MODEL_FROM_HOST:
- 'yes'
- tags: get,dataset-mixtral,openorca-mbxp-gsm8k-combined
names:
- openorca-mbxp-gsm8k-combined-preprocessed
Expand Down
2 changes: 1 addition & 1 deletion script/benchmark-program/customize.py
Original file line number Diff line number Diff line change
Expand Up @@ -45,7 +45,7 @@ def preprocess(i):

if os_info['platform'] != 'windows' and str(env.get('CM_SAVE_CONSOLE_LOG', True)).lower() not in [ "no", "false", "0"]:
logs_dir = env.get('CM_LOGS_DIR', env['CM_RUN_DIR'])
env['CM_RUN_CMD'] += r" 2>&1 | tee " + q+ os.path.join(logs_dir, "console.out") + q + r" echo \${PIPESTATUS[0]} > exitstatus"
env['CM_RUN_CMD'] += r" 2>&1 | tee " + q+ os.path.join(logs_dir, "console.out") + q + r"; echo \${PIPESTATUS[0]} > exitstatus"

# additional arguments and tags for measuring system informations(only if 'CM_PROFILE_NVIDIA_POWER' is 'on')
if env.get('CM_PROFILE_NVIDIA_POWER', '') == "on":
Expand Down
1 change: 1 addition & 0 deletions script/run-mlperf-inference-app/_cm.yaml
Original file line number Diff line number Diff line change
Expand Up @@ -116,6 +116,7 @@ input_mapping:
all_models: CM_MLPERF_ALL_MODELS
criteo_day23_raw_data_path: CM_CRITEO_DAY23_RAW_DATA_PATH
use_dataset_from_host: CM_USE_DATASET_FROM_HOST
use_model_from_host: CM_USE_MODEL_FROM_HOST

new_state_keys:
- app_mlperf_inference_*
Expand Down
2 changes: 1 addition & 1 deletion script/run-mlperf-inference-app/customize.py
Original file line number Diff line number Diff line change
Expand Up @@ -94,7 +94,7 @@ def preprocess(i):
env['CM_MLPERF_LOADGEN_SCENARIO'] = "Offline"

if env.get('CM_MLPERF_LOADGEN_ALL_SCENARIOS', '') == "yes":
env['CM_MLPERF_LOADGEN_SCENARIOS'] = get_valid_scenarios(env['CM_MODEL'], system_meta['system_type'], env['CM_MLPERF_LAST_RELEASE'], env['CM_MLPERF_INFERENCE_SOURCE'])
env['CM_MLPERF_LOADGEN_SCENARIOS'] = get_valid_scenarios(env['CM_MODEL'], system_meta.get('system_type', 'edge'), env['CM_MLPERF_LAST_RELEASE'], env['CM_MLPERF_INFERENCE_SOURCE'])
else:
system_meta = {}
env['CM_MLPERF_LOADGEN_SCENARIOS'] = [ env['CM_MLPERF_LOADGEN_SCENARIO'] ]
Expand Down
Loading