diff --git a/cm-mlops/script/app-mlperf-inference/_cm.yaml b/cm-mlops/script/app-mlperf-inference/_cm.yaml index e7911b0278..f5c1844ef1 100644 --- a/cm-mlops/script/app-mlperf-inference/_cm.yaml +++ b/cm-mlops/script/app-mlperf-inference/_cm.yaml @@ -1162,3 +1162,4 @@ docker: results_dir: RESULTS_DIR submission_dir: SUBMISSION_DIR dlrm_data_path: DLRM_DATA_PATH + intel_gptj_int8_model_path: CM_MLPERF_INFERENCE_INTEL_GPTJ_INT8_MODEL_PATH diff --git a/cm-mlops/script/reproduce-mlperf-inference-intel/customize.py b/cm-mlops/script/reproduce-mlperf-inference-intel/customize.py index a924508970..51a0ffd6b6 100644 --- a/cm-mlops/script/reproduce-mlperf-inference-intel/customize.py +++ b/cm-mlops/script/reproduce-mlperf-inference-intel/customize.py @@ -64,9 +64,11 @@ def preprocess(i): env['CM_MLPERF_INFERENCE_INTEL_HARNESS_PATH'] = os.path.join(os.getcwd(), "harness", "build", "gptj_inference") env['DATA_PATH'] = os.path.join(os.getcwd(), "harness", "gptj") env['MLPERF_INFERENCE_ROOT'] = env['CM_MLPERF_INFERENCE_SOURCE'] - final_model_path = os.path.join(harness_root, "data", "gpt-j-int8-model") + final_model_path = os.path.join(harness_root, "data", "gpt-j-int8-model", "best_model.pt") env['INT8_MODEL_DIR'] = os.path.dirname(final_model_path) env['CM_ML_MODEL_PATH'] = final_model_path + if env.get('CM_MLPERF_INFERENCE_INTEL_GPTJ_INT8_MODEL_PATH', '') != '': + shutil.copy(env['CM_MLPERF_INFERENCE_INTEL_GPTJ_INT8_MODEL_PATH'], final_model_path) elif env['CM_LOCAL_MLPERF_INFERENCE_INTEL_RUN_MODE'] == "run_harness": if env.get('CM_MLPERF_LOADGEN_MODE', '') == "compliance": diff --git a/cm-mlops/script/run-mlperf-inference-app/_cm.yaml b/cm-mlops/script/run-mlperf-inference-app/_cm.yaml index c0a277b401..f96f4f8028 100644 --- a/cm-mlops/script/run-mlperf-inference-app/_cm.yaml +++ b/cm-mlops/script/run-mlperf-inference-app/_cm.yaml @@ -118,10 +118,18 @@ deps: docker: fake_run_deps: true mounts: - - ${{ GPTJ_CHECKPOINT_PATH }}:${{ GPTJ_CHECKPOINT_PATH }} - ${{ INSTALL_DATA_PATH }}:/install_data - ${{ DATA_PATH }}:/data + - ${{ CM_MLPERF_INFERENCE_INTEL_GPTJ_INT8_MODEL_PATH }}:${{ CM_MLPERF_INFERENCE_INTEL_GPTJ_INT8_MODEL_PATH }} run: true + docker_input_mapping: + imagenet_path: IMAGENET_PATH + gptj_checkpoint_path: GPTJ_CHECKPOINT_PATH + criteo_preprocessed_path: CRITEO_PREPROCESSED_PATH + results_dir: RESULTS_DIR + submission_dir: SUBMISSION_DIR + dlrm_data_path: DLRM_DATA_PATH + intel_gptj_int8_model_path: CM_MLPERF_INFERENCE_INTEL_GPTJ_INT8_MODEL_PATH variations: