diff --git a/docs/benchmarks/recommendation/dlrm-v2.md b/docs/benchmarks/recommendation/dlrm-v2.md index 0657176c7..d1b41deb1 100644 --- a/docs/benchmarks/recommendation/dlrm-v2.md +++ b/docs/benchmarks/recommendation/dlrm-v2.md @@ -11,25 +11,25 @@ hide: ## MLPerf Reference Implementation in Python DLRM-v2-99 -{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99", "reference") }} +{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99", "reference") }} DLRM-v2-99.9 -{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99.9", "reference") }} +{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "reference") }} === "Nvidia" ## Nvidia MLPerf Implementation DLRM-v2-99 -{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99", "nvidia") }} +{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99", "nvidia") }} DLRM-v2-99.9 -{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99.9", "nvidia") }} +{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "nvidia") }} === "Intel" ## Intel MLPerf Implementation DLRM-v2-99 -{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99", "intel") }} +{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99", "intel") }} DLRM-v2-99.9 -{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99.9", "intel") }} \ No newline at end of file +{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "intel") }} \ No newline at end of file diff --git a/main.py b/main.py index d48d9f3d5..dc71f195a 100644 --- a/main.py +++ b/main.py @@ -39,7 +39,7 @@ def mlperf_inference_implementation_readme(spaces, model, implementation): frameworks = [ "pytorch" ] elif implementation == "intel": - if model not in [ "bert-99", "bert-99.9", "gptj-99", "gptj-99.9", "resnet50", "retinanet", "3d-unet-99", "3d-unet-99.9", "dlrm_v2-99", "dlrm_v2-99.9" ]: + if model not in [ "bert-99", "bert-99.9", "gptj-99", "gptj-99.9", "resnet50", "retinanet", "3d-unet-99", "3d-unet-99.9", "dlrm-v2-99", "dlrm-v2-99.9" ]: return pre_space+" WIP" if model in [ "bert-99", "bert-99.9", "retinanet", "3d-unet-99", "3d-unet-99.9" ]: code_version="r4.0" @@ -274,6 +274,9 @@ def mlperf_inference_run_command(spaces, model, implementation, framework, categ docker_cmd_suffix += f" \\\n{pre_space} --nvidia_llama2_dataset_file_path=" else: docker_cmd_suffix += f" \\\n{pre_space} --api_server=" + + if "dlrm-v2" in model and implementation == "nvidia": + docker_cmd_suffix += f" \\\n{pre_space} --criteo_day23_raw_data_path=" docker_setup_cmd = f"""\n {f_pre_space}```bash @@ -300,6 +303,9 @@ def mlperf_inference_run_command(spaces, model, implementation, framework, categ cmd_suffix += f" \\\n{pre_space} --nvidia_llama2_dataset_file_path=" else: cmd_suffix += f" \\\n{pre_space} --api_server=" + + if "dlrm-v2" in model and implementation == "nvidia": + cmd_suffix += f" \\\n{pre_space} --criteo_day23_raw_data_path=" run_cmd = f"""\n {f_pre_space}```bash