Skip to content

Commit

Permalink
lower cased implementation
Browse files Browse the repository at this point in the history
  • Loading branch information
anandhu-eng committed Aug 12, 2024
1 parent cd823cc commit 0c94ea9
Show file tree
Hide file tree
Showing 2 changed files with 6 additions and 6 deletions.
4 changes: 2 additions & 2 deletions docs/benchmarks/language/llama2-70b.md
Original file line number Diff line number Diff line change
Expand Up @@ -28,7 +28,7 @@ hide:
## Neural Magic MLPerf Implementation

LLAMA2-70b-99
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99", "NeuralMagic") }}
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99", "neuralmagic") }}

LLAMA2-70b-99.9
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99.9", "NeuralMagic") }}
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99.9", "neuralmagic") }}
8 changes: 4 additions & 4 deletions main.py
Original file line number Diff line number Diff line change
Expand Up @@ -34,9 +34,9 @@ def mlperf_inference_implementation_readme(spaces, model, implementation):
devices = [ "CUDA" ]
frameworks = [ "TensorRT" ]

elif implementation == "NeuralMagic":
elif implementation == "neuralmagic":
devices = [ "CUDA" ]
frameworks = [ "vLLM" ]
frameworks = [ "pytorch" ]

elif implementation == "intel":
if model not in [ "bert-99", "bert-99.9", "gptj-99", "gptj-99.9", "resnet50", "retinanet", "3d-unet-99", "3d-unet-99.9", "dlrm_v2-99", "dlrm_v2-99.9" ]:
Expand Down Expand Up @@ -269,7 +269,7 @@ def mlperf_inference_run_command(spaces, model, implementation, framework, categ
docker_cmd_suffix += f" \\\n{pre_space} --test_query_count={test_query_count}"

if "llama2-70b" in model:
if implementation != "NeuralMagic":
if implementation != "neuralmagic":
docker_cmd_suffix += f" \\\n{pre_space} --tp_size=<TP_SIZE>"
docker_cmd_suffix += f" \\\n{pre_space} --nvidia_llama2_dataset_file_path=<PATH_TO_PICKE_FILE>"
else:
Expand All @@ -295,7 +295,7 @@ def mlperf_inference_run_command(spaces, model, implementation, framework, categ
cmd_suffix += f" \\\n {pre_space} --test_query_count={test_query_count}"

if "llama2-70b" in model:
if implementation != "NeuralMagic":
if implementation != "neuralmagic":
cmd_suffix += f" \\\n{pre_space} --tp_size=<TP_SIZE>"
cmd_suffix += f" \\\n{pre_space} --nvidia_llama2_dataset_file_path=<PATH_TO_PICKE_FILE>"
else:
Expand Down

0 comments on commit 0c94ea9

Please sign in to comment.