Skip to content

Commit

Permalink
Merge pull request #9 from anandhu-eng/cm_readme_inference_update
Browse files Browse the repository at this point in the history
Cm readme inference update
  • Loading branch information
anandhu-eng authored Sep 3, 2024
2 parents 725b3c0 + 8815065 commit 44872c0
Show file tree
Hide file tree
Showing 33 changed files with 1,132 additions and 155 deletions.
5 changes: 5 additions & 0 deletions docs/benchmarks/image_classification/get-resnet50-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Image Classification using ResNet50

## Dataset
Expand Down
5 changes: 5 additions & 0 deletions docs/benchmarks/image_classification/mobilenets.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Image Classification using Mobilenet models

Mobilenet models are not official MLPerf models and so cannot be used for a Closed division MLPerf inference submission. But since they can be run with Imagenet dataset, we are allowed to use them for Open division submission. Only CPU runs are supported now.
Expand Down
5 changes: 5 additions & 0 deletions docs/benchmarks/image_classification/resnet50.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Image Classification using ResNet50

=== "MLCommons-Python"
Expand Down
27 changes: 0 additions & 27 deletions docs/benchmarks/index.md

This file was deleted.

14 changes: 6 additions & 8 deletions docs/benchmarks/language/bert.md
Original file line number Diff line number Diff line change
@@ -1,36 +1,34 @@
---
hide:
- toc
---

# Question Answering using Bert-Large

=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

BERT-99
{{ mlperf_inference_implementation_readme (4, "bert-99", "reference") }}

BERT-99.9
{{ mlperf_inference_implementation_readme (4, "bert-99.9", "reference") }}

=== "Nvidia"
## Nvidia MLPerf Implementation

BERT-99
{{ mlperf_inference_implementation_readme (4, "bert-99", "nvidia") }}

BERT-99.9
{{ mlperf_inference_implementation_readme (4, "bert-99.9", "nvidia") }}

=== "Intel"
## Intel MLPerf Implementation
BERT-99

{{ mlperf_inference_implementation_readme (4, "bert-99", "intel") }}

BERT-99.9
{{ mlperf_inference_implementation_readme (4, "bert-99.9", "intel") }}

=== "Qualcomm"
## Qualcomm AI100 MLPerf Implementation

BERT-99
{{ mlperf_inference_implementation_readme (4, "bert-99", "qualcomm") }}

BERT-99.9
{{ mlperf_inference_implementation_readme (4, "bert-99.9", "qualcomm") }}
5 changes: 5 additions & 0 deletions docs/benchmarks/language/get-bert-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Question Answering using Bert-Large

## Dataset
Expand Down
5 changes: 5 additions & 0 deletions docs/benchmarks/language/get-gptj-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Text Summarization using GPT-J

## Dataset
Expand Down
9 changes: 9 additions & 0 deletions docs/benchmarks/language/get-llama2-70b-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Text Summarization using LLAMA2-70b

## Dataset
Expand All @@ -23,4 +28,8 @@ Get the Official MLPerf LLAMA2-70b Model
```
cm run script --tags=get,ml-model,llama2-70b,_pytorch -j
```

!!! tip

Downloading llama2-70B model from Hugging Face will prompt you to enter the Hugging Face username and password. Please note that the password required is the [**access token**](https://huggingface.co/settings/tokens) generated for your account. Additionally, ensure that your account has access to the [llama2-70B](https://huggingface.co/meta-llama/Llama-2-70b-chat-hf) model.

5 changes: 5 additions & 0 deletions docs/benchmarks/language/get-mixtral-8x7b-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

## Dataset

The benchmark implementation run command will automatically download the preprocessed validation and calibration datasets. In case you want to download only the datasets, you can use the below commands.
Expand Down
13 changes: 6 additions & 7 deletions docs/benchmarks/language/gpt-j.md
Original file line number Diff line number Diff line change
@@ -1,39 +1,38 @@
---
hide:
- toc
---

# Text Summarization using GPT-J


=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

GPT-J-99


{{ mlperf_inference_implementation_readme (4, "gptj-99", "reference") }}

GPTJ-99.9

{{ mlperf_inference_implementation_readme (4, "gptj-99.9", "reference") }}

=== "Nvidia"
## Nvidia MLPerf Implementation

GPTJ-99

{{ mlperf_inference_implementation_readme (4, "gptj-99", "nvidia") }}

GPTJ-99.9

{{ mlperf_inference_implementation_readme (4, "gptj-99.9", "nvidia") }}

=== "Intel"
## Intel MLPerf Implementation
GPTJ-99

{{ mlperf_inference_implementation_readme (4, "gptj-99", "intel") }}


=== "Qualcomm"
## Qualcomm AI100 MLPerf Implementation

GPTJ-99

{{ mlperf_inference_implementation_readme (4, "gptj-99", "qualcomm") }}

20 changes: 10 additions & 10 deletions docs/benchmarks/language/llama2-70b.md
Original file line number Diff line number Diff line change
@@ -1,28 +1,28 @@
---
hide:
- toc
---

# Text Summarization using LLAMA2-70b


=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

LLAMA2-70b-99
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99", "reference") }}

LLAMA2-70b-99.9
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99.9", "reference") }}

=== "Nvidia"
## Nvidia MLPerf Implementation

LLAMA2-70b-99
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99", "nvidia") }}

LLAMA2-70b-99.9
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99.9", "nvidia") }}

=== "Neural Magic"
## Neural Magic MLPerf Implementation

{{ mlperf_inference_implementation_readme (4, "llama2-70b-99", "neuralmagic") }}

=== "Qualcomm"
## Qualcomm AI100 MLPerf Implementation

LLAMA2-70b-99
{{ mlperf_inference_implementation_readme (4, "llama2-70b-99", "qualcomm") }}

{{ mlperf_inference_implementation_readme (4, "llama2-70b-99.9", "neuralmagic") }}
5 changes: 4 additions & 1 deletion docs/benchmarks/language/mixtral-8x7b.md
Original file line number Diff line number Diff line change
@@ -1,6 +1,9 @@
---
hide:
- toc
---

=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

MIXTRAL-8x7b
{{ mlperf_inference_implementation_readme (4, "mixtral-8x7b", "reference") }}
11 changes: 5 additions & 6 deletions docs/benchmarks/medical_imaging/3d-unet.md
Original file line number Diff line number Diff line change
@@ -1,33 +1,32 @@
---
hide:
- toc
---

# Medical Imaging using 3d-unet (KiTS 2019 kidney tumor segmentation task)


=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

3d-unet-99

{{ mlperf_inference_implementation_readme (4, "3d-unet-99", "reference") }}

3d-unet-99.9

{{ mlperf_inference_implementation_readme (4, "3d-unet-99.9", "reference") }}

=== "Nvidia"
## Nvidia MLPerf Implementation
3d-unet-99

{{ mlperf_inference_implementation_readme (4, "3d-unet-99", "nvidia") }}

3d-unet-99.9

{{ mlperf_inference_implementation_readme (4, "3d-unet-99.9", "nvidia") }}

=== "Intel"
## Intel MLPerf Implementation
3d-unet-99

{{ mlperf_inference_implementation_readme (4, "3d-unet-99", "intel") }}

3d-unet-99.9

{{ mlperf_inference_implementation_readme (4, "3d-unet-99.9", "intel") }}
14 changes: 12 additions & 2 deletions docs/benchmarks/medical_imaging/get-3d-unet-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Medical Imaging using 3d-unet (KiTS 2019 kidney tumor segmentation task)

## Dataset
Expand All @@ -7,9 +12,14 @@ The benchmark implementation run command will automatically download the validat
=== "Validation"
3d-unet validation run uses the KiTS19 dataset performing [KiTS 2019](https://kits19.grand-challenge.org/) kidney tumor segmentation task

### Get Validation Dataset
### Get Validation Dataset(Original)
```
cm run script --tags=get,dataset,kits19,_validation -j
```

### Get Validation Dataset(Preprocessed)
```
cm run script --tags=get,dataset,kits19,validation -j
cm run script --tags=get,dataset,kits19,preprocessed -j
```

## Model
Expand Down
5 changes: 5 additions & 0 deletions docs/benchmarks/object_detection/get-retinanet-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Object Detection using Retinanet

## Dataset
Expand Down
5 changes: 5 additions & 0 deletions docs/benchmarks/object_detection/retinanet.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Object Detection using Retinanet

=== "MLCommons-Python"
Expand Down
25 changes: 16 additions & 9 deletions docs/benchmarks/recommendation/dlrm-v2.md
Original file line number Diff line number Diff line change
@@ -1,22 +1,29 @@
---
hide:
- toc
---

# Recommendation using DLRM v2


## Benchmark Implementations
=== "MLCommons-Python"
## MLPerf Reference Implementation in Python

DLRM-v2-99
{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99", "reference") }}
{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99", "reference") }}

DLRM-v2-99.9
{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99.9", "reference") }}
{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "reference") }}

=== "Nvidia"
## Nvidia MLPerf Implementation

DLRM-v2-99
{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99", "nvidia") }}

DLRM-v2-99.9
{{ mlperf_inference_implementation_readme (4, "dlrm_v2-99.9", "nvidia") }}
{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99", "nvidia") }}

{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "nvidia") }}

=== "Intel"
## Intel MLPerf Implementation

{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99", "intel") }}

{{ mlperf_inference_implementation_readme (4, "dlrm-v2-99.9", "intel") }}
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Recommendation using DLRM v2

## Dataset
Expand All @@ -9,7 +14,7 @@ The benchmark implementation run command will automatically download the validat

### Get Validation Dataset
```
cm run script --tags=get,dataset,criteo,validation -j
cm run script --tags=get,dataset,criteo,_validation -j
```
## Model
The benchmark implementation run command will automatically download the required model and do the necessary conversions. In case you want to only download the official model, you can use the below commands.
Expand All @@ -20,6 +25,6 @@ Get the Official MLPerf DLRM v2 Model

### Pytorch
```
cm run script --tags=get,ml-model,dlrm_v2,_pytorch -j
cm run script --tags=get,ml-model,dlrm,_pytorch -j
```

5 changes: 5 additions & 0 deletions docs/benchmarks/text_to_image/get-sdxl-data.md
Original file line number Diff line number Diff line change
@@ -1,3 +1,8 @@
---
hide:
- toc
---

# Text to Image using Stable Diffusion

## Dataset
Expand Down
Loading

0 comments on commit 44872c0

Please sign in to comment.