You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
The benchmark implementation run command will automatically download the validation and calibration datasets and do the necessary preprocessing. In case you want to download only the datasets, you can use the below commands.
11
+
12
+
=== "Validation"
13
+
14
+
### Get Validation Dataset
15
+
```
16
+
mlcr get,preprocessed,dataset,deepseek-r1,_validation,_mlc,_rclone --outdirname=<path to download> -j
17
+
```
18
+
19
+
=== "Calibration"
20
+
21
+
### Get Calibration Dataset
22
+
```
23
+
mlcr get,preprocessed,dataset,deepseek-r1,_calibration,_mlc,_rclone --outdirname=<path to download> -j
The benchmark implementation run command will automatically download the validation and calibration datasets and do the necessary preprocessing. In case you want to download only the datasets, you can use the below commands.
11
+
12
+
=== "Validation"
13
+
14
+
=== "Full dataset (Datacenter)"
15
+
16
+
### Get Validation Dataset
17
+
```
18
+
mlcr get,dataset,cnndm,_validation,_datacenter,_llama3,_mlc,_rclone --outdirname=<path to download> -j
19
+
```
20
+
21
+
=== "5000 samples (Edge)"
22
+
23
+
### Get Validation Dataset
24
+
```
25
+
mlcr get,dataset,cnndm,_validation,_edge,_llama3,_mlc,_rclone --outdirname=<path to download> -j
26
+
```
27
+
28
+
=== "Calibration"
29
+
30
+
### Get Calibration Dataset
31
+
```
32
+
mlcr get,dataset,cnndm,_calibration,_llama3,_mlc,_rclone --outdirname=<path to download> -j
33
+
```
34
+
35
+
-`--outdirname=<PATH_TO_DOWNLOAD_LLAMA3_405B_DATASET>` could be provided to download the dataset to a specific location.
36
+
37
+
## Model
38
+
The benchmark implementation run command will automatically download the required model and do the necessary conversions. In case you want to only download the official model, you can use the below commands.
39
+
40
+
=== "Pytorch"
41
+
42
+
=== "From MLCOMMONS Google Drive"
43
+
44
+
> **Note:** One has to accept the [MLCommons Llama 3.1 License Confidentiality Notice](http://llama3-1.mlcommons.org/) to access the model files in MLCOMMONS Google Drive.
45
+
46
+
### Get the Official MLPerf LLAMA3.1-405B model from MLCOMMONS Cloudfare R2
47
+
```
48
+
TBD
49
+
```
50
+
51
+
=== "From Hugging Face repo"
52
+
53
+
> **Note:** Access to the HuggingFace model could be requested [here](https://ai.meta.com/resources/models-and-libraries/llama-downloads/).
The benchmark implementation run command will automatically download the validation and calibration datasets and do the necessary preprocessing. In case you want to download only the datasets, you can use the below commands.
11
+
12
+
=== "Validation"
13
+
14
+
=== "Preprocessed"
15
+
16
+
### Get Preprocessed Validation Dataset
17
+
```
18
+
mlcr get,dataset,whisper,_preprocessed,_mlc,_rclone --outdirname=<path to download> -j
19
+
```
20
+
21
+
=== "Unprocessed"
22
+
23
+
### Get Unprocessed Validation Dataset
24
+
```
25
+
mlcr get,dataset,whisper,_unprocessed --outdirname=<path to download> -j
26
+
```
27
+
28
+
## Model
29
+
The benchmark implementation run command will automatically download the required model and do the necessary conversions if any. In case you want to only download the official model, you can use the below commands.
30
+
31
+
=== "Pytorch"
32
+
33
+
=== "From MLCOMMONS"
34
+
35
+
### Get the Official MLPerf Whisper model from MLCOMMONS Cloudflare R2
36
+
```
37
+
mlcr get,ml-model,whisper,_rclone,_mlc s-j
38
+
```
39
+
40
+
-`--outdirname=<PATH_TO_DOWNLOAD_WHISPER_MODEL>` could be provided to download the model to a specific location.
## Automated command to run the benchmark via MLFlow
4
+
5
+
Please see the [new docs site](https://docs.mlcommons.org/inference/benchmarks/language/deepseek-r1/) for an automated way to run this benchmark across different available implementations and do an end-to-end submission with or without docker.
6
+
7
+
You can also do pip install mlc-scripts and then use `mlcr` commands for downloading the model and datasets using the commands given in the later sections.
Copy file name to clipboardExpand all lines: language/llama3.1-8b/README.md
+40-12Lines changed: 40 additions & 12 deletions
Display the source diff
Display the rich diff
Original file line number
Diff line number
Diff line change
@@ -9,7 +9,7 @@
9
9
10
10
## Automated command to run the benchmark via MLFlow
11
11
12
-
Please see the [new docs site](https://docs.mlcommons.org/inference/benchmarks/language/llama3_1-8b/)(TBD) for an automated way to run this benchmark across different available implementations and do an end-to-end submission with or without docker.
12
+
Please see the [new docs site](https://docs.mlcommons.org/inference/benchmarks/language/llama3_1-8b/) for an automated way to run this benchmark across different available implementations and do an end-to-end submission with or without docker.
13
13
14
14
You can also do pip install mlc-scripts and then use `mlcr` commands for downloading the model and datasets using the commands given in the later sections.
15
15
@@ -99,7 +99,10 @@ pip install -e ../../loadgen
99
99
## Get Model
100
100
### MLCommons Members Download (Recommended for official submission)
101
101
102
-
You need to request for access to [MLCommons](http://llama3-1.mlcommons.org/) and you'll receive an email with the download instructions. You can download the model automatically via the below command
102
+
You need to request for access to [MLCommons](http://llama3-1.mlcommons.org/) and you'll receive an email with the download instructions.
103
+
104
+
**Official Model download using MLCFlow Automation**
105
+
You can download the model automatically via the below command
0 commit comments