-
Notifications
You must be signed in to change notification settings - Fork 538
Commit
This commit does not belong to any branch on this repository, and may belong to a fork outside of the repository.
add docs for llama3 + inference version upgrade (#2020)
* add docs for llama3 + inference version upgrade * add output path and hf token
- Loading branch information
1 parent
e3ea5ef
commit 2c67b24
Showing
4 changed files
with
63 additions
and
2 deletions.
There are no files selected for viewing
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,41 @@ | ||
--- | ||
hide: | ||
- toc | ||
--- | ||
|
||
# Text Summarization using LLAMA3.1-405b | ||
|
||
## Dataset | ||
|
||
The benchmark implementation run command will automatically download the validation and calibration datasets and do the necessary preprocessing. In case you want to download only the datasets, you can use the below commands. | ||
|
||
=== "Validation" | ||
|
||
### Get Validation Dataset | ||
``` | ||
cm run script --tags=get,dataset,mlperf,inference,llama3,_validation --outdirname=<path to download> -j | ||
``` | ||
|
||
=== "Calibration" | ||
|
||
### Get Calibration Dataset | ||
``` | ||
cm run script --tags=get,dataset,mlperf,inference,llama3,_calibration --outdirname=<path to download> -j | ||
``` | ||
|
||
## Model | ||
The benchmark implementation run command will automatically download the required model and do the necessary conversions. In case you want to only download the official model, you can use the below commands. | ||
|
||
Get the Official MLPerf LLAMA3.1-405b Model | ||
|
||
=== "Pytorch" | ||
|
||
### Pytorch | ||
``` | ||
cm run script --tags=get,ml-model,llama3 --outdirname=<path to download> --hf_token=<huggingface access token> -j | ||
``` | ||
|
||
!!! tip | ||
|
||
Downloading llama3.1-405B model from Hugging Face will require an [**access token**](https://huggingface.co/settings/tokens) which could be generated for your account. Additionally, ensure that your account has access to the [llama3.1-405B](https://huggingface.co/meta-llama/Llama-3.1-405B-Instruct) model. | ||
|
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Original file line number | Diff line number | Diff line change |
---|---|---|
@@ -0,0 +1,13 @@ | ||
--- | ||
hide: | ||
- toc | ||
--- | ||
|
||
# Text Summarization using LLAMA3_1-405b | ||
|
||
=== "MLCommons-Python" | ||
## MLPerf Reference Implementation in Python | ||
|
||
{{ mlperf_inference_implementation_readme (4, "llama3_1-405b-99", "reference", devices=["CPU","CUDA"]) }} | ||
|
||
{{ mlperf_inference_implementation_readme (4, "llama3_1-405b-99.9", "reference", devices=["CPU","CUDA"]) }} |
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
This file contains bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters