Skip to content

Commit ee99891

Browse files
SW publisherJenkins
SW publisher
authored and
Jenkins
committed
Model-References content for 1.20.0
Signed-off-by: SW publisher <[email protected]>
1 parent e14ce79 commit ee99891

File tree

99 files changed

+572
-12703
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

99 files changed

+572
-12703
lines changed

MLPERF4.0/Inference/llama/README.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -52,7 +52,7 @@ docker run --privileged --security-opt seccomp=unconfined \
5252
-v $INTEL_HABANALABS_DIR:/root/Intel-HabanaLabs/ \
5353
--cap-add=sys_nice --cap-add=SYS_PTRACE \
5454
--user root --workdir=/root --net=host \
55-
--ulimit memlock=-1:-1 vault.habana.ai/gaudi-docker/1.19.0/ubuntu22.04/habanalabs/pytorch-installer-2.5.1:latest
55+
--ulimit memlock=-1:-1 vault.habana.ai/gaudi-docker/1.20.0/ubuntu22.04/habanalabs/pytorch-installer-2.6.0:latest
5656
```
5757
```bash
5858
docker exec -it mlperf-intel-habanalabs bash
@@ -164,6 +164,6 @@ The Quantization Toolkit is described in the [Intel Gaudi documentation](https:/
164164

165165
## Supported Configurations
166166

167-
| Validated on | Intel Gaudi Software Version | Framework Version(s) | Mode |
168-
| :----------: | :--------------------------: | :------------------: | :------: |
169-
| Gaudi 2 | 1.19.0 | PyTorch 2.5.1 | Inference |
167+
| Validated on | Intel Gaudi Software Version | Framework Version(s) | Mode |
168+
| :----------: | :--------------------------: | :------------------: | :-------: |
169+
| Gaudi 2 | 1.20.0 | PyTorch 2.6.0 | Inference |
+1-1
Original file line numberDiff line numberDiff line change
@@ -1,4 +1,4 @@
11
*.Offline.min_query_count = 98304
22

3-
*.Server.target_qps = 21.1
3+
*.Server.target_qps = 20.2
44
*.Server.min_query_count = 24576

MLPERF4.0/Inference/llama/setup_tgi.sh

+3-2
Original file line numberDiff line numberDiff line change
@@ -23,15 +23,16 @@ pip install pip --upgrade
2323
# don't try to overwrite torch
2424
grep -v "torch==" requirements.txt | pip install --no-deps -r /dev/stdin
2525
pip install -e .
26+
# this stopped to be installed by TGI but is still required:
27+
pip install outlines==0.0.36
2628
cd ..
2729
# build router
2830
cd router
2931
cargo install --locked --path .
3032
cd ..
3133
# build launcher
3234
cd launcher
33-
cargo install --locked --path .
35+
cargo install --locked --path .
3436
cd ..
3537
popd
36-
3738
pip list

MLPERF4.0/Inference/stable-diffusion-xl/README.md

+3-3
Original file line numberDiff line numberDiff line change
@@ -131,9 +131,9 @@ All necessary environmental variables are enabled by default.
131131

132132
## Supported Configurations
133133

134-
| Validated on | Intel Gaudi Software Version | Framework Version(s) | Mode |
135-
| :----------: | :--------------------------: | :------------------: | :------: |
136-
| Gaudi 2 | 1.18.0 | PyTorch 2.4.0 | Inference |
134+
| Validated on | Intel Gaudi Software Version | Framework Version(s) | Mode |
135+
| :----------: | :--------------------------: | :------------------: | :-------: |
136+
| Gaudi 2 | 1.20.0 | PyTorch 2.6.0 | Inference |
137137

138138
## Changelog
139139

MLPERF4.0/Training/benchmarks/llm_finetune/README.md

+2-2
Original file line numberDiff line numberDiff line change
@@ -71,7 +71,7 @@ NOTE: The below is only an example value. Replace [Intel_Gaudi_software_version]
7171
7272
### Download Data and Model
7373
74-
MLCommons hosts the model for download exclusively by MLCommons Members. You must first agree to the [confidentiality notice](https://docs.google.com/forms/d/e/1FAIpQLSc_8VIvRmXM3I8KQaYnKf7gy27Z63BBoI_I1u02f4lw6rBp3g/viewform), then follow the [link[(https://drive.google.com/drive/folders/11tBZvvrh0FCm3XuR5E849K42TqftYdUF)] to a directory containing [Rclone download instructions](https://docs.google.com/document/d/1Yp2T_TsVfg8uEoEv0wa-dGP4R7r1EOHucTvDNWznWzE/edit#heading=h.at8a3matgbrk). Follow steps 1-3 to install and activate Rclone. Finally, download the model to the desired download directory (default ./models):
74+
MLCommons hosts the model for download exclusively by MLCommons Members. You must first agree to the [confidentiality notice](https://docs.google.com/forms/d/e/1FAIpQLSc_8VIvRmXM3I8KQaYnKf7gy27Z63BBoI_I1u02f4lw6rBp3g/viewform), then follow the [link](https://drive.google.com/drive/folders/11tBZvvrh0FCm3XuR5E849K42TqftYdUF) to a directory containing [Rclone download instructions](https://docs.google.com/document/d/1Yp2T_TsVfg8uEoEv0wa-dGP4R7r1EOHucTvDNWznWzE/edit#heading=h.at8a3matgbrk). Follow steps 1-3 to install and activate Rclone. Finally, download the model to the desired download directory (default ./models):
7575
Log into mlperf4.0 container and run:
7676
```bash
7777
rclone copy mlc-llama2:Llama2-70b-fused-qkv-mlperf /root/model/Llama2-70b-fused-qkv-mlperf -P
@@ -107,4 +107,4 @@ cp /root/MLPERF/Intel-HabanaLabs/benchmarks/llm_finetune/config.json /root/model
107107

108108
| Validated on | Intel Gaudi Software Version | Framework Version(s) | Mode |
109109
| :----------: | :--------------------------: | :------------------: | :------: |
110-
| Gaudi 2 | 1.18.0 | PyTorch 2.4.0 | Training |
110+
| Gaudi 2 | 1.20.0 | PyTorch 2.6.0 | Training |

PyTorch/computer_vision/classification/dino/LICENSE

-203
This file was deleted.

0 commit comments

Comments
 (0)