Skip to content

Commit ac4f30c

Browse files
committed
Sync master with v1.8.0 tag
2 parents 7e7f484 + 49c1ef2 commit ac4f30c

File tree

136 files changed

+11038
-2971
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

136 files changed

+11038
-2971
lines changed

Contribute.md

+4-4
Original file line numberDiff line numberDiff line change
@@ -71,10 +71,10 @@ required:
7171
first step where you setup the directories for your model. In this
7272
function, add commands to install any third-party dependencies within
7373
an `if [ ${NOINSTALL} != "True" ]; then` conditional block. The
74-
purpose of the `NOINSTALL` flag is to be able to skip the installs
75-
for quicker iteration when running on bare metal or debugging. If
76-
your model requires the `PYTHONPATH` environment variable to be setup
77-
to find model code or dependencies, that should be done in the
74+
purpose of the `--noinstall` flag or `NOINSTALL` env var is to be able
75+
to skip the installs for quicker iteration when running on bare metal
76+
or debugging. If your model requires the `PYTHONPATH` environment variable
77+
to be setup to find model code or dependencies, that should be done in the
7878
model's function. Next, setup the command that will be run. The
7979
standard launch script args are already added to the `CMD` variable,
8080
so your model function will only need to add on more args if you have

Jenkinsfile

+2-4
Original file line numberDiff line numberDiff line change
@@ -15,11 +15,9 @@ node('skx') {
1515
# don't know OS, so trying both apt-get and yum install
1616
sudo apt-get clean || sudo yum update -y
1717
sudo apt-get update -y || sudo yum install -y epel-release
18-
sudo apt-get install -y python-dev python-pip || sudo yum install -y python2-devel python2-pip
1918
sudo apt-get install -y python3-dev python3-pip || sudo yum install -y python36-devel python36-pip
2019
2120
# virtualenv 16.3.0 is broken do not use it
22-
python2 -m pip install --no-cache-dir --user --upgrade pip==19.0.3 virtualenv!=16.3.0 tox
2321
python3 -m pip install --no-cache-dir --user --upgrade pip==19.0.3 virtualenv!=16.3.0 tox
2422
"""
2523
}
@@ -29,7 +27,7 @@ node('skx') {
2927
set -e
3028
3129
cd intel-models
32-
~/.local/bin/tox -e py2.7-flake8 -e py3-flake8
30+
~/.local/bin/tox -e py3-flake8
3331
"""
3432
}
3533
stage('Unit tests') {
@@ -38,7 +36,7 @@ node('skx') {
3836
set -e
3937
4038
cd intel-models
41-
~/.local/bin/tox -e py2.7-py.test -e py3-py.test
39+
~/.local/bin/tox -e py3-py.test
4240
"""
4341
}
4442
// put benchmarks here later

Makefile

-22
Original file line numberDiff line numberDiff line change
@@ -37,8 +37,6 @@ venv: $(ACTIVATE)
3737
@echo -n "Using "
3838
@. $(ACTIVATE) && python --version
3939

40-
venv2: venv
41-
4240
venv3: PY_VERSION=3
4341
venv3: $(ACTIVATE)
4442
@echo -n "Using "
@@ -49,33 +47,13 @@ tox:
4947

5048
lint:
5149
@echo "Running style check..."
52-
tox -e py2.7-flake8 -e py3-flake8
53-
54-
lint2:
55-
@echo "Running style check..."
56-
tox -e py2.7-flake8
57-
58-
lint3:
59-
@echo "Running style check python 3 "
6050
tox -e py3-flake8
6151

6252
unit_test:
6353
@echo "Running unit tests..."
64-
tox -e py2.7-py.test -e py3-py.test
65-
66-
unit_test2:
67-
@echo "Running unit tests python 2..."
68-
tox -e py2.7-py.test
69-
70-
unit_test3:
71-
@echo "Running unit tests python 3..."
7254
tox -e py3-py.test
7355

7456
test: lint unit_test
7557

76-
test2: lint2 unit_test2
77-
78-
test3: lint3 unit_test3
79-
8058
clean:
8159
rm -rf .venv .venv3 .tox

README.md

-2
Original file line numberDiff line numberDiff line change
@@ -34,7 +34,5 @@ We hope this structure is intuitive and helps you find what you are looking for;
3434

3535
![Repo Structure](repo_structure.png)
3636

37-
*Note: For model quantization and optimization tools, see [https://github.com/IntelAI/tools](https://github.com/IntelAI/tools)*.
38-
3937
## How to Contribute
4038
If you would like to add a new benchmarking script, please use [this guide](/Contribute.md).

benchmarks/README.md

+12-1
Original file line numberDiff line numberDiff line change
@@ -25,16 +25,27 @@ dependencies to be installed:
2525
| Image Recognition | TensorFlow | [ResNet 50v1.5](https://github.com/tensorflow/models/tree/master/official/resnet) | Training | [FP32](image_recognition/tensorflow/resnet50v1_5/README.md#fp32-training-instructions) [BFloat16**](image_recognition/tensorflow/resnet50v1_5/README.md#bfloat16-training-instructions)|
2626
| Language Modeling | TensorFlow | [BERT](https://arxiv.org/pdf/1810.04805.pdf) | Inference | [FP32](language_modeling/tensorflow/bert_large/README.md#fp32-inference-instructions) [BFloat16**](language_modeling/tensorflow/bert_large/README.md#bfloat16-inference-instructions) |
2727
| Language Modeling | TensorFlow | [BERT](https://arxiv.org/pdf/1810.04805.pdf) | Training | [FP32](language_modeling/tensorflow/bert_large/README.md#fp32-training-instructions) [BFloat16**](language_modeling/tensorflow/bert_large/README.md#bfloat16-training-instructions) |
28+
| Language Translation | TensorFlow | [BERT](https://arxiv.org/pdf/1810.04805.pdf) | Inference | [FP32](language_translation/tensorflow/bert/README.md#fp32-inference-instructions) |
2829
| Language Translation | TensorFlow | [GNMT*](https://arxiv.org/pdf/1609.08144.pdf) | Inference | [FP32](language_translation/tensorflow/mlperf_gnmt/README.md#fp32-inference-instructions) |
29-
| Reinforcement | TensorFlow | [MiniGo](https://arxiv.org/abs/1712.01815.pdf) | Training | [FP32](reinforcement/tensorflow/minigo/README.md#fp32-training-instructions)|
3030
| Language Translation | TensorFlow | [Transformer_LT_Official ](https://arxiv.org/pdf/1706.03762.pdf)| Inference | [FP32](language_translation/tensorflow/transformer_lt_official/README.md#fp32-inference-instructions) |
3131
| Language Translation | TensorFlow | [Transformer_LT_mlperf ](https://arxiv.org/pdf/1706.03762.pdf)| Training | [FP32](language_translation/tensorflow/transformer_mlperf/README.md#fp32-training-instructions) [BFloat16**](language_translation/tensorflow/transformer_mlperf/README.md#bfloat16-training-instructions) |
3232
| Object Detection | TensorFlow | [R-FCN](https://arxiv.org/pdf/1605.06409.pdf) | Inference | [Int8](object_detection/tensorflow/rfcn/README.md#int8-inference-instructions) [FP32](object_detection/tensorflow/rfcn/README.md#fp32-inference-instructions) |
3333
| Object Detection | TensorFlow | [SSD-MobileNet*](https://arxiv.org/pdf/1704.04861.pdf) | Inference | [Int8](object_detection/tensorflow/ssd-mobilenet/README.md#int8-inference-instructions) [FP32](object_detection/tensorflow/ssd-mobilenet/README.md#fp32-inference-instructions) |
3434
| Object Detection | TensorFlow | [SSD-ResNet34*](https://arxiv.org/pdf/1512.02325.pdf) | Inference | [Int8](object_detection/tensorflow/ssd-resnet34/README.md#int8-inference-instructions) [FP32](object_detection/tensorflow/ssd-resnet34/README.md#fp32-inference-instructions) |
3535
| Object Detection | TensorFlow | [SSD-ResNet34](https://arxiv.org/pdf/1512.02325.pdf) | Training | [FP32](object_detection/tensorflow/ssd-resnet34/README.md#fp32-training-instructions) [BFloat16**](object_detection/tensorflow/ssd-resnet34/README.md#bf16-training-instructions) |
3636
| Recommendation | TensorFlow | [Wide & Deep Large Dataset](https://arxiv.org/pdf/1606.07792.pdf) | Inference | [Int8](recommendation/tensorflow/wide_deep_large_ds/README.md#int8-inference-instructions) [FP32](recommendation/tensorflow/wide_deep_large_ds/README.md#fp32-inference-instructions) |
37+
| Recommendation | TensorFlow | [Wide & Deep Large Dataset](https://arxiv.org/pdf/1606.07792.pdf) | Training | [FP32](recommendation/tensorflow/wide_deep_large_ds/README.md#fp32-training-instructions) |
3738
| Recommendation | TensorFlow | [Wide & Deep](https://arxiv.org/pdf/1606.07792.pdf) | Inference | [FP32](recommendation/tensorflow/wide_deep/README.md#fp32-inference-instructions) |
39+
| Reinforcement | TensorFlow | [MiniGo](https://arxiv.org/abs/1712.01815.pdf) | Training | [FP32](reinforcement/tensorflow/minigo/README.md#fp32-training-instructions)|
40+
41+
## TensorFlow Serving Use Cases
42+
43+
| Use Case | Framework | Model | Mode | Instructions |
44+
| -----------------------| --------------| ------------------- | --------- |------------------------------|
45+
| Image Recognition | TensorFlow Serving | [Inception V3](https://arxiv.org/pdf/1512.00567.pdf) | Inference | [FP32](image_recognition/tensorflow_serving/inceptionv3/README.md#fp32-inference-instructions) |
46+
| Image Recognition | TensorFlow Serving | [ResNet 50v1.5](https://github.com/tensorflow/models/tree/master/official/resnet) | Inference | [FP32](image_recognition/tensorflow_serving/resnet50v1_5/README.md#fp32-inference-instructions) |
47+
| Language Translation | TensorFlow Serving | [Transformer_LT_Official](https://arxiv.org/pdf/1706.03762.pdf) | Inference | [FP32](language_translation/tensorflow_serving/transformer_lt_official/README.md#fp32-inference-instructions) |
48+
| Object Detection | TensorFlow Serving | [SSD-MobileNet](https://arxiv.org/pdf/1704.04861.pdf) | Inference | [FP32](object_detection/tensorflow_serving/ssd-mobilenet/README.md#fp32-inference-instructions) |
3849

3950
*Means the model belongs to [MLPerf](https://mlperf.org/) models and will be supported long-term.
4051

benchmarks/common/base_benchmark_util.py

+23
Original file line numberDiff line numberDiff line change
@@ -97,6 +97,11 @@ def _define_args(self):
9797
help="Specify how many MPI processes to launch per socket",
9898
dest="num_mpi", default=1)
9999

100+
self._common_arg_parser.add_argument(
101+
"--mpi_hostnames",
102+
help="Specify MPI hostnames string of the form --mpi_hostnames host1,host2,host3",
103+
dest="mpi_hostnames", default=None)
104+
100105
self._common_arg_parser.add_argument(
101106
"-d", "--data-location",
102107
help="Specify the location of the data. If this parameter is not "
@@ -188,6 +193,18 @@ def _define_args(self):
188193
"with --accuracy-only and --mode=inference.",
189194
dest="output_results", action="store_true")
190195

196+
self._common_arg_parser.add_argument(
197+
"--optimized-softmax",
198+
help="Use tf.nn.softmax as opposed to basic math ops",
199+
dest="optimized_softmax", choices=["True", "False"],
200+
default=True)
201+
202+
self._common_arg_parser.add_argument(
203+
"--experimental-gelu",
204+
help="use tf.nn.gelu as opposed to basic math ops",
205+
dest="experimental_gelu", choices=["True", "False"],
206+
default=False)
207+
191208
# Note this can't be a normal boolean flag, because we need to know when the user
192209
# does not explicitly set the arg value so that we can apply the appropriate
193210
# default value, depending on the the precision.
@@ -222,6 +239,12 @@ def _define_args(self):
222239
"'models/benchmarks/common/tensorflow/logs' if no path is specified.",
223240
default="/models/benchmarks/common/tensorflow/logs")
224241

242+
self._common_arg_parser.add_argument(
243+
"--tf-serving-version",
244+
help="TF serving version to run the script with"
245+
"'master' if no value is specified.",
246+
default="master")
247+
225248
# Allow for additional command line args after --
226249
self._common_arg_parser.add_argument(
227250
"model_args", nargs="*",

benchmarks/common/base_model_init.py

+22-2
Original file line numberDiff line numberDiff line change
@@ -57,8 +57,28 @@ def __init__(self, args, custom_args=[], platform_util=None):
5757
if not platform_util:
5858
raise ValueError("Did not find any platform info.")
5959

60-
# Invoke mpirun if mpi_num_processes env is not None
61-
if os.environ["MPI_NUM_PROCESSES"] != "None":
60+
# use case: bare-metal with openmpi, horovod and multi-node
61+
if os.environ["MPI_HOSTNAMES"] != "None" and ("DOCKER" not in os.environ or os.environ["DOCKER"] == "False"):
62+
if os.environ["MPI_NUM_PROCESSES"] != "None":
63+
try:
64+
# slots per host calculation using MPI_NUM_PROCESSES and number of hosts
65+
host_names = os.environ["MPI_HOSTNAMES"]
66+
number_of_hosts = len(host_names.split(','))
67+
slots_per_host = int(int(os.environ["MPI_NUM_PROCESSES"]) / number_of_hosts)
68+
host_names = ",".join([host + ":" + str(slots_per_host) for host in host_names.split(',')])
69+
# see the [examples](https://horovod.readthedocs.io/en/latest/mpirun.html) for the mca flags
70+
self.python_exe = "mpirun " + " -x LD_LIBRARY_PATH " + " -x PYTHONPATH " \
71+
+ " --allow-run-as-root -n " + os.environ["MPI_NUM_PROCESSES"] + " -H " + host_names \
72+
+ " -mca pml ob1 -mca btl ^openib -mca btl_tcp_if_exclude " \
73+
"lo,docker0 --bind-to none --map-by slot " \
74+
+ self.python_exe
75+
except Exception as exception:
76+
raise ValueError("Caught exception calculating slots per host {}".format(str(exception)))
77+
else:
78+
raise ValueError("MPI_NUM_PROCESSES is required for MPI_HOSTNAMES and will be split evenly across the "
79+
"hosts.")
80+
# use case: docker with openmpi, single-node, multi-instance
81+
elif os.environ["MPI_NUM_PROCESSES"] != "None":
6282
if os.environ["MPI_NUM_PROCESSES_PER_SOCKET"] == "1":
6383
# Map by socket using OpenMPI by default (PPS=1).
6484
self.python_exe = "mpirun --allow-run-as-root -n " + os.environ["MPI_NUM_PROCESSES"] \
Original file line numberDiff line numberDiff line change
@@ -0,0 +1,21 @@
1+
#!/usr/bin/env bash
2+
#
3+
# Copyright (c) 2020 Intel Corporation
4+
#
5+
# Licensed under the Apache License, Version 2.0 (the "License");
6+
# you may not use this file except in compliance with the License.
7+
# You may obtain a copy of the License at
8+
#
9+
# http://www.apache.org/licenses/LICENSE-2.0
10+
#
11+
# Unless required by applicable law or agreed to in writing, software
12+
# distributed under the License is distributed on an "AS IS" BASIS,
13+
# WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
14+
# See the License for the specific language governing permissions and
15+
# limitations under the License.
16+
#
17+
18+
# This file includes runtime installs for model containers
19+
20+
apt-get install numactl -y
21+

0 commit comments

Comments
 (0)