-
Notifications
You must be signed in to change notification settings - Fork 223
/
Copy pathfaster-rcnn-int8-inference_spec.yml
74 lines (74 loc) · 3.72 KB
/
faster-rcnn-int8-inference_spec.yml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
releases:
dockerfiles:
tag_specs:
- '{intel-tf}{object-detection}{faster-rcnn-int8-inference}'
tf_1.15.2_containers:
tag_specs:
- '{_TAG_PREFIX}{intel-tf}{object-detection}{faster-rcnn-int8-inference}'
slice_sets:
faster-rcnn-int8-inference:
- add_to_name: -faster-rcnn-int8-inference
args:
- PACKAGE_NAME=faster-rcnn-int8-inference
- TF_MODELS_BRANCH=tags/v1.12.0
dockerfile_subdirectory: model_containers
documentation:
docs:
- name: Title
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/title.md
- name: Description
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/description.md
- name: Download link
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/download.md
- name: Datasets
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/datasets.md
- name: Quick Start Scripts
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/quickstart.md
- name: Bare Metal
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/baremetal.md
- name: Docker
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/docker.md
- name: License
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8/.docs/license.md
name: README.md
text_replace:
<docker image>: intel/object-detection:tf-1.15.2-imz-2.0.0-faster-rcnn-int8-inference
<mode>: inference
<model name>: Faster_RCNN
<package dir>: faster_rcnn-int8-inference
<package name>: faster_rcnn-int8-inference.tar.gz
<package url>: https://storage.googleapis.com/intel-optimized-tensorflow/models/v2_0_0/faster-rcnn-int8-inference.tar.gz
<precision>: Int8
<use case>: object_detection
uri: models/quickstart/object_detection/tensorflow/faster_rcnn/inference/int8
downloads:
- destination: faster_rcnn_int8_pretrained_model.pb
source: https://storage.googleapis.com/intel-optimized-tensorflow/models/faster_rcnn_int8_pretrained_model.pb
files:
- destination: benchmarks/common
source: benchmarks/common
- destination: benchmarks/launch_benchmark.py
source: benchmarks/launch_benchmark.py
- destination: benchmarks/object_detection/__init__.py
source: benchmarks/object_detection/__init__.py
- destination: benchmarks/object_detection/tensorflow/__init__.py
source: benchmarks/object_detection/tensorflow/__init__.py
- destination: benchmarks/object_detection/tensorflow/faster_rcnn/README.md
source: benchmarks/object_detection/tensorflow/faster_rcnn/README.md
- destination: benchmarks/object_detection/tensorflow/faster_rcnn/__init__.py
source: benchmarks/object_detection/tensorflow/faster_rcnn/__init__.py
- destination: benchmarks/object_detection/tensorflow/faster_rcnn/inference/__init__.py
source: benchmarks/object_detection/tensorflow/faster_rcnn/inference/__init__.py
- destination: benchmarks/object_detection/tensorflow/faster_rcnn/inference/int8
source: benchmarks/object_detection/tensorflow/faster_rcnn/inference/int8
- destination: models/common
source: models/common
- destination: models/object_detection/tensorflow/faster_rcnn/inference/int8
source: models/object_detection/tensorflow/faster_rcnn/inference/int8
- destination: quickstart/common
source: quickstart/common
- destination: quickstart
source: quickstart/object_detection/tensorflow/faster_rcnn/inference/int8
partials:
- model_package
- entrypoint