Skip to content

Commit 005c4c8

Browse files
Hotfix/docs (#19)
Updating documentation to reflect current purpose of this repo.
1 parent 659895b commit 005c4c8

File tree

35 files changed

+296
-314
lines changed

35 files changed

+296
-314
lines changed

README.md

Lines changed: 5 additions & 4 deletions
Original file line numberDiff line numberDiff line change
@@ -1,14 +1,15 @@
11
# Model Zoo for Intel® Architecture
22

3-
This repository contains **links to pre-trained models, benchmarking scripts, best practices, and step-by-step tutorials** for many popular open-source machine learning models optimized by Intel to run on Intel® Xeon® Scalable processors.
3+
This repository contains **links to pre-trained models, sample scripts, best practices, and step-by-step tutorials** for many popular open-source machine learning models optimized by Intel to run on Intel® Xeon® Scalable processors.
44

55
## Purpose of the Model Zoo
66

77
- Demonstrate the AI workloads and deep learning models Intel has optimized and validated to run on Intel hardware
88
- Show how to efficiently execute, train, and deploy Intel-optimized models
9-
- Make it easy to benchmark model performance on Intel hardware
109
- Make it easy to get started running Intel-optimized models on Intel hardware in the cloud or on bare metal
1110

11+
***DISCLAIMER: These scripts are not intended for benchmarking Intel platforms. Please see [https://www.intel.ai/blog](https://www.intel.ai/blog) for any information on performance and/or benchmarking information on specific Intel platforms.***
12+
1213
## How to Use the Model Zoo
1314

1415
### Getting Started
@@ -17,10 +18,10 @@ This repository contains **links to pre-trained models, benchmarking scripts, be
1718

1819
### Directory Structure
1920
The Model Zoo is divided into four main directories:
20-
- **[benchmarks](/benchmarks)**: Look here for benchmarking scripts and complete instructions on downloading and benchmarking each Intel-optimized pre-trained model.
21+
- **[benchmarks](/benchmarks)**: Look here for sample scripts and complete instructions on downloading and running each Intel-optimized pre-trained model.
2122
- **[docs](/docs)**: General best practices and detailed tutorials for a selection of models and frameworks can be found in this part of the repo.
2223
- **[models](/models)**: This directory contains optimized model code that has not yet been upstreamed to its respective official repository, such as dataset processing routines.
23-
There are no user-friendly READMEs in this directory, but many supporting modules used for benchmarking are here.
24+
There are no user-friendly READMEs in this directory, but many supporting modules are here.
2425
- **[tests](/tests)**: Look here for unit tests and information on how to run them.
2526

2627
The benchmarks, models, and docs folders share a common structure. Each model (or document) is organized first by *use case* and then by *framework*.

benchmarks/README.md

Lines changed: 5 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -1,10 +1,10 @@
1-
# Benchmark scripts
1+
# Model Zoo Scripts
22

33
Training and inference scripts with Intel-optimized MKL
44

55
## Prerequisites
66

7-
The benchmarking scripts can be run on Linux and require the following
7+
The model scripts can be run on Linux and require the following
88
dependencies to be installed:
99
* [Docker](https://docs.docker.com/install/)
1010
* [Python](https://www.python.org/downloads/) 2.7 or later
@@ -13,7 +13,7 @@ dependencies to be installed:
1313

1414
## Use Cases
1515

16-
| Use Case | Framework | Model | Mode | Benchmarking Instructions |
16+
| Use Case | Framework | Model | Mode | Instructions |
1717
| -----------------------| --------------| ------------------- | --------- |------------------------------|
1818
| Adversarial Networks | TensorFlow | [DCGAN](https://arxiv.org/pdf/1511.06434.pdf) | Inference | [FP32](adversarial_networks/tensorflow/dcgan/README.md#fp32-inference-instructions) |
1919
| Content Creation | TensorFlow | [DRAW](https://arxiv.org/pdf/1502.04623.pdf) | Inference | [FP32](content_creation/tensorflow/draw/README.md#fp32-inference-instructions) |
@@ -31,9 +31,9 @@ dependencies to be installed:
3131
| Language Translation | TensorFlow | [GNMT](https://arxiv.org/pdf/1609.08144.pdf) | Inference | [FP32](language_translation/tensorflow/gnmt/README.md#fp32-inference-instructions) |
3232
| Language Translation | TensorFlow | [Transformer Language](https://arxiv.org/pdf/1706.03762.pdf)| Inference | [FP32](language_translation/tensorflow/transformer_language/README.md#fp32-inference-instructions) |
3333
| Language Translation | TensorFlow | [Transformer_LT_Official ](https://arxiv.org/pdf/1706.03762.pdf)| Inference | [FP32](language_translation/tensorflow/transformer_lt_official/README.md#fp32-inference-instructions) |
34-
| Object Detection | TensorFlow | [R-FCN](https://arxiv.org/pdf/1605.06409.pdf) | Inference | [Int8](object_detection/tensorflow/rfcn/README.md#int8-inference-instructions) [FP32](object_detection/tensorflow/rfcn/README.md#fp32-inference-instructions) |
34+
| Object Detection | TensorFlow | [R-FCN](https://arxiv.org/pdf/1605.06409.pdf) | Inference | [FP32](object_detection/tensorflow/rfcn/README.md#fp32-inference-instructions) |
3535
| Object Detection | TensorFlow | [Faster R-CNN](https://arxiv.org/pdf/1506.01497.pdf) | Inference | [Int8](object_detection/tensorflow/faster_rcnn/README.md#int8-inference-instructions) [FP32](object_detection/tensorflow/faster_rcnn/README.md#fp32-inference-instructions) |
36-
| Object Detection | TensorFlow | [SSD-MobileNet](https://arxiv.org/pdf/1704.04861.pdf) | Inference | [Int8](object_detection/tensorflow/ssd-mobilenet/README.md#int8-inference-instructions) [FP32](object_detection/tensorflow/ssd-mobilenet/README.md#fp32-inference-instructions) |
36+
| Object Detection | TensorFlow | [SSD-MobileNet](https://arxiv.org/pdf/1704.04861.pdf) | Inference | [FP32](object_detection/tensorflow/ssd-mobilenet/README.md#fp32-inference-instructions) |
3737
| Object Detection | TensorFlow | [SSD-ResNet34](https://arxiv.org/pdf/1512.02325.pdf) | Inference | [FP32](object_detection/tensorflow/ssd-resnet34/README.md#fp32-inference-instructions) |
3838
| Recommendation | TensorFlow | [NCF](https://arxiv.org/pdf/1708.05031.pdf) | Inference | [FP32](recommendation/tensorflow/ncf/README.md#fp32-inference-instructions) |
3939
| Recommendation | TensorFlow | [Wide & Deep Large Dataset](https://arxiv.org/pdf/1606.07792.pdf) | Inference | [Int8](recommendation/tensorflow/wide_deep_large_ds/README.md#int8-inference-instructions) [FP32](recommendation/tensorflow/wide_deep_large_ds/README.md#fp32-inference-instructions) |

benchmarks/adversarial_networks/tensorflow/dcgan/README.md

Lines changed: 5 additions & 6 deletions
Original file line numberDiff line numberDiff line change
@@ -4,7 +4,7 @@ This document has instructions for how to run DCGAN for the
44
following modes/precisions:
55
* [FP32 inference](#fp32-inference-instructions)
66

7-
Benchmarking instructions and scripts for model training and inference.
7+
Script instructions for model training and inference.
88

99
## FP32 Inference Instructions
1010

@@ -35,19 +35,18 @@ repository:
3535
$ git clone https://github.com/IntelAI/models.git
3636
```
3737

38-
This repository includes launch scripts for running benchmarks and the
39-
an optimized version of the DCGAN model code.
38+
This repository includes launch scripts for running an optimized version of the DCGAN model code.
4039

4140
5. Navigate to the `benchmarks` directory in your local clone of
4241
the [intelai/models](https://github.com/IntelAI/models) repo from step 4.
4342
The `launch_benchmark.py` script in the `benchmarks` directory is
44-
used for starting a benchmarking run in a optimized TensorFlow docker
43+
used for starting a model script run in a optimized TensorFlow docker
4544
container. It has arguments to specify which model, framework, mode,
4645
precision, and docker image to use, along with your path to the external model directory
4746
for `--model-source-dir` (from step 1) `--data-location` (from step 2), and `--checkpoint` (from step 3).
4847

4948

50-
Run benchmarking for throughput and latency with `--batch-size=100` :
49+
Run the model script for batch and online inference with `--batch-size=100` :
5150
```
5251
$ cd /home/<user>/models/benchmarks
5352
@@ -66,7 +65,7 @@ $ python launch_benchmark.py \
6665

6766
5. Log files are located at the value of `--output-dir`.
6867

69-
Below is a sample log file tail when running benchmarking for throughput:
68+
Below is a sample log file tail when running for batch inference:
7069
```
7170
Batch size: 100
7271
Batches number: 500

benchmarks/content_creation/tensorflow/draw/README.md

Lines changed: 9 additions & 9 deletions
Original file line numberDiff line numberDiff line change
@@ -18,7 +18,7 @@ modes/precisions:
1818
```
1919

2020
The mnist directory will be passed as the dataset location when we
21-
run the benchmarking script in step 4.
21+
run the model script in step 4.
2222

2323
2. Download and extract the pretrained model:
2424
```
@@ -27,21 +27,21 @@ modes/precisions:
2727
```
2828

2929
3. Clone this [intelai/models](https://github.com/IntelAI/models) repo,
30-
which contains the scripts that we will be using to run benchmarking
31-
for DRAW. After the clone has completed, navigate to the `benchmarks`
30+
which contains the DRAW model scripts.
31+
After the clone has completed, navigate to the `benchmarks`
3232
directory in the repository.
3333

3434
```
3535
$ git clone https://github.com/IntelAI/models.git
3636
$ cd models/benchmarks
3737
```
3838

39-
4. Run benchmarking for either throughput or latency using the commands
39+
4. Run the model for either batch or online inference using the commands
4040
below. Replace in the path to the `--data-location` with your `mnist`
4141
dataset directory from step 1 and the `--checkpoint` files that you
4242
downloaded and extracted in step 2.
4343

44-
* Run benchmarking for latency (with `--batch-size 1`):
44+
* Run DRAW for online inference (with `--batch-size 1`):
4545
```
4646
python launch_benchmark.py \
4747
--precision fp32 \
@@ -54,7 +54,7 @@ modes/precisions:
5454
--batch-size 1 \
5555
--socket-id 0
5656
```
57-
* Run benchmarking for throughput (with `--batch-size 100`):
57+
* Run DRAW for batch inference (with `--batch-size 100`):
5858
```
5959
python launch_benchmark.py \
6060
--precision fp32 \
@@ -70,9 +70,9 @@ modes/precisions:
7070
Note that the `--verbose` or `--output-dir` flag can be added to any of the above
7171
commands to get additional debug output or change the default output location.
7272
73-
5. The log files for each benchmarking run are saved at the value of `--output-dir`.
73+
5. The log files for each run are saved at the value of `--output-dir`.
7474
75-
* Below is a sample log file tail when benchmarking latency:
75+
* Below is a sample log file tail when testing online inference:
7676
```
7777
...
7878
Elapsed Time 0.006622
@@ -88,7 +88,7 @@ modes/precisions:
8888
Log location outside container: {--output-dir value}/benchmark_draw_inference_fp32_20190123_012947.log
8989
```
9090
91-
* Below is a sample log file tail when benchmarking throughput:
91+
* Below is a sample log file tail when testing batch inference:
9292
```
9393
Elapsed Time 0.028355
9494
Elapsed Time 0.028221

benchmarks/face_detection_and_alignment/tensorflow/facenet/README.md

Lines changed: 9 additions & 11 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,7 @@ This document has instructions for how to run FaceNet for the
44
following modes/precisions:
55
* [FP32 inference](#fp32-inference-instructions)
66

7-
Benchmarking instructions and scripts for model training and inference
8-
other precisions are coming later.
7+
Script instructions for model training and inference for other precisions are coming later.
98

109
## FP32 Inference Instructions
1110

@@ -37,18 +36,17 @@ Instructions for downloading the dataset and converting it can be found in the d
3736
5. Navigate to the `benchmarks` directory in your local clone of
3837
the [intelai/models](https://github.com/IntelAI/models) repo from step 2.
3938
The `launch_benchmark.py` script in the `benchmarks` directory is
40-
used for starting a benchmarking run in a optimized TensorFlow docker
39+
used for starting a model run in a optimized TensorFlow docker
4140
container. It has arguments to specify which model, framework, mode,
4241
precision, and docker image.
4342

4443
Substitute in your own `--checkpoint` pretrained model file path (from step 3),
4544
and `--data-location` (from step 4).
4645

47-
FaceNet can be run for latency benchmarking, throughput
48-
benchmarking, or accuracy. Use one of the following examples below,
49-
depending on your use case.
46+
FaceNet can be run for testing online inference, batch inference, or accuracy.
47+
Use one of the following examples below, depending on your use case.
5048

51-
* For latency (using `--batch-size 1`):
49+
* For online inference (using `--batch-size 1`):
5250

5351
```
5452
python launch_benchmark.py \
@@ -63,7 +61,7 @@ python launch_benchmark.py \
6361
--model-source-dir /home/<user>/facenet/ \
6462
--docker-image intelaipg/intel-optimized-tensorflow:latest-devel-mkl
6563
```
66-
Example log tail when benchmarking for latency:
64+
Example log tail for online inference:
6765
```
6866
Batch 979 elapsed Time 0.0297989845276
6967
Batch 989 elapsed Time 0.029657125473
@@ -85,7 +83,7 @@ Ran inference with batch size 1
8583
Log location outside container: {--output-dir value}/benchmark_facenet_inference_fp32_20190328_205911.log
8684
```
8785

88-
* For throughput (using `--batch-size 100`):
86+
* For batch inference (using `--batch-size 100`):
8987

9088
```
9189
python launch_benchmark.py \
@@ -100,7 +98,7 @@ python launch_benchmark.py \
10098
--model-source-dir /home/<user>/facenet/ \
10199
--docker-image intelaipg/intel-optimized-tensorflow:latest-devel-mkl
102100
```
103-
Example log tail when benchmarking for throughput:
101+
Example log tail for batch inference:
104102
```
105103
Batch 219 elapsed Time 0.446497917175
106104
Batch 229 elapsed Time 0.422048091888
@@ -134,7 +132,7 @@ python launch_benchmark.py \
134132
--model-source-dir /home/<user>/facenet/ \
135133
--docker-image intelaipg/intel-optimized-tensorflow:latest-devel-mkl
136134
```
137-
Example log tail when benchmarking for accuracy:
135+
Example log tail for accuracy:
138136
```
139137
Batch 219 elapsed Time 0.398629188538
140138
Batch 229 elapsed Time 0.354953050613

benchmarks/face_detection_and_alignment/tensorflow/mtcc/README.md

Lines changed: 4 additions & 5 deletions
Original file line numberDiff line numberDiff line change
@@ -4,8 +4,7 @@ This document has instructions for how to run MTCC for the
44
following modes/precisions:
55
* [FP32 inference](#fp32-inference-instructions)
66

7-
Benchmarking instructions and scripts for the MTCC model training and inference
8-
other precisions are coming later.
7+
Instructions for MTCC model training and inference for other precisions are coming later.
98

109
## FP32 Inference Instructions
1110

@@ -33,7 +32,7 @@ other precisions are coming later.
3332
```
3433

3534
4. Clone the [intelai/models](https://github.com/intelai/models) repo.
36-
This repo has the launch script for running benchmarking.
35+
This repo has the launch script for running models.
3736

3837
```
3938
$ git clone https://github.com/IntelAI/models.git
@@ -43,7 +42,7 @@ This repo has the launch script for running benchmarking.
4342
5. Run the `launch_benchmark.py` script from the intelai/models repo with the appropriate parameters including: the `--model-source-dir` from step 1, `--data-location` from step 2,
4443
and the `--checkpoint` from step 3.
4544

46-
Run benchmarking:
45+
Run:
4746
```
4847
$ cd /home/<user>/models/benchmarks
4948
@@ -61,7 +60,7 @@ Run benchmarking:
6160

6261
6. The log file is saved to the value of `--output-dir`.
6362

64-
Below is a sample log file tail when running benchmarking for throughput,latency and accuracy:
63+
Below is a sample log file tail when running for batch inference, online inference, and accuracy:
6564

6665
```
6766
time cost 0.459 pnet 0.166 rnet 0.144 onet 0.149

0 commit comments

Comments
 (0)