|
1 | | -# Accelerating Deep Learning Inference for Model Zoo Workloads on Intel CPU and GPU |
| 1 | +# Accelerate Deep Learning Training and Inference for Model Zoo Workloads on Intel GPU |
2 | 2 |
|
3 | 3 | ## Introduction |
4 | | -This example shows the guideline to run Model Zoo workloads on Intel CPU and GPU with the optimizations from Intel® Extension for TensorFlow*, without any model code changes. |
| 4 | +This example shows the guideline to run Model Zoo workloads using TensorFlow* framework on Intel GPU with the optimizations from Intel® Extension for TensorFlow*. |
5 | 5 |
|
6 | | -## Prerequisites |
7 | | -For Intel CPU, refer to [Intel CPU software installation](../../docs/install/install_for_cpu.md#intel-cpu-software-installation). |
8 | | -For Intel GPU, refer to [Intel XPU software installation](../../docs/install/install_for_xpu.md#intel-gpu-software-installation). |
| 6 | +## Quick Start Guide |
9 | 7 |
|
10 | | -## Execute |
| 8 | +### Run Models in the Docker Container |
11 | 9 |
|
12 | | -### Prepare the Codes |
13 | | -```bash |
14 | | -git clone https://github.com/IntelAI/models |
15 | | -cd models |
16 | | -git checkout v2.8.0 |
17 | | -``` |
| 10 | +- For Intel® Data Center GPU Flex Series |
18 | 11 |
|
19 | | -### Sample Use cases |
| 12 | + Refer to [AI Model Zoo Containers on Flex Series](https://www.intel.com/content/www/us/en/developer/articles/containers/model-zoo-flex-series-containers.html) to run optimized Deep Learning inference workloads. |
20 | 13 |
|
| 14 | +- For Intel® Data Center GPU Max Series |
21 | 15 |
|
22 | | -|Model|Mode|Model Documentation| |
23 | | -|-|-|-| |
24 | | -|Inception V3|Inference|[FP32](https://github.com/IntelAI/models/blob/v2.8.0/benchmarks/image_recognition/tensorflow/inceptionv3/inference/fp32/README.md) [INT8](https://github.com/IntelAI/models/blob/v2.8.0/benchmarks/image_recognition/tensorflow/inceptionv3/inference/int8/README.md)| |
25 | | -|Inception V4|Inference|[FP32](https://github.com/IntelAI/models/blob/v2.8.0/benchmarks/image_recognition/tensorflow/inceptionv4/inference/fp32/README.md) [INT8](https://github.com/IntelAI/models/blob/v2.8.0/benchmarks/image_recognition/tensorflow/inceptionv4/inference/int8/README.md)| |
26 | | -|ResNet50 V1.5|Inference|[FP32](https://github.com/IntelAI/models/blob/v2.8.0/benchmarks/image_recognition/tensorflow/resnet50v1_5/inference/fp32/README.md) [INT8](https://github.com/IntelAI/models/blob/v2.8.0/benchmarks/image_recognition/tensorflow/resnet50v1_5/inference/int8/README.md)| |
| 16 | + Refer to [AI Model Zoo Containers on Max Series](https://www.intel.com/content/www/us/en/developer/articles/containers/model-zoo-max-series-containers/model-zoo-max-series-containers.html) to run optimized Deep Learning training and inference workloads. |
27 | 17 |
|
28 | | -### Performance Optimization |
29 | | -- FP16/BF16 INT8 Inference Optimization |
| 18 | +### Run Models on Bare Metal |
30 | 19 |
|
31 | | - Refer to the above FP32 model documentation, and only set one extra environment variable to enable advanced auto mixed precision Graph optimization before running inference. |
32 | | - ```bash |
33 | | - export ITEX_AUTO_MIXED_PRECISION=1 |
34 | | - ``` |
35 | | -- INT8 Inference Optimization |
36 | | - |
37 | | - To avoid memory copy on GPU, we provide a tool to convert the const to host const for INT8 pretrained-models. |
38 | | - |
39 | | - Take the ResNet50 v1.5 INT8 pb for example, |
40 | | - ``` |
41 | | -
|
42 | | - wget https://storage.googleapis.com/intel-optimized-tensorflow/models/v1_8/resnet50v1_5_int8_pretrained_model.pb |
43 | | -
|
44 | | - python host_const.py -i <path to the frozen graph downloaded above>/resnet50v1_5_int8_pretrained_model.pb -b -o <path to save the converted frozen graph>/resnet50v1_5_int8_pretrained_model-hostconst.pb |
45 | | - ``` |
46 | | - Use the new INT8 pb for INT8 inference, After converting to the new INT8 pb. |
47 | | - |
48 | | -## FAQ |
49 | | -1. During the Inception V3 INT8 batch inference, if running with real data, you might encounter a message "Running out of images from dataset". It is a known issue of Model Zoo script. |
50 | | - |
51 | | -Solution: |
52 | | - |
53 | | -- Option 1: Please use dummy data instead. |
54 | | - |
55 | | -- Option 2: If you want to run inference with real data, use the command below. And comment the last line of below int8_batch_inference.sh script to unspecify the warmup_steps and steps. |
56 | | -```bash |
57 | | -cd models |
58 | | -vi ./quickstart/image_recognition/tensorflow/inceptionv3/inference/cpu/int8/int8_batch_inference.sh |
59 | | -``` |
| 20 | +Refer to [AI Model Zoo Examples on Intel® Data Center GPU](https://github.com/IntelAI/models/tree/master#intel-data-center-gpu-workloads) to run optimized Deep Learning training and inference workloads on bare metal. |
0 commit comments