You are viewing an old version of this page. View the current version.

Compare with Current View Page History

« Previous Version 17 Next »

Google doc for comments : 

https://docs.google.com/document/d/1He-CGz8bTb151vhSeky-chCL1n93KyOLTxEFx91EuPU/edit?usp=sharing

Purpose

Two advanced features, fused computation and reduced-precision kernels, are introduced by MKL-DNN in the recent version [1,2]. These features can significantly speed up the inference performance on CPU for a broad range of deep learning topologies.

However, MXNet still cannot benefit from them because of the limitation of graph representation and the lack of graph optimization previously. Fortunately, the new subgraph feature [3] of MXNet makes these improvements possible now.

This document serves the purpose to illustrate the subgraph based solution to leverage MKL-DNN's new features into MXNet in Q3'2018. In general, the new solution will partition MKL-DNN operators into a subgraph, and then replace those MKL-DNN operators with fused kernels if possible, within the subgraph. Furthermore, users can choose the quantization flow to accelerate the inference procedure of their models by using the reduced-precision kernels of MKL-DNN, such as INT8 kernels.

Milestone

The whole project of this proposal is targeting to be released in 1.4.0 of MXNet. Three steps in the above can be considered as milestones of this project and used to track the process.

  • PR#12157 the subgraph branch into the master branch based on [3] (merge to master branch at the end of August)
  • submit MKL-DNN fusion implementation into subgraph branch (merge to subgraph branch at the middle of August)
  • submit the quantization flow implementation into the master branch (merge to master branch at the middle of September)

Workflow

Step 1. Partition MKL-DNN operators into the subgraph

This is one of the main purposes of the subgraph. More details are described at “Unified integration with external backend libraries” [3] . To achieve this, we can list all operators MKL-DNN supported and pass it to DefaultSubgraphProperty. After applying graph partitioning, all MKL-DNN operators will group into subgraph node, and data format will be converted between MKL-DNN internal format and NDArray default format on the subgraph boundary automatically. Subgraph border is a bit different according to executing mode.

1.         Symbolic mode

Subgraph will try to cover adjacent MKL-DNN operators as much as possible.

2.         Imperative mode

Each MKL-DNN operator will execute in an independent subgraph to ensure MKL-DNN internal format won’t be exposed outside subgraph.

Step 2. MKL-DNN operator fusion

MKL-DNN library supports running several certain patterned operators in a single execution. Such as convolution + relu. We can define new SubgraphSelector to capture such operator pattern and generate new MKL-DNN specific operator to replace the original ones.

New fused operators are standalone operators which represent the operations MKL-DNN library defined. For example, in MKL-DNN library, convolution can support post operations that allow executing relu following convolution. So, for convolution + relu fusion, we will create a MKL-DNN convolution (named _sg_MKL-DNN_conv) and describe relu as a post operation of it.

 

NNVM_REGISTER_OP(_sg_MKL-DNN_conv)
.set_attr<FStatefulComputeEx>("FStatefulComputeEx<cpu>",  SgMKL-DNNConvOpForward)

 

In general, new fused operators follow the abstraction of MKL-DNN library, and subgraph fusion pass is the lowering process to convert NNVM common graph to MKL-DNN graph. Fusion pass only happens inside the subgraph created by step1, so won’t affect MXNet default operators.

Step 3. Quantization

MKL-DNN supports most reduced precision primitives in convolutional neural networks, especially for fused primitives. MXNet INT8 inference consists of two steps:

1.       Prepare INT8 model, quantize parameters and collect calibration data

We perform this step only once. To create INT8 model based on FP32 model, we will run QuantizeGraph pass inside subgraph to replace FP32 operators with INT8 operators if MKL-DNN supports and insert dequantize operator on proper position.

2.       Run INT8 inference

When INT8 symbols and parameters are ready for inference, the user can run inference on new input data which is the same as before.

Accuracy Validation

Regardless of quantization, subgraph solution won’t introduce accuracy lost itself, on the contrary, it will enhance framework stability when using MKL-DNN. 

NETWORK

FP32

FP32 with Fusion

Top1

Top5

Top1

Top5

Resnet-152 by MKL-DNN

77.16

92.98

77.16

92.98

Inception by MKL-DNN

72.36

90.58

72.36

90.58

For quantization enabled case, MKL-DNN can get huge performance improvement within less than 0.5% accuracy loss comparing with FP32 with entropy mode calibration.

Accuracy data from the internal branch are shown in below table [4].

 

FP32

INT8

  

No Calibration

Calibration using 5 batches

Calibration Method

NETWORK

Top1

Top5Top1Top5Top1Top5
Resnet-152 by GPU77.1993.0175.5692.3275.5892.24Threshold by min/max
75.6592.35Threshold by entropy loss
Resnet-152 by MKL-DNN77.1692.9876.5992.7176.4492.69Threshold by min/max
76.7692.79Threshold by entropy loss
Inception-bn by GPU72.3890.6171.9890.2671.7890.26Threshold by min/max
71.9890.36Threshold by entropy loss
Inception-bn by MKL-DNN72.3690.5872.2190.5072.1690.43Threshold by min/max
72.1990.45Threshold by entropy loss


Performance

On performance part, both fusion and quantization can provide the huge improvement on various kinds of topologies.

Below is the inference performance data(images/sec) from the internal branch based on SKX-8180 1 socket with batch size 128. 

TopologyBaseWith FusionWith Quantization
resnet50-v1198333480
resnet50-v2190244In progress
resnet152-v279101In progress
vgg-168897In progress
vgg-197177In progress
inception-bn445583In progress
inception-v3167209In progress
inception-v485103In progress
googlenet642724In progress

 The quantization solution is still under development. We will share more data when it’s ready.

Testcase

Tests need to cover 2 parts. First one is the graph conversion test. We need to ensure that:

Step

Criterion 

1All MKL-DNN operators are partitioned into one or more subgraphs according to executing mode.
2Desired patterns can be captured and desired fused operators will be created.
3Quantization pass can convert desired operators to the quantized version with the correct data connection.

Another one is the unit test for MKL-DNN specific fused operators. The test should cover all the fusion scenarios to ensure the fused operators can provide the accurate result.

 

Reference

  1. Intel MKL-DNN, Introduction to Low-Precision 8-bit Integer Computations
    http://intel.github.io/mkl-dnn/ex_int8_simplenet.html
  2. Intel MKL-DNN, Post OPs Struct Reference
    http://intel.github.io/mkl-dnn/structmkldnn_1_1post__ops.html
  3. Da Zheng, Unified integration with external backend libraries
    https://cwiki.apache.org/confluence/display/MXNET/Unified+integration+with+external+backend+libraries
  4. Wu Jun, Model Quantization with Calibration
    https://github.com/apache/incubator-mxnet/files/1662073/quantization_github.pptx
    https://github.com/apache/incubator-mxnet/pull/9552


  • No labels