Skip to content

Commit 219a55c

Browse files
authored
Change the float16 inference report to README.md (#10465)
* change to readme * Update README.md
1 parent 380471f commit 219a55c

File tree

1 file changed

+4
-0
lines changed

1 file changed

+4
-0
lines changed

contrib/float16/float16_inference_report.md renamed to contrib/float16/README.md

+4
Original file line numberDiff line numberDiff line change
@@ -1,3 +1,7 @@
1+
# Float16 Inference in PaddlePaddle Fluid
2+
3+
Kexin Zhao <zhaokexin01@baidu.com>
4+
15
## Introduction
26
Working with deep neural networks (DNN) is a two-stage process. First we train DNN using labeled examples of inputs and desired outputs to obtain the model parameters (weights), then we deploy DNN along with the trained weights to run inference on unknown inputs. Typically, these weights are in float data type and hence we run inference in float mode using these weights. This post focuses on the discussion of how to use low precision float16 data type to represent these trained weights and run inference in float16 mode as well as the advantages of float16 inference over its float counterpart by showing some experiment results.
37

0 commit comments

Comments
 (0)