-
Notifications
You must be signed in to change notification settings - Fork 5.9k
Closed
Labels
预测原名Inference,包含Capi预测问题等原名Inference,包含Capi预测问题等
Description
The usual scenario for float16 inference is as follows:
- We first train a model in float32 mode
- We convert the float32 weights into float16, and save them on dist
- During inference, we load the float16 weights and model, run inference engine in float16 mode.
To support this, we need to make save op capable of saving weights as float16 type.
Metadata
Metadata
Assignees
Labels
预测原名Inference,包含Capi预测问题等原名Inference,包含Capi预测问题等