Skip to content

[PIR][oneDNN] Fix bf16 quantization for cast #68057

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Merged
merged 4 commits into from
Sep 14, 2024

Conversation

LLee233
Copy link
Contributor

@LLee233 LLee233 commented Sep 6, 2024

PR Category

Inference

PR Types

Others

Description

There are situations when Cast is receiving input other than float32/bf16, in which case Cast shall not be transformed to oneDNN op. Besides, Cast itself has the ability to transform tensor dtype, no need to add extra Quantize.

This PR is targeting to issue and improvement both as described above.

Copy link

paddle-bot bot commented Sep 6, 2024

你的PR提交成功,感谢你对开源项目的贡献!
请关注后续CI自动化测试结果,详情请参考Paddle-CI手册
Your PR has been submitted. Thanks for your contribution!
Please wait for the result of CI firstly. See Paddle CI Manual for details.

@LLee233 LLee233 added the contributor External developers label Sep 6, 2024
zhanglirong1999
zhanglirong1999 previously approved these changes Sep 6, 2024
@LLee233 LLee233 force-pushed the fix_cast branch 2 times, most recently from 1d36143 to 078616e Compare September 9, 2024 07:25
Copy link
Contributor

@zhanglirong1999 zhanglirong1999 left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

LGTM

@zhanglirong1999 zhanglirong1999 merged commit 4bb1f2f into PaddlePaddle:develop Sep 14, 2024
28 of 30 checks passed
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
contributor External developers Intel
Projects
None yet
Development

Successfully merging this pull request may close these issues.

2 participants