-
Notifications
You must be signed in to change notification settings - Fork 3.5k
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
[VTA][HotFix] Relay->VTA quantization fix #4433
Conversation
@liangfu This is the follow up to the fix you requested. Right now quantization breaks compilation to VTA, so this will require further investigation. |
@tmoreau89 Default opt_level has been changed to 2, batch_norm won't be folded during quantization, will this cause issue in vta? You can still wrap |
bingo - that's very likely the root issue which explains why we have multiplication in there |
@vinx13 looks like the |
Hotfix should be ready for review. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM
Thanks @tmoreau89 @vinx13 @liangfu this is merged and will be included in v0.6.0 release |
* relay -> vta fix * setting optlevel to 3 for quantization to fold batchnorm
* relay -> vta fix * setting optlevel to 3 for quantization to fold batchnorm
* relay -> vta fix * setting optlevel to 3 for quantization to fold batchnorm
* relay -> vta fix * setting optlevel to 3 for quantization to fold batchnorm
* relay -> vta fix * setting optlevel to 3 for quantization to fold batchnorm
This addresses a compilation bug introduced by: #4295
Other than interface changes to quantization pass (graph vs. module) it appears that this has broken quantization pass for VTA by inserting multiplication at the end of the layer when multiplication is not supported by VTA (and instead must rely on shift and add).
Investigation in progress.
@vinx13 @ZihengJiang