Skip to content

Quantized op implementation: quantized_decomposed::dequantize_per_channel_group #7676

@sheetalarkadam

Description

@sheetalarkadam

🚀 The feature, motivation and pitch

The op quantized_decomposed::dequantize_per_channel_group implementation would help in using executorch's quantized CPU ops in models like llama3

Came across this issue when trying to build llama3 without XNNPack. More details here #6975

Alternatives

No response

Additional context

No response

RFC (Optional)

No response

cc @larryliu0820 @manuelcandales @kimishpatel @jerryzh168

Metadata

Metadata

Labels

module: kernelsIssues related to kernel libraries and utilities, and code under kernels/module: quantizationIssues related to quantizationrfcRequest for comment and feedback on a post, proposal, etc.triagedThis issue has been looked at a team member, and triaged and prioritized into an appropriate module

Projects

No projects

Milestone

No milestone

Relationships

None yet

Development

No branches or pull requests

Issue actions