Open
Description
🚀 The feature, motivation and pitch
The op quantized_decomposed::dequantize_per_channel_group
implementation would help in using executorch's quantized CPU ops in models like llama3
Came across this issue when trying to build llama3 without XNNPack. More details here #6975
Alternatives
No response
Additional context
No response
RFC (Optional)
No response