You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Update on "[ET-VK][int4] patch 4-bit linear op for ensuring w-packed in/out"
If the partitioner is using channels-packed setting for activations, then the checks will throw.
Remove the checks and conditionally re-pack the input/output tensors if they are not width-packed.
Differential Revision: [D68813946](https://our.internmc.facebook.com/intern/diff/D68813946/)
[ghstack-poisoned]
- func: cadence::quantized_fully_connected.per_tensor_out(Tensor src, Tensor weight, Tensor bias, int src_zero_point, int weight_zero_point, int out_multiplier, int out_shift, int out_zero_point, Tensor? offset, *, Tensor(a!) out) -> Tensor(a!)
0 commit comments