Skip to content

Conversation

@Elubrazione
Copy link

  • Add complete quantized_matmul_impl_typed template function for CPU (float16, float32, and bfloat16).
  • Add fp32 test cases for quantized_matmul.
  • Relax float32 tolerance in test utils.

- Add complete quantized_matmul_impl_typed template function for CPU, which support float16, float32, and bfloat16 data types
- Add float32 test cases for quantized_matmul
- Adjust float32 tolerance in test utils for better precision
Copy link
Owner

@skyzh skyzh left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Thanks!!

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment

Labels

None yet

Projects

None yet

Development

Successfully merging this pull request may close these issues.

2 participants