- 
                Notifications
    You must be signed in to change notification settings 
- Fork 247
fix: Fall back to Spark when hashing decimals with precision > 18 #1325
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
| Codecov ReportAttention: Patch coverage is  
 
 Additional details and impacted files@@              Coverage Diff              @@
##               main    #1325       +/-   ##
=============================================
- Coverage     56.12%   39.13%   -17.00%     
- Complexity      976     2065     +1089     
=============================================
  Files           119      262      +143     
  Lines         11743    60262    +48519     
  Branches       2251    12819    +10568     
=============================================
+ Hits           6591    23581    +16990     
- Misses         4012    32201    +28189     
- Partials       1140     4480     +3340     ☔ View full report in Codecov by Sentry. | 
| Does this implicitly affect any data read that originated as  | 
| 
 Yes, in the context of a user calling the  | 
        
          
                spark/src/test/scala/org/apache/comet/CometExpressionSuite.scala
              
                Outdated
          
            Show resolved
            Hide resolved
        
      | def isSupportedType(expr: Expression): Boolean = { | ||
| for (child <- expr.children) { | ||
| child.dataType match { | ||
| case dt: DecimalType if dt.precision > 18 => | 
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
From the issue it seemed that we get test failure even if the precision is less than 18. So do we want to fallback to Spark for all DecimalType values?
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There was an earlier PR that implemented the correct behavior for the case where precision is < 18.
| Thanks for the review @kazuyukitanimura and @parthchandra | 
Which issue does this PR close?
Closes #1294
Rationale for this change
Fix correctness issue
What changes are included in this PR?
Fall back to Spark when hashing decimals with precision > 18 when using Murmur3Hash or XXHash64.
How are these changes tested?