-
Notifications
You must be signed in to change notification settings - Fork 28.6k
[SPARK-17679] [PYSPARK] remove unnecessary Py4J ListConverter patch #15254
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
@davies you authored #5570 and reported the issue in Py4J py4j/py4j#160. I happened across this while spelunking through Py4J code in PySpark, it seems like it's no longer needed. Do you agree? |
I guess we can also remove another workaround here ? |
cc @JoshRosen and @davies |
+1 on @lins05's suggestion of going further and removing any non-unnecessary explicit usages of |
@JoshRosen @lins05 As requested, I've removed all remaining explicit mentions of I'm not familiar with the code in pyspark-ml and pyspark-mllib, but it seemed straightforward and didn't introduce any regressions in the tests. |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
LGTM. The purpose of the workarounds was to:
- avoid auto converting bytearray to java ArrayList (fixed in auto_convert does not work with bytearray py4j/py4j#160)
- avoid the bug trigged when calling java class constructor with non-JavaObject args (fixed in JavaClass does not work with auto_convert py4j/py4j#161)
Since both bugs have been fixed, I think we are safe to remove the workarounds.
Jenkins, this is ok to test. |
lgtm, pending on jenkins. |
Test build #66045 has finished for PR 15254 at commit
|
+1 :) |
Merging this into master, thanks! |
What changes were proposed in this pull request?
This PR removes a patch on ListConverter from #5570, as it is no longer necessary. The underlying issue in Py4J py4j/py4j#160 was patched in py4j/py4j@224b94b and is present in 0.10.3, the version currently in use in Spark.
How was this patch tested?
The original test added in #5570 remains.