-
Notifications
You must be signed in to change notification settings - Fork 28.6k
SPARK-1209 [CORE] SparkHadoop{MapRed,MapReduce}Util should not use package org.apache.hadoop #2814
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
QA tests have started for PR 2814 at commit
|
Test FAILed. |
jenkins, test this please. |
Jenkins, test this please. |
QA tests have finished for PR 2814 at commit
|
Test FAILed. |
Test FAILed. |
I addressed the MIMA warnings, two of which were just caused by the class being renamed. I'm less sure about the warning around |
QA tests have started for PR 2814 at commit
|
QA tests have finished for PR 2814 at commit
|
Test PASSed. |
QA tests have started for PR 2814 at commit
|
QA tests have finished for PR 2814 at commit
|
Test PASSed. |
Test build #22248 has started for PR 2814 at commit
|
Test build #22248 has finished for PR 2814 at commit
|
Test FAILed. |
retest this please |
Test build #22373 has started for PR 2814 at commit
|
Test build #22373 timed out for PR 2814 at commit |
Test FAILed. |
retest this please |
Test build #22388 has started for PR 2814 at commit
|
Test build #22388 has finished for PR 2814 at commit
|
Test FAILed. |
retest this please |
Test build #22456 has started for PR 2814 at commit
|
Test build #22456 has finished for PR 2814 at commit
|
Test PASSed. |
LGTM. Looks like there are now merge conflicts. |
….hadoop to org.apache.spark
…o part of the PairRDDFunctions type hierarchy though?
@andrewor14 Yep, rebased it. |
Test build #22563 has started for PR 2814 at commit
|
Test build #22563 has finished for PR 2814 at commit
|
Test PASSed. |
Ok I'm merging this. Thanks. |
Hey @srowen I reverted this because it was failing tests for Spark built with Hadoop 1.0.4. I haven't dug into it deeply but it seems to be failing at run time due to some reflection issue. Do you know what's going on?
|
Hm, that's weird since I thought I ran the test against the default Hadoop, and that's 1.0.4. Which test failed? (I'll go look around jenkins too) I can't find 1.0.4 source at the moment, but it looks like in ancient times, the JobContextImpl constructor was package-private: So All of this would go away once Hadoop 1.x is not supported anymore too, along with some more acrobatics inside |
…ot use package org.apache.hadoop andrewor14 Another try at SPARK-1209, to address #2814 (comment) I successfully tested with `mvn -Dhadoop.version=1.0.4 -DskipTests clean package; mvn -Dhadoop.version=1.0.4 test` I assume that is what failed Jenkins last time. I also tried `-Dhadoop.version1.2.1` and `-Phadoop-2.4 -Pyarn -Phive` for more coverage. So this is why the class was put in `org.apache.hadoop` to begin with, I assume. One option is to leave this as-is for now and move it only when Hadoop 1.0.x support goes away. This is the other option, which adds a call to force the constructor to be public at run-time. It's probably less surprising than putting Spark code in `org.apache.hadoop`, but, does involve reflection. A `SecurityManager` might forbid this, but it would forbid a lot of stuff Spark does. This would also only affect Hadoop 1.0.x it seems. Author: Sean Owen <sowen@cloudera.com> Closes #3048 from srowen/SPARK-1209 and squashes the following commits: 0d48f4b [Sean Owen] For Hadoop 1.0.x, make certain constructors public, which were public in later versions 466e179 [Sean Owen] Disable MIMA warnings resulting from moving the class -- this was also part of the PairRDDFunctions type hierarchy though? eb61820 [Sean Owen] Move SparkHadoopMapRedUtil / SparkHadoopMapReduceUtil from org.apache.hadoop to org.apache.spark (cherry picked from commit f8e5732) Signed-off-by: Patrick Wendell <pwendell@gmail.com>
…ot use package org.apache.hadoop andrewor14 Another try at SPARK-1209, to address #2814 (comment) I successfully tested with `mvn -Dhadoop.version=1.0.4 -DskipTests clean package; mvn -Dhadoop.version=1.0.4 test` I assume that is what failed Jenkins last time. I also tried `-Dhadoop.version1.2.1` and `-Phadoop-2.4 -Pyarn -Phive` for more coverage. So this is why the class was put in `org.apache.hadoop` to begin with, I assume. One option is to leave this as-is for now and move it only when Hadoop 1.0.x support goes away. This is the other option, which adds a call to force the constructor to be public at run-time. It's probably less surprising than putting Spark code in `org.apache.hadoop`, but, does involve reflection. A `SecurityManager` might forbid this, but it would forbid a lot of stuff Spark does. This would also only affect Hadoop 1.0.x it seems. Author: Sean Owen <sowen@cloudera.com> Closes #3048 from srowen/SPARK-1209 and squashes the following commits: 0d48f4b [Sean Owen] For Hadoop 1.0.x, make certain constructors public, which were public in later versions 466e179 [Sean Owen] Disable MIMA warnings resulting from moving the class -- this was also part of the PairRDDFunctions type hierarchy though? eb61820 [Sean Owen] Move SparkHadoopMapRedUtil / SparkHadoopMapReduceUtil from org.apache.hadoop to org.apache.spark
(This is just a look at what completely moving the classes would look like. I know Patrick flagged that as maybe not OK, although, it's private?)