Skip to content

Commit aa39c55

Browse files
author
Davies Liu
committed
Merge branch 'master' of github.com:apache/spark into python_tests
Conflicts: python/pyspark/mllib/tests.py python/pyspark/mllib/tree.py python/pyspark/tests.py
2 parents 55bb451 + 04e44b3 commit aa39c55

File tree

98 files changed

+1414
-1402
lines changed

Some content is hidden

Large Commits have some content hidden by default. Use the searchbox below for content that may be hidden.

98 files changed

+1414
-1402
lines changed

R/pkg/DESCRIPTION

Lines changed: 3 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -17,19 +17,19 @@ License: Apache License (== 2.0)
1717
Collate:
1818
'generics.R'
1919
'jobj.R'
20-
'SQLTypes.R'
2120
'RDD.R'
2221
'pairRDD.R'
22+
'SQLTypes.R'
2323
'column.R'
2424
'group.R'
2525
'DataFrame.R'
2626
'SQLContext.R'
27+
'backend.R'
2728
'broadcast.R'
29+
'client.R'
2830
'context.R'
2931
'deserialize.R'
3032
'serialize.R'
3133
'sparkR.R'
32-
'backend.R'
33-
'client.R'
3434
'utils.R'
3535
'zzz.R'

R/pkg/R/DataFrame.R

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717

1818
# DataFrame.R - DataFrame class and methods implemented in S4 OO classes
1919

20-
#' @include jobj.R SQLTypes.R RDD.R pairRDD.R column.R group.R
20+
#' @include generics.R jobj.R SQLTypes.R RDD.R pairRDD.R column.R group.R
2121
NULL
2222

2323
setOldClass("jobj")

R/pkg/R/column.R

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -17,7 +17,7 @@
1717

1818
# Column Class
1919

20-
#' @include generics.R jobj.R
20+
#' @include generics.R jobj.R SQLTypes.R
2121
NULL
2222

2323
setOldClass("jobj")

R/pkg/R/group.R

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -17,6 +17,9 @@
1717

1818
# group.R - GroupedData class and methods implemented in S4 OO classes
1919

20+
#' @include generics.R jobj.R SQLTypes.R column.R
21+
NULL
22+
2023
setOldClass("jobj")
2124

2225
#' @title S4 class that represents a GroupedData

R/pkg/R/jobj.R

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -18,6 +18,9 @@
1818
# References to objects that exist on the JVM backend
1919
# are maintained using the jobj.
2020

21+
#' @include generics.R
22+
NULL
23+
2124
# Maintain a reference count of Java object references
2225
# This allows us to GC the java object when it is safe
2326
.validJobjs <- new.env(parent = emptyenv())

R/pkg/R/pairRDD.R

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -16,6 +16,8 @@
1616
#
1717

1818
# Operations supported on RDDs contains pairs (i.e key, value)
19+
#' @include generics.R jobj.R RDD.R
20+
NULL
1921

2022
############ Actions and Transformations ############
2123

bin/pyspark

Lines changed: 1 addition & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -89,6 +89,7 @@ export PYTHONSTARTUP="$SPARK_HOME/python/pyspark/shell.py"
8989
if [[ -n "$SPARK_TESTING" ]]; then
9090
unset YARN_CONF_DIR
9191
unset HADOOP_CONF_DIR
92+
export PYTHONHASHSEED=0
9293
if [[ -n "$PYSPARK_DOC_TEST" ]]; then
9394
exec "$PYSPARK_DRIVER_PYTHON" -m doctest $1
9495
else

bin/spark-submit

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -19,6 +19,9 @@
1919

2020
SPARK_HOME="$(cd "`dirname "$0"`"/..; pwd)"
2121

22+
# disable randomized hash for string in Python 3.3+
23+
export PYTHONHASHSEED=0
24+
2225
# Only define a usage function if an upstream script hasn't done so.
2326
if ! type -t usage >/dev/null 2>&1; then
2427
usage() {

bin/spark-submit2.cmd

Lines changed: 3 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -20,6 +20,9 @@ rem
2020
rem This is the entry point for running Spark submit. To avoid polluting the
2121
rem environment, it just launches a new cmd to do the real work.
2222

23+
rem disable randomized hash for string in Python 3.3+
24+
set PYTHONHASHSEED=0
25+
2326
set CLASS=org.apache.spark.deploy.SparkSubmit
2427
call %~dp0spark-class2.cmd %CLASS% %*
2528
set SPARK_ERROR_LEVEL=%ERRORLEVEL%

core/src/main/scala/org/apache/spark/rpc/RpcEnv.scala

Lines changed: 4 additions & 3 deletions
Original file line numberDiff line numberDiff line change
@@ -284,7 +284,7 @@ private[spark] abstract class RpcEndpointRef(@transient conf: SparkConf)
284284

285285
private[this] val maxRetries = conf.getInt("spark.akka.num.retries", 3)
286286
private[this] val retryWaitMs = conf.getLong("spark.akka.retry.wait", 3000)
287-
private[this] val defaultTimeout = conf.getLong("spark.akka.lookupTimeout", 30) seconds
287+
private[this] val defaultAskTimeout = conf.getLong("spark.akka.askTimeout", 30) seconds
288288

289289
/**
290290
* return the address for the [[RpcEndpointRef]]
@@ -304,7 +304,8 @@ private[spark] abstract class RpcEndpointRef(@transient conf: SparkConf)
304304
*
305305
* This method only sends the message once and never retries.
306306
*/
307-
def sendWithReply[T: ClassTag](message: Any): Future[T] = sendWithReply(message, defaultTimeout)
307+
def sendWithReply[T: ClassTag](message: Any): Future[T] =
308+
sendWithReply(message, defaultAskTimeout)
308309

309310
/**
310311
* Send a message to the corresponding [[RpcEndpoint.receiveAndReply)]] and return a `Future` to
@@ -327,7 +328,7 @@ private[spark] abstract class RpcEndpointRef(@transient conf: SparkConf)
327328
* @tparam T type of the reply message
328329
* @return the reply message from the corresponding [[RpcEndpoint]]
329330
*/
330-
def askWithReply[T: ClassTag](message: Any): T = askWithReply(message, defaultTimeout)
331+
def askWithReply[T: ClassTag](message: Any): T = askWithReply(message, defaultAskTimeout)
331332

332333
/**
333334
* Send a message to the corresponding [[RpcEndpoint.receive]] and get its result within a

dev/run-tests

Lines changed: 2 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -235,6 +235,8 @@ echo "========================================================================="
235235

236236
CURRENT_BLOCK=$BLOCK_PYSPARK_UNIT_TESTS
237237

238+
# add path for python 3 in jenkins
239+
export PATH="${PATH}:/home/anaonda/envs/py3k/bin"
238240
./python/run-tests
239241

240242
echo ""

dev/run-tests-jenkins

Lines changed: 1 addition & 1 deletion
Original file line numberDiff line numberDiff line change
@@ -47,7 +47,7 @@ COMMIT_URL="https://github.com/apache/spark/commit/${ghprbActualCommit}"
4747
# GitHub doesn't auto-link short hashes when submitted via the API, unfortunately. :(
4848
SHORT_COMMIT_HASH="${ghprbActualCommit:0:7}"
4949

50-
TESTS_TIMEOUT="120m" # format: http://linux.die.net/man/1/timeout
50+
TESTS_TIMEOUT="150m" # format: http://linux.die.net/man/1/timeout
5151

5252
# Array to capture all tests to run on the pull request. These tests are held under the
5353
#+ dev/tests/ directory.

0 commit comments

Comments
 (0)