File tree Expand file tree Collapse file tree 2 files changed +6
-6
lines changed
repl/scala-2.11/src/main/scala/org/apache/spark/repl Expand file tree Collapse file tree 2 files changed +6
-6
lines changed Original file line number Diff line number Diff line change @@ -91,13 +91,13 @@ install_mvn() {
91
91
92
92
# Install zinc under the build/ folder
93
93
install_zinc () {
94
- local zinc_path=" zinc-0.3.15 /bin/zinc"
94
+ local zinc_path=" zinc-0.3.11 /bin/zinc"
95
95
[ ! -f " ${_DIR} /${zinc_path} " ] && ZINC_INSTALL_FLAG=1
96
96
local TYPESAFE_MIRROR=${TYPESAFE_MIRROR:- https:// downloads.typesafe.com}
97
97
98
98
install_app \
99
- " ${TYPESAFE_MIRROR} /zinc/0.3.15 " \
100
- " zinc-0.3.15 .tgz" \
99
+ " ${TYPESAFE_MIRROR} /zinc/0.3.11 " \
100
+ " zinc-0.3.11 .tgz" \
101
101
" ${zinc_path} "
102
102
ZINC_BIN=" ${_DIR} /${zinc_path} "
103
103
}
Original file line number Diff line number Diff line change @@ -97,13 +97,13 @@ class SparkILoop(in0: Option[BufferedReader], out: JPrintWriter)
97
97
override def commands : List [LoopCommand ] = sparkStandardCommands
98
98
99
99
/**
100
- * We override `createInterpreter ` because we need to initialize Spark *before* the REPL
100
+ * We override `loadFiles ` because we need to initialize Spark *before* the REPL
101
101
* sees any files, so that the Spark context is visible in those files. This is a bit of a
102
102
* hack, but there isn't another hook available to us at this point.
103
103
*/
104
- override def createInterpreter (): Unit = {
105
- super .createInterpreter()
104
+ override def loadFiles (settings : Settings ): Unit = {
106
105
initializeSpark()
106
+ super .loadFiles(settings)
107
107
}
108
108
109
109
override def resetCommand (line : String ): Unit = {
You can’t perform that action at this time.
0 commit comments