-
Notifications
You must be signed in to change notification settings - Fork 3k
Miniconda py38 and Spark 3.0 #1115
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Conversation
We should consider swapping out Toree and spylon for https://almond.sh/ |
We could add it but Almond has a Dockerfile already which extends from base-notebook, not sure what value does a copy of others code adds here. It will make the image larger than it already is. Maybe add a link to the Almond repo in the docs? |
Good call. That it does.
Without Toree or Spylon in the all-spark-notebook image, there's no support for using Scala in Jupyter in the image any longer. Maybe it's fine to support Python and R only but it is a significant departure from the original purpose of the image. |
These's a good chance Toree will be upgraded to work with JDK 11+ now that Spark 3.0 is out. |
@@ -88,7 +88,7 @@ lint-build-test-all: $(foreach I,$(ALL_IMAGES),lint/$(I) arch_patch/$(I) build/$ | |||
|
|||
lint-install: ## install hadolint | |||
@echo "Installing hadolint at $(HADOLINT) ..." | |||
@curl -sL -o $(HADOLINT) "https://github.com/hadolint/hadolint/releases/download/v1.17.6/hadolint-$(shell uname -s)-$(shell uname -m)" | |||
@curl -sL -o $(HADOLINT) "https://github.com/hadolint/hadolint/releases/download/v1.18.0/hadolint-$(shell uname -s)-$(shell uname -m)" |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Thank you for this update. I didn't realize that the version was hardcoded -- my bad. It could be an opportunity to put the hadolint version in a variable and to print it in the echo
to know the version that was used.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Hello,
Just a remark if Scala kernel(s) is(are) removed we need also to remove the corresponding example(s) from the Image Specifics page.
Best
@@ -76,14 +76,14 @@ RUN mkdir /home/$NB_USER/work && \ | |||
|
|||
# Install conda as jovyan and check the md5 sum provided on the download site | |||
ENV MINICONDA_VERSION=4.8.2 \ |
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
There is a new version 4.8.3
(Miniconda3-py38_4.8.3-Linux-x86_64.sh
) in the miniconda repo. We should also use this new version.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I was trying to limit the change set to avoid breaking things, I figured once this PR gets merged then I can do it in the next PR, unless disagree.
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Ok it makes sense. Thanks.
I would remove spylon from the image as unmaintained. |
Do you want to do it in this PR or next one? It still works with Spark 3.0. |
If it's working with 3.0, go ahead and leave it for now. |
Looks like Travis CI failed due to a web error? It builds locally for me. Note the warning:
I don't know where the aforementioned conflicting explicit spec is, but perhaps it would be better after all to just bump to 4.8.3:
|
@parente Can we merge this PR yet? Thx |
To be more specific, base-notebook builds for me. I didn't try the others... |
The PR builds fine, the failure is a Travis issue not the code. |
@Bidek56 I'm planning on tagging the last images containing Spark 2.x and then merging this PR when I get a few minutes in the next day or two. |
I've tagged the latest pyspark and all-spark images with |
Yes, release coming soon. |
@lresende Nothing prevents coexistence. There was a request + push to get Spark 3.0 into the images. We can add toree back as soon as it's compatible. |
Allow to build `pyspark-notebook` image with an alternative Spark version. - Define arguments for Spark installation - Add a note in "Image Specifics" explaining how to build an image with an alternative Spark version - Remove Toree documentation from "Image Specifics" since its support has been droped in jupyter#1115
Upgrading Spark to 3.0
Upgrading Python to 3.8
Removing Toree due to lack of support
Open JDK needs to stay at 11 because of SparkR limitation
Spylon kernel may need to be removed soon since it has not been maintained in years