Skip to content

[SPARK-27563][SQL][TEST] automatically get the latest Spark versions in HiveExternalCatalogVersionsSuite #24454

New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

Closed
wants to merge 2 commits into from

Conversation

cloud-fan
Copy link
Contributor

What changes were proposed in this pull request?

We can get the latest downloadable Spark versions from https://dist.apache.org/repos/dist/release/spark/

How was this patch tested?

manually.

@cloud-fan
Copy link
Contributor Author

.filter(_.contains("""<li><a href="spark-"""))
.map("""<a href="spark-(\d.\d.\d)/">""".r.findFirstMatchIn(_).get.group(1))
.filter(_ < org.apache.spark.SPARK_VERSION)
logInfo(s"Testing ${org.apache.spark.SPARK_VERSION} with ${versions.mkString(", ")}.")
Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

we can check the testing log to make sure we picked the right versions to test.

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Nice! @cloud-fan

Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Cool!

.split("\n")
.filter(_.contains("""<li><a href="spark-"""))
.map("""<a href="spark-(\d.\d.\d)/">""".r.findFirstMatchIn(_).get.group(1))
.filter(_ < org.apache.spark.SPARK_VERSION)
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

isn;t this always true?

Copy link
Contributor Author

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

This PR should be merged to all the active branches: 2.3, 2.4 and master. Branch 2.3 should not test with Spark 2.4.x

val testingVersions = Seq("2.3.3", "2.4.2")
lazy val testingVersions: Seq[String] = {
import scala.io.Source
val versions = Source.fromURL("https://dist.apache.org/repos/dist/release/spark/").mkString
Copy link
Member

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

BTW, what if this link is temporarily unavailable?
Even if this is a potential issue, I think we should merge this and see if the issue is virtually ignorable/trivial or not.

@srowen
Copy link
Member

srowen commented Apr 25, 2019

This assumes that dist has only the latest versions in active branches, which it should, but depends on us deleting previous releases diligently. Not sure if it's in the release process. Ideally this would have further logic to only pick the latest version per branch. But not sure if that's complicated to write. Right now this would actually run an extra set of tests.

@SparkQA
Copy link

SparkQA commented Apr 25, 2019

Test build #104887 has finished for PR 24454 at commit 86e6039.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

@cloud-fan
Copy link
Contributor Author

@srowen yea we may need to run an extra set of tests for a while during the release. I think it's ok as the failure window is short and the release doesn't happen very frequently.

@cloud-fan
Copy link
Contributor Author

cloud-fan commented Apr 25, 2019

seems logInfo doesn't work in test suites. I searched the testing log and found SparkContext:54 - Running Spark version 2.3.3 and SparkContext - Running Spark version 2.4.2, so it does work.

@SparkQA
Copy link

SparkQA commented Apr 25, 2019

Test build #104899 has finished for PR 24454 at commit 8700851.

  • This patch passes all tests.
  • This patch merges cleanly.
  • This patch adds no public classes.

Copy link
Member

@srowen srowen left a comment

Choose a reason for hiding this comment

The reason will be displayed to describe this comment to others. Learn more.

Looks fine; the 2.4.1 release has been removed from dist (thanks)

@HyukjinKwon
Copy link
Member

Merged to master.

cloud-fan added a commit that referenced this pull request Apr 26, 2019
…in HiveExternalCatalogVersionsSuite

## What changes were proposed in this pull request?

We can get the latest downloadable Spark versions from https://dist.apache.org/repos/dist/release/spark/

## How was this patch tested?

manually.

Closes #24454 from cloud-fan/test.

Authored-by: Wenchen Fan <wenchen@databricks.com>
Signed-off-by: HyukjinKwon <gurwls223@apache.org>
cloud-fan added a commit that referenced this pull request Apr 26, 2019
…in HiveExternalCatalogVersionsSuite

We can get the latest downloadable Spark versions from https://dist.apache.org/repos/dist/release/spark/

manually.

Closes #24454 from cloud-fan/test.

Authored-by: Wenchen Fan <wenchen@databricks.com>
Signed-off-by: HyukjinKwon <gurwls223@apache.org>
@cloud-fan
Copy link
Contributor Author

I've backported it to 2.4/2.3, so that we can completely ignore it during release process. I'll update https://spark.apache.org/release-process.html soon.

@HyukjinKwon
Copy link
Member

Oops thanks

kai-chi pushed a commit to kai-chi/spark that referenced this pull request Jul 23, 2019
…in HiveExternalCatalogVersionsSuite

## What changes were proposed in this pull request?

We can get the latest downloadable Spark versions from https://dist.apache.org/repos/dist/release/spark/

## How was this patch tested?

manually.

Closes apache#24454 from cloud-fan/test.

Authored-by: Wenchen Fan <wenchen@databricks.com>
Signed-off-by: HyukjinKwon <gurwls223@apache.org>
kai-chi pushed a commit to kai-chi/spark that referenced this pull request Jul 25, 2019
…in HiveExternalCatalogVersionsSuite

## What changes were proposed in this pull request?

We can get the latest downloadable Spark versions from https://dist.apache.org/repos/dist/release/spark/

## How was this patch tested?

manually.

Closes apache#24454 from cloud-fan/test.

Authored-by: Wenchen Fan <wenchen@databricks.com>
Signed-off-by: HyukjinKwon <gurwls223@apache.org>
kai-chi pushed a commit to kai-chi/spark that referenced this pull request Aug 1, 2019
…in HiveExternalCatalogVersionsSuite

## What changes were proposed in this pull request?

We can get the latest downloadable Spark versions from https://dist.apache.org/repos/dist/release/spark/

## How was this patch tested?

manually.

Closes apache#24454 from cloud-fan/test.

Authored-by: Wenchen Fan <wenchen@databricks.com>
Signed-off-by: HyukjinKwon <gurwls223@apache.org>
@dongjoon-hyun
Copy link
Member

Hi, All.
Unfortunately, 3.0.0-preview breaks this patch. I made a PR to recover it.

Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
None yet
Projects
None yet
Development

Successfully merging this pull request may close these issues.

6 participants