-
Notifications
You must be signed in to change notification settings - Fork 1.5k
Add Spark Job Launcher tool #9288
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Changes from all commits
5634a2c
c3292fb
3348ea5
4aeac5a
0d4a40e
41f9049
b678399
b7e5cab
2ba3254
6854682
File filter
Filter by extension
Conversations
Jump to
Diff view
Diff view
There are no files selected for viewing
| Original file line number | Diff line number | Diff line change |
|---|---|---|
|
|
@@ -34,6 +34,8 @@ | |
| <properties> | ||
| <pinot.root>${basedir}/..</pinot.root> | ||
| <aws.version>2.14.28</aws.version> | ||
| <scala.version>2.12</scala.version> | ||
| <spark.version>3.2.1</spark.version> | ||
| </properties> | ||
| <dependencies> | ||
| <dependency> | ||
|
|
@@ -268,6 +270,24 @@ | |
| <artifactId>mockito-core</artifactId> | ||
| <scope>test</scope> | ||
| </dependency> | ||
|
|
||
| <!-- | ||
| This dependency is needed for LaunchSparkDataIngestionJobCommand. | ||
| The dependency only contains a few classes and scala library which has been excluded. | ||
| Hence, it will not interfere with spark-core classes present in runtime env | ||
| and will use the env spark version to actually execute the spark job | ||
| --> | ||
| <dependency> | ||
| <groupId>org.apache.spark</groupId> | ||
| <artifactId>spark-launcher_${scala.version}</artifactId> | ||
| <version>${spark.version}</version> | ||
| <exclusions> | ||
|
||
| <exclusion> | ||
| <groupId>org.scala-lang</groupId> | ||
| <artifactId>scala-library</artifactId> | ||
| </exclusion> | ||
| </exclusions> | ||
| </dependency> | ||
| </dependencies> | ||
| <build> | ||
| <plugins> | ||
|
|
||
There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
Once you start pulling in Scala code, you need to ensure that every dependency that's also using Scala is using the same version. So I think this should go in the top-level
pom.xmlfile. Also note that pinot-kafka and pinot-spark have dependencies on Scala 2.11, which I believe will cause runtime problems if they are on the classpath when the tool is being run (and it's using 2.12).There was a problem hiding this comment.
Choose a reason for hiding this comment
The reason will be displayed to describe this comment to others. Learn more.
I have actually excluded the scala code from plugin. It is just that the plugin requires it in name.