|
17 | 17 |
|
18 | 18 | package org.apache.kyuubi.kubernetes.test.spark |
19 | 19 |
|
| 20 | +import java.util.UUID |
| 21 | + |
20 | 22 | import scala.collection.JavaConverters._ |
21 | 23 | import scala.concurrent.duration._ |
22 | 24 |
|
23 | 25 | import org.apache.hadoop.conf.Configuration |
24 | 26 | import org.apache.hadoop.net.NetUtils |
25 | 27 |
|
26 | | -import org.apache.kyuubi.{BatchTestHelper, KyuubiException, Logging, Utils, WithKyuubiServer, WithSimpleDFSService} |
| 28 | +import org.apache.kyuubi._ |
27 | 29 | import org.apache.kyuubi.config.KyuubiConf |
28 | 30 | import org.apache.kyuubi.config.KyuubiConf.FRONTEND_THRIFT_BINARY_BIND_HOST |
29 | 31 | import org.apache.kyuubi.engine.{ApplicationInfo, ApplicationOperation, KubernetesApplicationOperation} |
@@ -134,7 +136,8 @@ class KyuubiOperationKubernetesClusterClientModeSuite |
134 | 136 | server.backendService.sessionManager.asInstanceOf[KyuubiSessionManager] |
135 | 137 |
|
136 | 138 | test("Spark Client Mode On Kubernetes Kyuubi KubernetesApplicationOperation Suite") { |
137 | | - val batchRequest = newSparkBatchRequest(conf.getAll) |
| 139 | + val batchRequest = newSparkBatchRequest(conf.getAll ++ Map( |
| 140 | + "kyuubi.batch.id" -> UUID.randomUUID().toString)) |
138 | 141 |
|
139 | 142 | val sessionHandle = sessionManager.openBatchSession( |
140 | 143 | "kyuubi", |
@@ -193,7 +196,8 @@ class KyuubiOperationKubernetesClusterClusterModeSuite |
193 | 196 | "spark.kubernetes.driver.pod.name", |
194 | 197 | driverPodNamePrefix + "-" + System.currentTimeMillis()) |
195 | 198 |
|
196 | | - val batchRequest = newSparkBatchRequest(conf.getAll) |
| 199 | + val batchRequest = newSparkBatchRequest(conf.getAll ++ Map( |
| 200 | + "kyuubi.batch.id" -> UUID.randomUUID().toString)) |
197 | 201 |
|
198 | 202 | val sessionHandle = sessionManager.openBatchSession( |
199 | 203 | "runner", |
|
0 commit comments