@@ -291,15 +291,18 @@ class ExecutorMonitorSuite extends SparkFunSuite {
291
291
292
292
// First a failed task, to make sure it does not count.
293
293
monitor.onTaskStart(SparkListenerTaskStart (1 , 0 , taskInfo(" 1" , 1 )))
294
- monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , TaskResultLost , taskInfo(" 1" , 1 ), null ))
294
+ monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , TaskResultLost , taskInfo(" 1" , 1 ),
295
+ new ExecutorMetrics , null ))
295
296
assert(monitor.timedOutExecutors(idleDeadline) === Seq (" 1" ))
296
297
297
298
monitor.onTaskStart(SparkListenerTaskStart (1 , 0 , taskInfo(" 1" , 1 )))
298
- monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , Success , taskInfo(" 1" , 1 ), null ))
299
+ monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , Success , taskInfo(" 1" , 1 ),
300
+ new ExecutorMetrics , null ))
299
301
assert(monitor.timedOutExecutors(idleDeadline).isEmpty)
300
302
301
303
monitor.onTaskStart(SparkListenerTaskStart (3 , 0 , taskInfo(" 1" , 1 )))
302
- monitor.onTaskEnd(SparkListenerTaskEnd (3 , 0 , " foo" , Success , taskInfo(" 1" , 1 ), null ))
304
+ monitor.onTaskEnd(SparkListenerTaskEnd (3 , 0 , " foo" , Success , taskInfo(" 1" , 1 ),
305
+ new ExecutorMetrics , null ))
303
306
assert(monitor.timedOutExecutors(idleDeadline).isEmpty)
304
307
305
308
// Finish the jobs, now the executor should be idle, but with the shuffle timeout, since the
@@ -351,11 +354,13 @@ class ExecutorMonitorSuite extends SparkFunSuite {
351
354
monitor.onJobStart(SparkListenerJobStart (2 , clock.getTimeMillis(), Seq (stage3, stage4)))
352
355
353
356
monitor.onTaskStart(SparkListenerTaskStart (1 , 0 , taskInfo(" 1" , 1 )))
354
- monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , Success , taskInfo(" 1" , 1 ), null ))
357
+ monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , Success , taskInfo(" 1" , 1 ),
358
+ new ExecutorMetrics , null ))
355
359
assert(monitor.timedOutExecutors(idleDeadline) === Seq (" 2" ))
356
360
357
361
monitor.onTaskStart(SparkListenerTaskStart (3 , 0 , taskInfo(" 2" , 1 )))
358
- monitor.onTaskEnd(SparkListenerTaskEnd (3 , 0 , " foo" , Success , taskInfo(" 2" , 1 ), null ))
362
+ monitor.onTaskEnd(SparkListenerTaskEnd (3 , 0 , " foo" , Success , taskInfo(" 2" , 1 ),
363
+ new ExecutorMetrics , null ))
359
364
assert(monitor.timedOutExecutors(idleDeadline).isEmpty)
360
365
361
366
monitor.onJobEnd(SparkListenerJobEnd (1 , clock.getTimeMillis(), JobSucceeded ))
@@ -385,7 +390,8 @@ class ExecutorMonitorSuite extends SparkFunSuite {
385
390
clock.advance(1000L )
386
391
monitor.onExecutorAdded(SparkListenerExecutorAdded (clock.getTimeMillis(), " 1" , null ))
387
392
monitor.onTaskStart(SparkListenerTaskStart (1 , 0 , taskInfo(" 1" , 1 )))
388
- monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , Success , taskInfo(" 1" , 1 ), null ))
393
+ monitor.onTaskEnd(SparkListenerTaskEnd (1 , 0 , " foo" , Success , taskInfo(" 1" , 1 ),
394
+ new ExecutorMetrics , null ))
389
395
monitor.onJobEnd(SparkListenerJobEnd (1 , clock.getTimeMillis(), JobSucceeded ))
390
396
391
397
assert(monitor.timedOutExecutors(idleDeadline).isEmpty)
0 commit comments