Skip to content

Commit

Permalink
style: line length exceeds 100 characters
Browse files Browse the repository at this point in the history
Signed-off-by: Josh Soref <jsoref@users.noreply.github.com>
  • Loading branch information
jsoref committed Dec 7, 2020
1 parent d3d6324 commit db599a5
Show file tree
Hide file tree
Showing 3 changed files with 8 additions and 5 deletions.
Original file line number Diff line number Diff line change
Expand Up @@ -113,7 +113,8 @@ class DatasetCacheSuite extends QueryTest
ds.unpersist(blocking = true)
assert(ds.storageLevel == StorageLevel.NONE, "The Dataset ds should not be cached.")
aggregated.unpersist(blocking = true)
assert(aggregated.storageLevel == StorageLevel.NONE, "The Dataset aggregated should not be cached.")
assert(aggregated.storageLevel == StorageLevel.NONE,
"The Dataset aggregated should not be cached.")
}

test("persist and then withColumn") {
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -857,7 +857,7 @@ class ParquetV1QuerySuite extends ParquetQuerySuite {
val df = spark.range(10).select(Seq.tabulate(11) {i => ('id + i).as(s"c$i")} : _*)
df.write.mode(SaveMode.Overwrite).parquet(path)

// do not return batch, because whole stage codegen is disabled for wide table (>200 columns)
// do not return batch - whole stage codegen is disabled for wide table (>200 columns)
val df2 = spark.read.parquet(path)
val fileScan2 = df2.queryExecution.sparkPlan.find(_.isInstanceOf[FileSourceScanExec]).get
assert(!fileScan2.asInstanceOf[FileSourceScanExec].supportsColumnar)
Expand Down Expand Up @@ -890,7 +890,7 @@ class ParquetV2QuerySuite extends ParquetQuerySuite {
val df = spark.range(10).select(Seq.tabulate(11) {i => ('id + i).as(s"c$i")} : _*)
df.write.mode(SaveMode.Overwrite).parquet(path)

// do not return batch, because whole stage codegen is disabled for wide table (>200 columns)
// do not return batch - whole stage codegen is disabled for wide table (>200 columns)
val df2 = spark.read.parquet(path)
val fileScan2 = df2.queryExecution.sparkPlan.find(_.isInstanceOf[BatchScanExec]).get
val parquetScan2 = fileScan2.asInstanceOf[BatchScanExec].scan.asInstanceOf[ParquetScan]
Expand Down
Original file line number Diff line number Diff line change
Expand Up @@ -644,7 +644,8 @@ abstract class BucketedReadSuite extends QueryTest with SQLTestUtils {
aggregated.sort("i", "j"),
df1.groupBy("i", "j").agg(max("k")).sort("i", "j"))

assert(aggregated.queryExecution.executedPlan.find(_.isInstanceOf[ShuffleExchangeExec]).isEmpty)
assert(
aggregated.queryExecution.executedPlan.find(_.isInstanceOf[ShuffleExchangeExec]).isEmpty)
}
}

Expand Down Expand Up @@ -684,7 +685,8 @@ abstract class BucketedReadSuite extends QueryTest with SQLTestUtils {
aggregated.sort("i", "j"),
df1.groupBy("i", "j").agg(max("k")).sort("i", "j"))

assert(aggregated.queryExecution.executedPlan.find(_.isInstanceOf[ShuffleExchangeExec]).isEmpty)
assert(
aggregated.queryExecution.executedPlan.find(_.isInstanceOf[ShuffleExchangeExec]).isEmpty)
}
}

Expand Down

0 comments on commit db599a5

Please sign in to comment.