From 4ef73e0b70e3471916e13ab55fe5cc5e47a27628 Mon Sep 17 00:00:00 2001 From: Prashant Singh Date: Mon, 24 Apr 2023 14:31:06 -0700 Subject: [PATCH] Address review feedback --- .../source/TestMetadataTableReadableMetrics.java | 8 ++++---- .../source/TestMetadataTableReadableMetrics.java | 11 ++++++----- 2 files changed, 10 insertions(+), 9 deletions(-) diff --git a/spark/v3.3/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java b/spark/v3.3/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java index f80ea4cdc11d..350add71dc5a 100644 --- a/spark/v3.3/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java +++ b/spark/v3.3/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java @@ -77,8 +77,6 @@ public class TestMetadataTableReadableMetrics extends SparkTestBaseWithCatalog { optional(8, "fixedCol", Types.FixedType.ofLength(3)), optional(9, "binaryCol", Types.BinaryType.get())); - private DataFile dataFile; - public TestMetadataTableReadableMetrics() { // only SparkCatalog supports metadata table sql queries super(SparkCatalogConfig.HIVE); @@ -125,7 +123,8 @@ private Table createPrimitiveTable() throws IOException { createPrimitiveRecord( false, 2, 2L, Float.NaN, 2.0D, new BigDecimal("2.00"), "2", null, null)); - dataFile = FileHelpers.writeDataFile(table, Files.localOutput(temp.newFile()), records); + DataFile dataFile = + FileHelpers.writeDataFile(table, Files.localOutput(temp.newFile()), records); table.newAppend().appendFile(dataFile).commit(); return table; } @@ -192,7 +191,8 @@ private GenericRecord createNestedRecord(Long longCol, Double doubleCol) { @Test public void testPrimitiveColumns() throws Exception { - createPrimitiveTable(); + Table table = createPrimitiveTable(); + DataFile dataFile = table.currentSnapshot().addedDataFiles(table.io()).iterator().next(); Map columnSizeStats = dataFile.columnSizes(); Object[] binaryCol = diff --git a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java index f80ea4cdc11d..416d5eed5b65 100644 --- a/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java +++ b/spark/v3.4/spark/src/test/java/org/apache/iceberg/spark/source/TestMetadataTableReadableMetrics.java @@ -77,8 +77,6 @@ public class TestMetadataTableReadableMetrics extends SparkTestBaseWithCatalog { optional(8, "fixedCol", Types.FixedType.ofLength(3)), optional(9, "binaryCol", Types.BinaryType.get())); - private DataFile dataFile; - public TestMetadataTableReadableMetrics() { // only SparkCatalog supports metadata table sql queries super(SparkCatalogConfig.HIVE); @@ -125,7 +123,8 @@ private Table createPrimitiveTable() throws IOException { createPrimitiveRecord( false, 2, 2L, Float.NaN, 2.0D, new BigDecimal("2.00"), "2", null, null)); - dataFile = FileHelpers.writeDataFile(table, Files.localOutput(temp.newFile()), records); + DataFile dataFile = + FileHelpers.writeDataFile(table, Files.localOutput(temp.newFile()), records); table.newAppend().appendFile(dataFile).commit(); return table; } @@ -143,7 +142,8 @@ private void createNestedTable() throws IOException { createNestedRecord(0L, 0.0), createNestedRecord(1L, Double.NaN), createNestedRecord(null, null)); - dataFile = FileHelpers.writeDataFile(table, Files.localOutput(temp.newFile()), records); + DataFile dataFile = + FileHelpers.writeDataFile(table, Files.localOutput(temp.newFile()), records); table.newAppend().appendFile(dataFile).commit(); } @@ -192,7 +192,8 @@ private GenericRecord createNestedRecord(Long longCol, Double doubleCol) { @Test public void testPrimitiveColumns() throws Exception { - createPrimitiveTable(); + Table table = createPrimitiveTable(); + DataFile dataFile = table.currentSnapshot().addedDataFiles(table.io()).iterator().next(); Map columnSizeStats = dataFile.columnSizes(); Object[] binaryCol =