@@ -256,7 +256,7 @@ def test_data_files(spark: SparkSession, session_catalog: Catalog, arrow_table_w
256256 identifier = "default.arrow_data_files"
257257 tbl = _create_table (session_catalog , identifier , {"format-version" : "1" }, [])
258258
259- tbl .overwrite (arrow_table_with_null )
259+ tbl .append (arrow_table_with_null )
260260 # should produce a DELETE entry
261261 tbl .overwrite (arrow_table_with_null )
262262 # Since we don't rewrite, this should produce a new manifest with an ADDED entry
@@ -288,7 +288,7 @@ def get_current_snapshot_id(identifier: str) -> int:
288288 .snapshot_id
289289 )
290290
291- tbl .overwrite (arrow_table_with_null )
291+ tbl .append (arrow_table_with_null )
292292 assert tbl .current_snapshot ().snapshot_id == get_current_snapshot_id (identifier ) # type: ignore
293293 tbl .overwrite (arrow_table_with_null )
294294 assert tbl .current_snapshot ().snapshot_id == get_current_snapshot_id (identifier ) # type: ignore
@@ -330,7 +330,7 @@ def test_python_writes_special_character_column_with_spark_reads(
330330 arrow_table_with_special_character_column = pa .Table .from_pydict (TEST_DATA_WITH_SPECIAL_CHARACTER_COLUMN , schema = pa_schema )
331331 tbl = _create_table (session_catalog , identifier , {"format-version" : format_version }, schema = pa_schema )
332332
333- tbl .overwrite (arrow_table_with_special_character_column )
333+ tbl .append (arrow_table_with_special_character_column )
334334 spark_df = spark .sql (f"SELECT * FROM { identifier } " ).toPandas ()
335335 pyiceberg_df = tbl .scan ().to_pandas ()
336336 assert spark_df .equals (pyiceberg_df )
@@ -354,7 +354,7 @@ def test_python_writes_dictionary_encoded_column_with_spark_reads(
354354
355355 tbl = _create_table (session_catalog , identifier , {"format-version" : format_version }, schema = pa_schema )
356356
357- tbl .overwrite (arrow_table )
357+ tbl .append (arrow_table )
358358 spark_df = spark .sql (f"SELECT * FROM { identifier } " ).toPandas ()
359359 pyiceberg_df = tbl .scan ().to_pandas ()
360360 assert spark_df .equals (pyiceberg_df )
@@ -393,7 +393,7 @@ def test_python_writes_with_small_and_large_types_spark_reads(
393393 arrow_table = pa .Table .from_pydict (TEST_DATA , schema = pa_schema )
394394 tbl = _create_table (session_catalog , identifier , {"format-version" : format_version }, schema = pa_schema )
395395
396- tbl .overwrite (arrow_table )
396+ tbl .append (arrow_table )
397397 spark_df = spark .sql (f"SELECT * FROM { identifier } " ).toPandas ()
398398 pyiceberg_df = tbl .scan ().to_pandas ()
399399 assert spark_df .equals (pyiceberg_df )
@@ -429,7 +429,7 @@ def get_data_files_count(identifier: str) -> int:
429429
430430 # writes 1 data file since the table is smaller than default target file size
431431 assert arrow_table_with_null .nbytes < TableProperties .WRITE_TARGET_FILE_SIZE_BYTES_DEFAULT
432- tbl .overwrite (arrow_table_with_null )
432+ tbl .append (arrow_table_with_null )
433433 assert get_data_files_count (identifier ) == 1
434434
435435 # writes 1 data file as long as table is smaller than default target file size
@@ -820,7 +820,7 @@ def test_inspect_snapshots(
820820 identifier = "default.table_metadata_snapshots"
821821 tbl = _create_table (session_catalog , identifier , properties = {"format-version" : format_version })
822822
823- tbl .overwrite (arrow_table_with_null )
823+ tbl .append (arrow_table_with_null )
824824 # should produce a DELETE entry
825825 tbl .overwrite (arrow_table_with_null )
826826 # Since we don't rewrite, this should produce a new manifest with an ADDED entry
@@ -979,6 +979,7 @@ def test_table_write_subset_of_schema(session_catalog: Catalog, arrow_table_with
979979
980980@pytest .mark .integration
981981@pytest .mark .parametrize ("format_version" , [1 , 2 ])
982+ @pytest .mark .filterwarnings ("ignore:Delete operation did not match any records" )
982983def test_table_write_out_of_order_schema (session_catalog : Catalog , arrow_table_with_null : pa .Table , format_version : int ) -> None :
983984 identifier = "default.test_table_write_out_of_order_schema"
984985 # rotate the schema fields by 1
@@ -989,6 +990,7 @@ def test_table_write_out_of_order_schema(session_catalog: Catalog, arrow_table_w
989990 tbl = _create_table (session_catalog , identifier , {"format-version" : format_version }, schema = rotated_schema )
990991
991992 tbl .overwrite (arrow_table_with_null )
993+
992994 tbl .append (arrow_table_with_null )
993995 # overwrite and then append should produce twice the data
994996 assert len (tbl .scan ().to_arrow ()) == len (arrow_table_with_null ) * 2
0 commit comments