@@ -2065,49 +2065,6 @@ index 07e2849ce6f..3e73645b638 100644
20652065 val extraOptions = Map[String, String](
20662066 ParquetOutputFormat.WRITER_VERSION -> ParquetProperties.WriterVersion.PARQUET_2_0.toString
20672067 )
2068- diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala
2069- index c10e1799702..ba6629abfd9 100644
2070- --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala
2071- +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala
2072- @@ -16,7 +16,7 @@
2073- */
2074- package org.apache.spark.sql.execution.datasources.parquet
2075-
2076- - import org.apache.spark.sql.{AnalysisException, DataFrame, QueryTest}
2077- + import org.apache.spark.sql.{AnalysisException, DataFrame, IgnoreCometNativeDataFusion, QueryTest}
2078- import org.apache.spark.sql.execution.datasources.FileFormat
2079- import org.apache.spark.sql.functions.{col, lit}
2080- import org.apache.spark.sql.internal.SQLConf
2081- @@ -154,7 +154,8 @@ class ParquetFileMetadataStructRowIndexSuite extends QueryTest with SharedSparkS
2082- }
2083- }
2084-
2085- - test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - not present in a table") {
2086- + test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - not present in a table",
2087- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3317")) {
2088- // File format supporting row index generation populates the column with row indexes.
2089- withReadDataFrame("parquet", extraSchemaFields =
2090- Seq(StructField(ROW_INDEX_TEMPORARY_COLUMN_NAME, LongType))) { df =>
2091- @@ -172,7 +173,8 @@ class ParquetFileMetadataStructRowIndexSuite extends QueryTest with SharedSparkS
2092- }
2093- }
2094-
2095- - test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - present in a table") {
2096- + test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - present in a table",
2097- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3317")) {
2098- withReadDataFrame("parquet", extraCol = ROW_INDEX_TEMPORARY_COLUMN_NAME) { df =>
2099- // Values of ROW_INDEX_TEMPORARY_COLUMN_NAME column are always populated with
2100- // generated row indexes, rather than read from the file.
2101- @@ -189,7 +191,8 @@ class ParquetFileMetadataStructRowIndexSuite extends QueryTest with SharedSparkS
2102- }
2103- }
2104-
2105- - test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - as partition col") {
2106- + test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - as partition col",
2107- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3317")) {
2108- withReadDataFrame("parquet", partitionCol = ROW_INDEX_TEMPORARY_COLUMN_NAME) { df =>
2109- // Column values are set for each partition, rather than populated with generated row indexes.
2110- assert(df
21112068diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
21122069index 8e88049f51e..49f2001dc6b 100644
21132070--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
0 commit comments