@@ -2138,49 +2138,6 @@ index 5e01d3f447c..284d6657d4f 100644
21382138 withTempDir { dir =>
21392139 val readSchema =
21402140 new StructType()
2141- diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala
2142- index c10e1799702..ba6629abfd9 100644
2143- --- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala
2144- +++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFileMetadataStructRowIndexSuite.scala
2145- @@ -16,7 +16,7 @@
2146- */
2147- package org.apache.spark.sql.execution.datasources.parquet
2148-
2149- - import org.apache.spark.sql.{AnalysisException, DataFrame, QueryTest}
2150- + import org.apache.spark.sql.{AnalysisException, DataFrame, IgnoreCometNativeDataFusion, QueryTest}
2151- import org.apache.spark.sql.execution.datasources.FileFormat
2152- import org.apache.spark.sql.functions.{col, lit}
2153- import org.apache.spark.sql.internal.SQLConf
2154- @@ -154,7 +154,8 @@ class ParquetFileMetadataStructRowIndexSuite extends QueryTest with SharedSparkS
2155- }
2156- }
2157-
2158- - test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - not present in a table") {
2159- + test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - not present in a table",
2160- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3317")) {
2161- // File format supporting row index generation populates the column with row indexes.
2162- withReadDataFrame("parquet", extraSchemaFields =
2163- Seq(StructField(ROW_INDEX_TEMPORARY_COLUMN_NAME, LongType))) { df =>
2164- @@ -172,7 +173,8 @@ class ParquetFileMetadataStructRowIndexSuite extends QueryTest with SharedSparkS
2165- }
2166- }
2167-
2168- - test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - present in a table") {
2169- + test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - present in a table",
2170- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3317")) {
2171- withReadDataFrame("parquet", extraCol = ROW_INDEX_TEMPORARY_COLUMN_NAME) { df =>
2172- // Values of ROW_INDEX_TEMPORARY_COLUMN_NAME column are always populated with
2173- // generated row indexes, rather than read from the file.
2174- @@ -189,7 +191,8 @@ class ParquetFileMetadataStructRowIndexSuite extends QueryTest with SharedSparkS
2175- }
2176- }
2177-
2178- - test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - as partition col") {
2179- + test(s"reading ${ROW_INDEX_TEMPORARY_COLUMN_NAME} - as partition col",
2180- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3317")) {
2181- withReadDataFrame("parquet", partitionCol = ROW_INDEX_TEMPORARY_COLUMN_NAME) { df =>
2182- // Column values are set for each partition, rather than populated with generated row indexes.
2183- assert(df
21842141diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
21852142index 8e88049f51e..49f2001dc6b 100644
21862143--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetFilterSuite.scala
0 commit comments