@@ -589,7 +589,7 @@ index 81713c777bc..b5f92ed9742 100644
589589 assert(exchanges.size == 2)
590590 }
591591diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
592- index 2c24cc7d570..63047ef482e 100644
592+ index 2c24cc7d570..753737a1057 100644
593593--- a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
594594+++ b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
595595@@ -22,6 +22,7 @@ import org.scalatest.GivenWhenThen
@@ -640,17 +640,7 @@ index 2c24cc7d570..63047ef482e 100644
640640 withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "false",
641641 SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "-1") {
642642 withTable("large", "dimTwo", "dimThree") {
643- @@ -1151,7 +1162,8 @@ abstract class DynamicPartitionPruningSuiteBase
644- }
645- }
646-
647- - test("join key with multiple references on the filtering plan") {
648- + test("join key with multiple references on the filtering plan",
649- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
650- withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true",
651- SQLConf.ADAPTIVE_OPTIMIZER_EXCLUDED_RULES.key -> AQEPropagateEmptyRelation.ruleName,
652- SQLConf.ANSI_ENABLED.key -> "false" // ANSI mode doesn't support "String + String"
653- @@ -1204,10 +1216,16 @@ abstract class DynamicPartitionPruningSuiteBase
643+ @@ -1204,10 +1215,16 @@ abstract class DynamicPartitionPruningSuiteBase
654644
655645 val plan = df.queryExecution.executedPlan
656646 val countSubqueryBroadcasts =
@@ -669,7 +659,7 @@ index 2c24cc7d570..63047ef482e 100644
669659
670660 assert(countSubqueryBroadcasts == 1)
671661 assert(countReusedSubqueryBroadcasts == 1)
672- @@ -1215,7 +1233 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
662+ @@ -1215,7 +1232 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
673663 }
674664
675665 test("SPARK-32509: Unused Dynamic Pruning filter shouldn't affect " +
@@ -679,15 +669,15 @@ index 2c24cc7d570..63047ef482e 100644
679669 withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true") {
680670 withSQLConf(SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "-1") {
681671 val df = sql(
682- @@ -1330,6 +1349 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
672+ @@ -1330,6 +1348 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
683673 }
684674
685675 test("Subquery reuse across the whole plan",
686- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 "),
676+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3313 "),
687677 DisableAdaptiveExecution("DPP in AQE must reuse broadcast")) {
688678 withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_ENABLED.key -> "true",
689679 SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "false",
690- @@ -1424,7 +1444 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
680+ @@ -1424,7 +1443 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
691681 }
692682 }
693683
@@ -697,15 +687,15 @@ index 2c24cc7d570..63047ef482e 100644
697687 withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true") {
698688 val df = sql(
699689 """ WITH v as (
700- @@ -1578,6 +1599 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
690+ @@ -1578,6 +1598 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
701691
702692 val subqueryBroadcastExecs = collectWithSubqueries(df.queryExecution.executedPlan) {
703693 case s: SubqueryBroadcastExec => s
704694+ case s: CometSubqueryBroadcastExec => s
705695 }
706696 assert(subqueryBroadcastExecs.size === 1)
707697 subqueryBroadcastExecs.foreach { subqueryBroadcastExec =>
708- @@ -1730,6 +1752 ,10 @@ abstract class DynamicPartitionPruningV1Suite extends DynamicPartitionPruningDat
698+ @@ -1730,6 +1751 ,10 @@ abstract class DynamicPartitionPruningV1Suite extends DynamicPartitionPruningDat
709699 case s: BatchScanExec =>
710700 // we use f1 col for v2 tables due to schema pruning
711701 s.output.exists(_.exists(_.argString(maxFields = 100).contains("f1")))
@@ -742,7 +732,7 @@ index 9c90e0105a4..fadf2f0f698 100644
742732
743733 test("SPARK-35884: Explain Formatted") {
744734diff --git a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
745- index 9c529d14221..6d5db65b5d8 100644
735+ index 9c529d14221..a046f1ed1ca 100644
746736--- a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
747737+++ b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
748738@@ -33,6 +33,8 @@ import org.apache.spark.sql.catalyst.expressions.{AttributeReference, GreaterTha
@@ -761,7 +751,7 @@ index 9c529d14221..6d5db65b5d8 100644
761751- testQuietly(s"Enabling/disabling ignoreMissingFiles using $format") {
762752+ val ignoreMissingTags: Seq[org.scalatest.Tag] = if (format == "parquet") {
763753+ Seq(IgnoreCometNativeDataFusion(
764- + "https://github.com/apache/datafusion-comet/issues/3321 "))
754+ + "https://github.com/apache/datafusion-comet/issues/3314 "))
765755+ } else Seq.empty
766756+ test(s"Enabling/disabling ignoreMissingFiles using $format", ignoreMissingTags: _*) { quietly {
767757 def testIgnoreMissingFiles(options: Map[String, String]): Unit = {
@@ -1329,10 +1319,10 @@ index 0df7f806272..92390bd819f 100644
13291319
13301320 test("non-matching optional group") {
13311321diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
1332- index 2e33f6505ab..6d4a75d02ff 100644
1322+ index 2e33f6505ab..3a8b154b565 100644
13331323--- a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
13341324+++ b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
1335- @@ -23,12 +23,14 @@ import org.apache.spark.SparkRuntimeException
1325+ @@ -23,11 +23,13 @@ import org.apache.spark.SparkRuntimeException
13361326 import org.apache.spark.sql.catalyst.expressions.SubqueryExpression
13371327 import org.apache.spark.sql.catalyst.plans.{LeftAnti, LeftSemi}
13381328 import org.apache.spark.sql.catalyst.plans.logical.{Aggregate, Filter, Join, LogicalPlan, Project, Sort, Union}
@@ -1343,11 +1333,10 @@ index 2e33f6505ab..6d4a75d02ff 100644
13431333- import org.apache.spark.sql.execution.exchange.ShuffleExchangeExec
13441334+ import org.apache.spark.sql.execution.exchange.ShuffleExchangeLike
13451335 import org.apache.spark.sql.execution.joins.{BaseJoinExec, BroadcastHashJoinExec, BroadcastNestedLoopJoinExec}
1346- import org.apache.spark.sql.internal.SQLConf
13471336+ import org.apache.spark.sql.IgnoreCometNativeDataFusion
1337+ import org.apache.spark.sql.internal.SQLConf
13481338 import org.apache.spark.sql.test.SharedSparkSession
13491339
1350- class SubquerySuite extends QueryTest
13511340@@ -1529,6 +1531,18 @@ class SubquerySuite extends QueryTest
13521341 fs.inputRDDs().forall(
13531342 _.asInstanceOf[FileScanRDD].filePartitions.forall(
@@ -1382,7 +1371,7 @@ index 2e33f6505ab..6d4a75d02ff 100644
13821371
13831372- test("SPARK-43402: FileSourceScanExec supports push down data filter with scalar subquery") {
13841373+ test("SPARK-43402: FileSourceScanExec supports push down data filter with scalar subquery",
1385- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")) {
1374+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3315 ")) {
13861375 def checkFileSourceScan(query: String, answer: Seq[Row]): Unit = {
13871376 val df = sql(query)
13881377 checkAnswer(df, answer)
@@ -1902,7 +1891,7 @@ index 47679ed7865..9ffbaecb98e 100644
19021891 assert(collectWithSubqueries(plan) { case s: SortAggregateExec => s }.length == sortAggCount)
19031892 }
19041893diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala
1905- index 77a988f340e..e4deeb6b1d8 100644
1894+ index 77a988f340e..263208a67d9 100644
19061895--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala
19071896+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala
19081897@@ -1061,7 +1061,8 @@ abstract class SQLViewSuite extends QueryTest with SQLTestUtils {
@@ -1911,7 +1900,7 @@ index 77a988f340e..e4deeb6b1d8 100644
19111900
19121901- test("alter temporary view should follow current storeAnalyzedPlanForView config") {
19131902+ test("alter temporary view should follow current storeAnalyzedPlanForView config",
1914- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")) {
1903+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3314 ")) {
19151904 withTable("t") {
19161905 Seq(2, 3, 1).toDF("c1").write.format("parquet").saveAsTable("t")
19171906 withView("v1") {
@@ -2898,7 +2887,7 @@ index 4474ec1fd42..05fa0257c82 100644
28982887 checkAnswer(
28992888 // "fruit" column in this file is encoded using DELTA_LENGTH_BYTE_ARRAY.
29002889diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
2901- index bba71f1c48d..e1b0c25a354 100644
2890+ index bba71f1c48d..faee9b4ce83 100644
29022891--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
29032892+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
29042893@@ -27,6 +27,7 @@ import org.apache.parquet.hadoop.ParquetOutputFormat
@@ -2925,7 +2914,7 @@ index bba71f1c48d..e1b0c25a354 100644
29252914
29262915- test("Enabling/disabling ignoreCorruptFiles") {
29272916+ test("Enabling/disabling ignoreCorruptFiles",
2928- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")) {
2917+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3314 ")) {
29292918 def testIgnoreCorruptFiles(options: Map[String, String]): Unit = {
29302919 withTempDir { dir =>
29312920 val basePath = dir.getCanonicalPath
@@ -3122,7 +3111,7 @@ index 0acb21f3e6f..1f9c3fd13fc 100644
31223111
31233112 withTempPath { dir =>
31243113diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala
3125- index 09ed6955a51..98e313cddd4 100644
3114+ index 09ed6955a51..5cd856ff7b6 100644
31263115--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala
31273116+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala
31283117@@ -24,7 +24,7 @@ import org.apache.parquet.format.converter.ParquetMetadataConverter
@@ -3154,72 +3143,63 @@ index 09ed6955a51..98e313cddd4 100644
31543143 }
31553144 }
31563145
3157- @@ -190,10 +192,16 @@ class ParquetTypeWideningSuite
3146+ @@ -190,7 +192,8 @@ class ParquetTypeWideningSuite
31583147 (Seq("1", "2", Short.MinValue.toString), ShortType, DoubleType),
31593148 (Seq("1", "2", Int.MinValue.toString), IntegerType, DoubleType),
31603149 (Seq("1.23", "10.34"), FloatType, DoubleType),
31613150- (Seq("2020-01-01", "2020-01-02", "1312-02-27"), DateType, TimestampNTZType)
31623151+ // TODO: Comet cannot handle older than "1582-10-15"
31633152+ (Seq("2020-01-01", "2020-01-02"/* , "1312-02-27" */), DateType, TimestampNTZType)
31643153 )
3165- + wideningTags: Seq[org.scalatest.Tag] =
3166- + if (fromType == DateType && toType == TimestampNTZType) {
3167- + Seq(IgnoreCometNativeDataFusion(
3168- + "https://github.com/apache/datafusion-comet/issues/3321"))
3169- + } else Seq.empty
3170- }
3171- - test(s"parquet widening conversion $fromType -> $toType") {
3172- + test(s"parquet widening conversion $fromType -> $toType", wideningTags: _*) {
3173- checkAllParquetReaders(values, fromType, toType, expectError = false)
31743154 }
3175-
3176- @@ -231,7 +239 ,8 @@ class ParquetTypeWideningSuite
3155+ test(s"parquet widening conversion $fromType -> $toType") {
3156+ @@ -231,7 +234 ,8 @@ class ParquetTypeWideningSuite
31773157 (Seq("2020-01-01", "2020-01-02", "1312-02-27"), DateType, TimestampType)
31783158 )
31793159 }
31803160- test(s"unsupported parquet conversion $fromType -> $toType") {
31813161+ test(s"unsupported parquet conversion $fromType -> $toType",
3182- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")) {
3162+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720 ")) {
31833163 checkAllParquetReaders(values, fromType, toType, expectError = true)
31843164 }
31853165
3186- @@ -257,7 +266 ,8 @@ class ParquetTypeWideningSuite
3166+ @@ -257,7 +261 ,8 @@ class ParquetTypeWideningSuite
31873167 (Seq("1", "2"), LongType, DecimalType(LongDecimal.precision, 1))
31883168 )
31893169 }
31903170- test(s"unsupported parquet conversion $fromType -> $toType") {
31913171+ test(s"unsupported parquet conversion $fromType -> $toType",
3192- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")) {
3172+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720 ")) {
31933173 checkAllParquetReaders(values, fromType, toType,
31943174 expectError =
31953175 // parquet-mr allows reading decimals into a smaller precision decimal type without
3196- @@ -271,7 +281 ,8 @@ class ParquetTypeWideningSuite
3176+ @@ -271,7 +276 ,8 @@ class ParquetTypeWideningSuite
31973177 (Seq("2020-01-01", "2020-01-02", "1312-02-27"), TimestampNTZType, DateType))
31983178 outputTimestampType <- ParquetOutputTimestampType.values
31993179 }
32003180- test(s"unsupported parquet timestamp conversion $fromType ($outputTimestampType) -> $toType") {
32013181+ test(s"unsupported parquet timestamp conversion $fromType ($outputTimestampType) -> $toType",
3202- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")) {
3182+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720 ")) {
32033183 withSQLConf(
32043184 SQLConf.PARQUET_OUTPUT_TIMESTAMP_TYPE.key -> outputTimestampType.toString,
32053185 SQLConf.PARQUET_INT96_REBASE_MODE_IN_WRITE.key -> LegacyBehaviorPolicy.CORRECTED.toString
3206- @@ -291,7 +302 ,8 @@ class ParquetTypeWideningSuite
3186+ @@ -291,7 +297 ,8 @@ class ParquetTypeWideningSuite
32073187 Seq(7 -> 5, 10 -> 5, 20 -> 5, 12 -> 10, 20 -> 10, 22 -> 20)
32083188 }
32093189 test(
32103190- s"parquet decimal precision change Decimal($fromPrecision, 2) -> Decimal($toPrecision, 2)") {
32113191+ s"parquet decimal precision change Decimal($fromPrecision, 2) -> Decimal($toPrecision, 2)",
3212- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")) {
3192+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720 ")) {
32133193 checkAllParquetReaders(
32143194 values = Seq("1.23", "10.34"),
32153195 fromType = DecimalType(fromPrecision, 2),
3216- @@ -322,7 +334 ,8 @@ class ParquetTypeWideningSuite
3196+ @@ -322,7 +329 ,8 @@ class ParquetTypeWideningSuite
32173197 Seq((5, 2) -> (6, 4), (10, 4) -> (12, 7), (20, 5) -> (22, 8))
32183198 }
32193199 test(s"parquet decimal precision and scale change Decimal($fromPrecision, $fromScale) -> " +
32203200- s"Decimal($toPrecision, $toScale)"
32213201+ s"Decimal($toPrecision, $toScale)",
3222- + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321 ")
3202+ + IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720 ")
32233203 ) {
32243204 checkAllParquetReaders(
32253205 values = Seq("1.23", "10.34"),
0 commit comments