@@ -589,7 +589,7 @@ index 81713c777bc..b5f92ed9742 100644
589589 assert(exchanges.size == 2)
590590 }
591591diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
592- index 2c24cc7d570..753737a1057 100644
592+ index 2c24cc7d570..5a1fe7017c3 100644
593593--- a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
594594+++ b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
595595@@ -22,6 +22,7 @@ import org.scalatest.GivenWhenThen
@@ -630,17 +630,7 @@ index 2c24cc7d570..753737a1057 100644
630630 case _ => Nil
631631 }
632632 }
633- @@ -1027,7 +1037,8 @@ abstract class DynamicPartitionPruningSuiteBase
634- }
635- }
636-
637- - test("avoid reordering broadcast join keys to match input hash partitioning") {
638- + test("avoid reordering broadcast join keys to match input hash partitioning",
639- + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/1839")) {
640- withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "false",
641- SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "-1") {
642- withTable("large", "dimTwo", "dimThree") {
643- @@ -1204,10 +1215,16 @@ abstract class DynamicPartitionPruningSuiteBase
633+ @@ -1204,10 +1214,16 @@ abstract class DynamicPartitionPruningSuiteBase
644634
645635 val plan = df.queryExecution.executedPlan
646636 val countSubqueryBroadcasts =
@@ -659,43 +649,43 @@ index 2c24cc7d570..753737a1057 100644
659649
660650 assert(countSubqueryBroadcasts == 1)
661651 assert(countReusedSubqueryBroadcasts == 1)
662- @@ -1215,7 +1232 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
652+ @@ -1215,7 +1231 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
663653 }
664654
665655 test("SPARK-32509: Unused Dynamic Pruning filter shouldn't affect " +
666656- "canonicalization and exchange reuse") {
667657+ "canonicalization and exchange reuse",
668- + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/1839 ")) {
658+ + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/4045 ")) {
669659 withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true") {
670660 withSQLConf(SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "-1") {
671661 val df = sql(
672- @@ -1330,6 +1348 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
662+ @@ -1330,6 +1347 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
673663 }
674664
675665 test("Subquery reuse across the whole plan",
676666+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3313"),
677667 DisableAdaptiveExecution("DPP in AQE must reuse broadcast")) {
678668 withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_ENABLED.key -> "true",
679669 SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "false",
680- @@ -1424,7 +1443 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
670+ @@ -1424,7 +1442 ,8 @@ abstract class DynamicPartitionPruningSuiteBase
681671 }
682672 }
683673
684674- test("SPARK-34637: DPP side broadcast query stage is created firstly") {
685675+ test("SPARK-34637: DPP side broadcast query stage is created firstly",
686- + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/1839 ")) {
676+ + IgnoreComet("TODO: https://github.com/apache/datafusion-comet/issues/4045 ")) {
687677 withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true") {
688678 val df = sql(
689679 """ WITH v as (
690- @@ -1578,6 +1598 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
680+ @@ -1578,6 +1597 ,7 @@ abstract class DynamicPartitionPruningSuiteBase
691681
692682 val subqueryBroadcastExecs = collectWithSubqueries(df.queryExecution.executedPlan) {
693683 case s: SubqueryBroadcastExec => s
694684+ case s: CometSubqueryBroadcastExec => s
695685 }
696686 assert(subqueryBroadcastExecs.size === 1)
697687 subqueryBroadcastExecs.foreach { subqueryBroadcastExec =>
698- @@ -1730,6 +1751 ,10 @@ abstract class DynamicPartitionPruningV1Suite extends DynamicPartitionPruningDat
688+ @@ -1730,6 +1750 ,10 @@ abstract class DynamicPartitionPruningV1Suite extends DynamicPartitionPruningDat
699689 case s: BatchScanExec =>
700690 // we use f1 col for v2 tables due to schema pruning
701691 s.output.exists(_.exists(_.argString(maxFields = 100).contains("f1")))
@@ -1319,24 +1309,23 @@ index 0df7f806272..92390bd819f 100644
13191309
13201310 test("non-matching optional group") {
13211311diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
1322- index 2e33f6505ab..3a8b154b565 100644
1312+ index 2e33f6505ab..949fdea0003 100644
13231313--- a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
13241314+++ b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
1325- @@ -23,11 +23,13 @@ import org.apache.spark.SparkRuntimeException
1315+ @@ -23,10 +23,12 @@ import org.apache.spark.SparkRuntimeException
13261316 import org.apache.spark.sql.catalyst.expressions.SubqueryExpression
13271317 import org.apache.spark.sql.catalyst.plans.{LeftAnti, LeftSemi}
13281318 import org.apache.spark.sql.catalyst.plans.logical.{Aggregate, Filter, Join, LogicalPlan, Project, Sort, Union}
13291319+ import org.apache.spark.sql.comet.{CometNativeColumnarToRowExec, CometNativeScanExec, CometScanExec}
13301320 import org.apache.spark.sql.execution._
1321+ + import org.apache.spark.sql.IgnoreCometNativeDataFusion
13311322 import org.apache.spark.sql.execution.adaptive.{AdaptiveSparkPlanHelper, DisableAdaptiveExecution}
13321323 import org.apache.spark.sql.execution.datasources.FileScanRDD
13331324- import org.apache.spark.sql.execution.exchange.ShuffleExchangeExec
13341325+ import org.apache.spark.sql.execution.exchange.ShuffleExchangeLike
13351326 import org.apache.spark.sql.execution.joins.{BaseJoinExec, BroadcastHashJoinExec, BroadcastNestedLoopJoinExec}
1336- + import org.apache.spark.sql.IgnoreCometNativeDataFusion
13371327 import org.apache.spark.sql.internal.SQLConf
13381328 import org.apache.spark.sql.test.SharedSparkSession
1339-
13401329@@ -1529,6 +1531,18 @@ class SubquerySuite extends QueryTest
13411330 fs.inputRDDs().forall(
13421331 _.asInstanceOf[FileScanRDD].filePartitions.forall(
0 commit comments