Skip to content

Commit b98a736

Browse files
authored
test: unignore passing Spark 4.0 SQL tests and retag remaining to specific issues (#4041)
1 parent 33f2764 commit b98a736

1 file changed

Lines changed: 31 additions & 51 deletions

File tree

dev/diffs/4.0.1.diff

Lines changed: 31 additions & 51 deletions
Original file line numberDiff line numberDiff line change
@@ -589,7 +589,7 @@ index 81713c777bc..b5f92ed9742 100644
589589
assert(exchanges.size == 2)
590590
}
591591
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
592-
index 2c24cc7d570..63047ef482e 100644
592+
index 2c24cc7d570..753737a1057 100644
593593
--- a/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
594594
+++ b/sql/core/src/test/scala/org/apache/spark/sql/DynamicPartitionPruningSuite.scala
595595
@@ -22,6 +22,7 @@ import org.scalatest.GivenWhenThen
@@ -640,17 +640,7 @@ index 2c24cc7d570..63047ef482e 100644
640640
withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "false",
641641
SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "-1") {
642642
withTable("large", "dimTwo", "dimThree") {
643-
@@ -1151,7 +1162,8 @@ abstract class DynamicPartitionPruningSuiteBase
644-
}
645-
}
646-
647-
- test("join key with multiple references on the filtering plan") {
648-
+ test("join key with multiple references on the filtering plan",
649-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
650-
withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true",
651-
SQLConf.ADAPTIVE_OPTIMIZER_EXCLUDED_RULES.key -> AQEPropagateEmptyRelation.ruleName,
652-
SQLConf.ANSI_ENABLED.key -> "false" // ANSI mode doesn't support "String + String"
653-
@@ -1204,10 +1216,16 @@ abstract class DynamicPartitionPruningSuiteBase
643+
@@ -1204,10 +1215,16 @@ abstract class DynamicPartitionPruningSuiteBase
654644

655645
val plan = df.queryExecution.executedPlan
656646
val countSubqueryBroadcasts =
@@ -669,7 +659,7 @@ index 2c24cc7d570..63047ef482e 100644
669659

670660
assert(countSubqueryBroadcasts == 1)
671661
assert(countReusedSubqueryBroadcasts == 1)
672-
@@ -1215,7 +1233,8 @@ abstract class DynamicPartitionPruningSuiteBase
662+
@@ -1215,7 +1232,8 @@ abstract class DynamicPartitionPruningSuiteBase
673663
}
674664

675665
test("SPARK-32509: Unused Dynamic Pruning filter shouldn't affect " +
@@ -679,15 +669,15 @@ index 2c24cc7d570..63047ef482e 100644
679669
withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true") {
680670
withSQLConf(SQLConf.AUTO_BROADCASTJOIN_THRESHOLD.key -> "-1") {
681671
val df = sql(
682-
@@ -1330,6 +1349,7 @@ abstract class DynamicPartitionPruningSuiteBase
672+
@@ -1330,6 +1348,7 @@ abstract class DynamicPartitionPruningSuiteBase
683673
}
684674

685675
test("Subquery reuse across the whole plan",
686-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321"),
676+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3313"),
687677
DisableAdaptiveExecution("DPP in AQE must reuse broadcast")) {
688678
withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_ENABLED.key -> "true",
689679
SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "false",
690-
@@ -1424,7 +1444,8 @@ abstract class DynamicPartitionPruningSuiteBase
680+
@@ -1424,7 +1443,8 @@ abstract class DynamicPartitionPruningSuiteBase
691681
}
692682
}
693683

@@ -697,15 +687,15 @@ index 2c24cc7d570..63047ef482e 100644
697687
withSQLConf(SQLConf.DYNAMIC_PARTITION_PRUNING_REUSE_BROADCAST_ONLY.key -> "true") {
698688
val df = sql(
699689
""" WITH v as (
700-
@@ -1578,6 +1599,7 @@ abstract class DynamicPartitionPruningSuiteBase
690+
@@ -1578,6 +1598,7 @@ abstract class DynamicPartitionPruningSuiteBase
701691

702692
val subqueryBroadcastExecs = collectWithSubqueries(df.queryExecution.executedPlan) {
703693
case s: SubqueryBroadcastExec => s
704694
+ case s: CometSubqueryBroadcastExec => s
705695
}
706696
assert(subqueryBroadcastExecs.size === 1)
707697
subqueryBroadcastExecs.foreach { subqueryBroadcastExec =>
708-
@@ -1730,6 +1752,10 @@ abstract class DynamicPartitionPruningV1Suite extends DynamicPartitionPruningDat
698+
@@ -1730,6 +1751,10 @@ abstract class DynamicPartitionPruningV1Suite extends DynamicPartitionPruningDat
709699
case s: BatchScanExec =>
710700
// we use f1 col for v2 tables due to schema pruning
711701
s.output.exists(_.exists(_.argString(maxFields = 100).contains("f1")))
@@ -742,7 +732,7 @@ index 9c90e0105a4..fadf2f0f698 100644
742732

743733
test("SPARK-35884: Explain Formatted") {
744734
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
745-
index 9c529d14221..6d5db65b5d8 100644
735+
index 9c529d14221..a046f1ed1ca 100644
746736
--- a/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
747737
+++ b/sql/core/src/test/scala/org/apache/spark/sql/FileBasedDataSourceSuite.scala
748738
@@ -33,6 +33,8 @@ import org.apache.spark.sql.catalyst.expressions.{AttributeReference, GreaterTha
@@ -761,7 +751,7 @@ index 9c529d14221..6d5db65b5d8 100644
761751
- testQuietly(s"Enabling/disabling ignoreMissingFiles using $format") {
762752
+ val ignoreMissingTags: Seq[org.scalatest.Tag] = if (format == "parquet") {
763753
+ Seq(IgnoreCometNativeDataFusion(
764-
+ "https://github.com/apache/datafusion-comet/issues/3321"))
754+
+ "https://github.com/apache/datafusion-comet/issues/3314"))
765755
+ } else Seq.empty
766756
+ test(s"Enabling/disabling ignoreMissingFiles using $format", ignoreMissingTags: _*) { quietly {
767757
def testIgnoreMissingFiles(options: Map[String, String]): Unit = {
@@ -1329,10 +1319,10 @@ index 0df7f806272..92390bd819f 100644
13291319

13301320
test("non-matching optional group") {
13311321
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
1332-
index 2e33f6505ab..6d4a75d02ff 100644
1322+
index 2e33f6505ab..3a8b154b565 100644
13331323
--- a/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
13341324
+++ b/sql/core/src/test/scala/org/apache/spark/sql/SubquerySuite.scala
1335-
@@ -23,12 +23,14 @@ import org.apache.spark.SparkRuntimeException
1325+
@@ -23,11 +23,13 @@ import org.apache.spark.SparkRuntimeException
13361326
import org.apache.spark.sql.catalyst.expressions.SubqueryExpression
13371327
import org.apache.spark.sql.catalyst.plans.{LeftAnti, LeftSemi}
13381328
import org.apache.spark.sql.catalyst.plans.logical.{Aggregate, Filter, Join, LogicalPlan, Project, Sort, Union}
@@ -1343,11 +1333,10 @@ index 2e33f6505ab..6d4a75d02ff 100644
13431333
-import org.apache.spark.sql.execution.exchange.ShuffleExchangeExec
13441334
+import org.apache.spark.sql.execution.exchange.ShuffleExchangeLike
13451335
import org.apache.spark.sql.execution.joins.{BaseJoinExec, BroadcastHashJoinExec, BroadcastNestedLoopJoinExec}
1346-
import org.apache.spark.sql.internal.SQLConf
13471336
+import org.apache.spark.sql.IgnoreCometNativeDataFusion
1337+
import org.apache.spark.sql.internal.SQLConf
13481338
import org.apache.spark.sql.test.SharedSparkSession
13491339

1350-
class SubquerySuite extends QueryTest
13511340
@@ -1529,6 +1531,18 @@ class SubquerySuite extends QueryTest
13521341
fs.inputRDDs().forall(
13531342
_.asInstanceOf[FileScanRDD].filePartitions.forall(
@@ -1382,7 +1371,7 @@ index 2e33f6505ab..6d4a75d02ff 100644
13821371

13831372
- test("SPARK-43402: FileSourceScanExec supports push down data filter with scalar subquery") {
13841373
+ test("SPARK-43402: FileSourceScanExec supports push down data filter with scalar subquery",
1385-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
1374+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3315")) {
13861375
def checkFileSourceScan(query: String, answer: Seq[Row]): Unit = {
13871376
val df = sql(query)
13881377
checkAnswer(df, answer)
@@ -1902,7 +1891,7 @@ index 47679ed7865..9ffbaecb98e 100644
19021891
assert(collectWithSubqueries(plan) { case s: SortAggregateExec => s }.length == sortAggCount)
19031892
}
19041893
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala
1905-
index 77a988f340e..e4deeb6b1d8 100644
1894+
index 77a988f340e..263208a67d9 100644
19061895
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala
19071896
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/SQLViewSuite.scala
19081897
@@ -1061,7 +1061,8 @@ abstract class SQLViewSuite extends QueryTest with SQLTestUtils {
@@ -1911,7 +1900,7 @@ index 77a988f340e..e4deeb6b1d8 100644
19111900

19121901
- test("alter temporary view should follow current storeAnalyzedPlanForView config") {
19131902
+ test("alter temporary view should follow current storeAnalyzedPlanForView config",
1914-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
1903+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3314")) {
19151904
withTable("t") {
19161905
Seq(2, 3, 1).toDF("c1").write.format("parquet").saveAsTable("t")
19171906
withView("v1") {
@@ -2898,7 +2887,7 @@ index 4474ec1fd42..05fa0257c82 100644
28982887
checkAnswer(
28992888
// "fruit" column in this file is encoded using DELTA_LENGTH_BYTE_ARRAY.
29002889
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
2901-
index bba71f1c48d..e1b0c25a354 100644
2890+
index bba71f1c48d..faee9b4ce83 100644
29022891
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
29032892
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetQuerySuite.scala
29042893
@@ -27,6 +27,7 @@ import org.apache.parquet.hadoop.ParquetOutputFormat
@@ -2925,7 +2914,7 @@ index bba71f1c48d..e1b0c25a354 100644
29252914

29262915
- test("Enabling/disabling ignoreCorruptFiles") {
29272916
+ test("Enabling/disabling ignoreCorruptFiles",
2928-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
2917+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3314")) {
29292918
def testIgnoreCorruptFiles(options: Map[String, String]): Unit = {
29302919
withTempDir { dir =>
29312920
val basePath = dir.getCanonicalPath
@@ -3122,7 +3111,7 @@ index 0acb21f3e6f..1f9c3fd13fc 100644
31223111

31233112
withTempPath { dir =>
31243113
diff --git a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala
3125-
index 09ed6955a51..98e313cddd4 100644
3114+
index 09ed6955a51..5cd856ff7b6 100644
31263115
--- a/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala
31273116
+++ b/sql/core/src/test/scala/org/apache/spark/sql/execution/datasources/parquet/ParquetTypeWideningSuite.scala
31283117
@@ -24,7 +24,7 @@ import org.apache.parquet.format.converter.ParquetMetadataConverter
@@ -3154,72 +3143,63 @@ index 09ed6955a51..98e313cddd4 100644
31543143
}
31553144
}
31563145

3157-
@@ -190,10 +192,16 @@ class ParquetTypeWideningSuite
3146+
@@ -190,7 +192,8 @@ class ParquetTypeWideningSuite
31583147
(Seq("1", "2", Short.MinValue.toString), ShortType, DoubleType),
31593148
(Seq("1", "2", Int.MinValue.toString), IntegerType, DoubleType),
31603149
(Seq("1.23", "10.34"), FloatType, DoubleType),
31613150
- (Seq("2020-01-01", "2020-01-02", "1312-02-27"), DateType, TimestampNTZType)
31623151
+ // TODO: Comet cannot handle older than "1582-10-15"
31633152
+ (Seq("2020-01-01", "2020-01-02"/* , "1312-02-27" */), DateType, TimestampNTZType)
31643153
)
3165-
+ wideningTags: Seq[org.scalatest.Tag] =
3166-
+ if (fromType == DateType && toType == TimestampNTZType) {
3167-
+ Seq(IgnoreCometNativeDataFusion(
3168-
+ "https://github.com/apache/datafusion-comet/issues/3321"))
3169-
+ } else Seq.empty
3170-
}
3171-
- test(s"parquet widening conversion $fromType -> $toType") {
3172-
+ test(s"parquet widening conversion $fromType -> $toType", wideningTags: _*) {
3173-
checkAllParquetReaders(values, fromType, toType, expectError = false)
31743154
}
3175-
3176-
@@ -231,7 +239,8 @@ class ParquetTypeWideningSuite
3155+
test(s"parquet widening conversion $fromType -> $toType") {
3156+
@@ -231,7 +234,8 @@ class ParquetTypeWideningSuite
31773157
(Seq("2020-01-01", "2020-01-02", "1312-02-27"), DateType, TimestampType)
31783158
)
31793159
}
31803160
- test(s"unsupported parquet conversion $fromType -> $toType") {
31813161
+ test(s"unsupported parquet conversion $fromType -> $toType",
3182-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
3162+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720")) {
31833163
checkAllParquetReaders(values, fromType, toType, expectError = true)
31843164
}
31853165

3186-
@@ -257,7 +266,8 @@ class ParquetTypeWideningSuite
3166+
@@ -257,7 +261,8 @@ class ParquetTypeWideningSuite
31873167
(Seq("1", "2"), LongType, DecimalType(LongDecimal.precision, 1))
31883168
)
31893169
}
31903170
- test(s"unsupported parquet conversion $fromType -> $toType") {
31913171
+ test(s"unsupported parquet conversion $fromType -> $toType",
3192-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
3172+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720")) {
31933173
checkAllParquetReaders(values, fromType, toType,
31943174
expectError =
31953175
// parquet-mr allows reading decimals into a smaller precision decimal type without
3196-
@@ -271,7 +281,8 @@ class ParquetTypeWideningSuite
3176+
@@ -271,7 +276,8 @@ class ParquetTypeWideningSuite
31973177
(Seq("2020-01-01", "2020-01-02", "1312-02-27"), TimestampNTZType, DateType))
31983178
outputTimestampType <- ParquetOutputTimestampType.values
31993179
}
32003180
- test(s"unsupported parquet timestamp conversion $fromType ($outputTimestampType) -> $toType") {
32013181
+ test(s"unsupported parquet timestamp conversion $fromType ($outputTimestampType) -> $toType",
3202-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
3182+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720")) {
32033183
withSQLConf(
32043184
SQLConf.PARQUET_OUTPUT_TIMESTAMP_TYPE.key -> outputTimestampType.toString,
32053185
SQLConf.PARQUET_INT96_REBASE_MODE_IN_WRITE.key -> LegacyBehaviorPolicy.CORRECTED.toString
3206-
@@ -291,7 +302,8 @@ class ParquetTypeWideningSuite
3186+
@@ -291,7 +297,8 @@ class ParquetTypeWideningSuite
32073187
Seq(7 -> 5, 10 -> 5, 20 -> 5, 12 -> 10, 20 -> 10, 22 -> 20)
32083188
}
32093189
test(
32103190
- s"parquet decimal precision change Decimal($fromPrecision, 2) -> Decimal($toPrecision, 2)") {
32113191
+ s"parquet decimal precision change Decimal($fromPrecision, 2) -> Decimal($toPrecision, 2)",
3212-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")) {
3192+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720")) {
32133193
checkAllParquetReaders(
32143194
values = Seq("1.23", "10.34"),
32153195
fromType = DecimalType(fromPrecision, 2),
3216-
@@ -322,7 +334,8 @@ class ParquetTypeWideningSuite
3196+
@@ -322,7 +329,8 @@ class ParquetTypeWideningSuite
32173197
Seq((5, 2) -> (6, 4), (10, 4) -> (12, 7), (20, 5) -> (22, 8))
32183198
}
32193199
test(s"parquet decimal precision and scale change Decimal($fromPrecision, $fromScale) -> " +
32203200
- s"Decimal($toPrecision, $toScale)"
32213201
+ s"Decimal($toPrecision, $toScale)",
3222-
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3321")
3202+
+ IgnoreCometNativeDataFusion("https://github.com/apache/datafusion-comet/issues/3720")
32233203
) {
32243204
checkAllParquetReaders(
32253205
values = Seq("1.23", "10.34"),

0 commit comments

Comments
 (0)