@@ -38,7 +38,7 @@ import org.apache.spark.sql.types._
3838import org .apache .spark .sql .vectorized .ColumnVector
3939
4040import org .apache .comet .{CometConf , WithHdfsCluster }
41- import org .apache .comet .CometConf .{SCAN_NATIVE_COMET , SCAN_NATIVE_DATAFUSION , SCAN_NATIVE_ICEBERG_COMPAT }
41+ import org .apache .comet .CometConf .{SCAN_NATIVE_DATAFUSION , SCAN_NATIVE_ICEBERG_COMPAT }
4242import org .apache .comet .parquet .BatchReader
4343
4444/**
@@ -67,14 +67,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
6767 spark.sql(s " select $query from parquetV1Table " ).noop()
6868 }
6969
70- sqlBenchmark.addCase(" SQL Parquet - Comet" ) { _ =>
71- withSQLConf(
72- CometConf .COMET_ENABLED .key -> " true" ,
73- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
74- spark.sql(s " select $query from parquetV1Table " ).noop()
75- }
76- }
77-
7870 sqlBenchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
7971 withSQLConf(
8072 CometConf .COMET_ENABLED .key -> " true" ,
@@ -175,21 +167,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
175167 }
176168 }
177169
178- sqlBenchmark.addCase(" SQL Parquet - Comet" ) { _ =>
179- withSQLConf(
180- " spark.memory.offHeap.enabled" -> " true" ,
181- " spark.memory.offHeap.size" -> " 10g" ,
182- CometConf .COMET_ENABLED .key -> " true" ,
183- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ,
184- DecryptionPropertiesFactory .CRYPTO_FACTORY_CLASS_PROPERTY_NAME -> cryptoFactoryClass,
185- KeyToolkit .KMS_CLIENT_CLASS_PROPERTY_NAME ->
186- " org.apache.parquet.crypto.keytools.mocks.InMemoryKMS" ,
187- InMemoryKMS .KEY_LIST_PROPERTY_NAME ->
188- s " footerKey: ${footerKey}, key1: ${key1}" ) {
189- spark.sql(s " select $query from parquetV1Table " ).noop()
190- }
191- }
192-
193170 sqlBenchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
194171 withSQLConf(
195172 " spark.memory.offHeap.enabled" -> " true" ,
@@ -245,14 +222,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
245222 spark.sql(" select sum(id) from parquetV1Table" ).noop()
246223 }
247224
248- sqlBenchmark.addCase(" SQL Parquet - Comet" ) { _ =>
249- withSQLConf(
250- CometConf .COMET_ENABLED .key -> " true" ,
251- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
252- spark.sql(" select sum(id) from parquetV1Table" ).noop()
253- }
254- }
255-
256225 sqlBenchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
257226 withSQLConf(
258227 CometConf .COMET_ENABLED .key -> " true" ,
@@ -373,14 +342,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
373342 spark.sql(" select sum(c2) from parquetV1Table where c1 + 1 > 0" ).noop()
374343 }
375344
376- benchmark.addCase(" SQL Parquet - Comet" ) { _ =>
377- withSQLConf(
378- CometConf .COMET_ENABLED .key -> " true" ,
379- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
380- spark.sql(" select sum(c2) from parquetV1Table where c1 + 1 > 0" ).noop()
381- }
382- }
383-
384345 benchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
385346 withSQLConf(
386347 CometConf .COMET_ENABLED .key -> " true" ,
@@ -431,14 +392,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
431392 spark.sql(" select sum(length(id)) from parquetV1Table" ).noop()
432393 }
433394
434- sqlBenchmark.addCase(" SQL Parquet - Comet" ) { _ =>
435- withSQLConf(
436- CometConf .COMET_ENABLED .key -> " true" ,
437- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
438- spark.sql(" select sum(length(id)) from parquetV1Table" ).noop()
439- }
440- }
441-
442395 sqlBenchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
443396 withSQLConf(
444397 CometConf .COMET_ENABLED .key -> " true" ,
@@ -482,17 +435,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
482435 .noop()
483436 }
484437
485- benchmark.addCase(" SQL Parquet - Comet" ) { _ =>
486- withSQLConf(
487- CometConf .COMET_ENABLED .key -> " true" ,
488- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
489- spark
490- .sql(" select sum(length(c2)) from parquetV1Table where c1 is " +
491- " not NULL and c2 is not NULL" )
492- .noop()
493- }
494- }
495-
496438 benchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
497439 withSQLConf(
498440 CometConf .COMET_ENABLED .key -> " true" ,
@@ -538,14 +480,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
538480 spark.sql(s " SELECT sum(c $middle) FROM parquetV1Table " ).noop()
539481 }
540482
541- benchmark.addCase(" SQL Parquet - Comet" ) { _ =>
542- withSQLConf(
543- CometConf .COMET_ENABLED .key -> " true" ,
544- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
545- spark.sql(s " SELECT sum(c $middle) FROM parquetV1Table " ).noop()
546- }
547- }
548-
549483 benchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
550484 withSQLConf(
551485 CometConf .COMET_ENABLED .key -> " true" ,
@@ -589,14 +523,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
589523 spark.sql(" SELECT * FROM parquetV1Table WHERE c1 + 1 > 0" ).noop()
590524 }
591525
592- benchmark.addCase(" SQL Parquet - Comet" ) { _ =>
593- withSQLConf(
594- CometConf .COMET_ENABLED .key -> " true" ,
595- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
596- spark.sql(" SELECT * FROM parquetV1Table WHERE c1 + 1 > 0" ).noop()
597- }
598- }
599-
600526 benchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
601527 withSQLConf(
602528 CometConf .COMET_ENABLED .key -> " true" ,
@@ -640,14 +566,6 @@ class CometReadBaseBenchmark extends CometBenchmarkBase {
640566 spark.sql(" SELECT * FROM parquetV1Table WHERE c1 + 1 > 0" ).noop()
641567 }
642568
643- benchmark.addCase(" SQL Parquet - Comet" ) { _ =>
644- withSQLConf(
645- CometConf .COMET_ENABLED .key -> " true" ,
646- CometConf .COMET_NATIVE_SCAN_IMPL .key -> SCAN_NATIVE_COMET ) {
647- spark.sql(" SELECT * FROM parquetV1Table WHERE c1 + 1 > 0" ).noop()
648- }
649- }
650-
651569 benchmark.addCase(" SQL Parquet - Comet Native DataFusion" ) { _ =>
652570 withSQLConf(
653571 CometConf .COMET_ENABLED .key -> " true" ,
0 commit comments