Compile datagen tool (derived from maropu/spark-tpcds-datagen).
cd tpcds/datagen
mvn package -DskipTestsGenerate 1TB dataset with spark.
# use correct SPARK_HOME and output data location
# --use-double-for-decimal and --use-string-for-char are optional, see dsdgen usage
SPARK_HOME=$HOME/software/spark ./bin/dsdgen \
--output-location /user/hive/data/tpcds-1000 \
--scale-factor 1000 \
--format parquet \
--overwrite \
--use-double-for-decimal \
--use-string-for-charCompile benchmark tool (derived from databricks/spark-sql-perf).
cd tpcds/benchmark-runner
mvn package -DskipTestsEdit your $SPARK_HOME/conf/spark-default.conf to enable/disable Auron (see the following conf), then launch benchmark runner.
If benchmarking with Auron, ensure that the Auron jar package is correctly built and moved into $SPARK_HOME/jars. (How to build Auron?)
# use correct SPARK_HOME and data location
SPARK_HOME=$HOME/software/spark ./bin/run \
--data-location /user/hive/data/tpcds-1000 \
--format parquet \
--output-dir ./benchmark-resultMonitor benchmark status:
tail -f ./benchmark-result/YYYYMMDDHHmm/logSummarize query times of all cases:
./bin/stat ./benchmark-result/YYYYMMDDHHmmspark.master yarn
spark.yarn.stagingDir.list hdfs://auron-test/home/spark/user/
spark.eventLog.enabled true
spark.eventLog.dir hdfs://auron-test/home/spark-eventlog
spark.history.fs.logDirectory hdfs://auron-test/home/spark-eventlog
spark.externalBlockStore.url.list hdfs://auron-test/home/platform
spark.driver.extraJavaOptions -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=/media/disk1/spark/ -Djava.io.tmpdir=/media/disk1/tmp -Dlog4j2.formatMsgNoLookups=true
spark.local.dir /media/disk1/spark/localdir
spark.shuffle.service.enabled true
spark.shuffle.service.port 7337
spark.driver.memory 20g
spark.driver.memoryOverhead 4096
spark.executor.instances 10000
spark.dynamicallocation.maxExecutors 10000
spark.executor.cores 5
spark.io.compression.codec zstd
spark.sql.parquet.compression.codec zstd
# benchmark without auron
#spark.executor.memory 6g
#spark.executor.memoryOverhead 2048
# benchmark with auron
spark.executor.memory 4g
spark.executor.memoryOverhead 4096
spark.auron.enable true
spark.sql.extensions org.apache.spark.sql.auron.AuronSparkSessionExtension
spark.shuffle.manager org.apache.spark.sql.execution.auron.shuffle.AuronShuffleManager
spark.memory.offHeap.enabled false
# spark3.3+ disable char/varchar padding
#spark.sql.readSideCharPadding false
# enable shuffled hash join
#spark.sql.join.preferSortMergeJoin false