Page 1 of 1

The default value of spark.sql.shuffle.partitions is 200. Which of the following describes what that means?

Posted: Mon Mar 27, 2023 1:39 pm
by answerhappygod
The default value of spark.sql.shuffle.partitions is 200. Which of the following describes what that means?

A. By default, all DataFrames in Spark will be spit to perfectly fill the memory of 200 executors.
B. By default, new DataFrames created by Spark will be split to perfectly fill the memory of 200 executors.
C. By default, Spark will only read the first 200 partitions of DataFrames to improve speed.
D. By default, all DataFrames in Spark, including existing DataFrames, will be split into 200 unique segments for parallelization.
E. By default, DataFrames will be split into 200 unique partitions when data is being shuffled.