1

we want to find the values for the following spark parameters according to inputs as memory on datanode machine , CPU CORE on data node machine , numbers of datanode machine etc ,,

spark.dynamicAllocation.initialExecutors = ?
spark.dynamicAllocation.minExecutors = ?
spark.dynamicAllocation.maxExecutors = ?

I guess we cant give exactly the values so it will help to find the range for this parameters

for example

we have ambari cluster with

3 master machines

21 worker machine ( each worker machine have 32G memory & 16 CPU CORE )

3 kafka machines

please help me to calculate the values for the parameters ( I will happy to get then range of the value for each parameters since I know we never shut the value exactly )

I just give here my opinion ( based on the site - https://community.microstrategy.com/s/article/Best-Practices-for-Spark-Thrift-Server-on-YARN

can we set both parameters to 21 ( as number of workers machines ) ?? , is it OK?

spark.dynamicAllocation.initialExecutors = 21
spark.dynamicAllocation.minExecutors = 21
spark.dynamicAllocation.maxExecutors = 21
shalom
  • 451
  • 12
  • 26

0 Answers0