spark.master.url | By default, left unset. This property is only specified when using an external Spark cluster; when Fusion is using its own standalone Spark cluster, this property is not set. |
spark.cores.max | The maximum number of cores across the cluster assigned to the application. If not specified, there is no limit. The default is unset, i.e., an unlimited number of cores. |
spark.executor.memory | Amount of memory assigned to each application’s executor. The default is 2G. |
spark.scheduler.mode | Controls how tasks are assigned to available resources. Can be either ‘FIFO’ or ‘FAIR’. Default value is ‘FAIR’. |
spark.dynamicAllocation.enabled | Boolean - whether or not to enable dynamic allocation of executors. Default value is ‘TRUE’. |
spark.shuffle.service.enabled | Boolean - whether or not to enable internal shuffle service for standalone Spark cluster. Default value is ‘TRUE’. |
spark.dynamicAllocation.executorIdleTimeout | Number of seconds after which idle executors are removed. Default value is ‘60s’. |
spark.dynamicAllocation.minExecutors | Number of executors to leave running even when idle. Default value is 0. |
spark.eventLog.enabled | Boolean - whether or not event log is enabled. Event log stores job details and can be accessed after application finishes. Default value is ‘TRUE’. |
spark.eventLog.dir | Directory that stores event logs. Default location is $FUSION_HOME/var/spark-eventlog. |
spark.eventLog.compress | Boolean - whether or not to compress event log data. Default value is ‘TRUE’. |
spark.logConf | Boolean - whether or not to log effective SparkConf of new SparkContext-s. Default value is ‘TRUE’. |
spark.deploy.recoveryMode | Default value is ‘ZOOKEEPER’ |
spark.deploy.zookeeper.url | ZooKeeper connect string. Default value is $FUSION_ZK |
spark.deploy.zookeeper.dir | ZooKeeper path, default value is /lucid/spark |
spark.worker.cleanup.enabled | Boolean - whether or not to periodically cleanup worker data. Default value is ‘TRUE’. |
spark.worker.cleanup.appDataTtl | Time-to-live in seconds. Default value is 86400 (24h). |
spark.deploy.retainedApplications | The maximum number of applications to show in the UI. Default value is 50. |
spark.deploy.retainedDrivers | The maximum number of drivers. Default value is 50. |
spark.worker.timeout | The maximum timeout in seconds allowed before a worker is considered lost. The default value is 30. |
spark.worker.memory | The maximum total heap allocated to all executors running on this worker. Defaults to value of the executor memory heap. |