Spark related parameters¶
To import data, you must set parameters for Spark. This table lists some generally-used parameters. For more Spark-related parameters, see Apache Spark documentation. For more information, see the examples.
Parameters | Default | Data type | Required? | Description |
---|---|---|---|---|
spark.app.name |
Spark Writer | string |
No | Specifies the name of the Spark Driver Program. |
spark.driver.cores |
1 | int |
No | Specifies the number of cores to use for the driver process, only in cluster mode. |
spark.driver.maxResultSize |
1G | string |
No | Specifies the limit of the total size of serialized results of all partitions for each Spark action (e.g. collect) in bytes. Should be at least 1M, or 0 for unlimited. |
spark.cores.max |
None | int |
No | When the driver program runs on a standalone deployed cluster or a Mesos cluster in "coarse-grained" sharing mode, the maximum amount of CPU cores to request for the application from across the cluster (not from each machine). If not set, the default will be spark.deploy.defaultCores on the standalone cluster manager of Spark, or infinite (all available cores) on Mesos. |
Last update: April 8, 2021