美文网首页
Spark提交参数

Spark提交参数

作者: Jorvi | 来源:发表于2019-05-14 19:43 被阅读0次
参数名 参数说明 备注
--master MASTER_URL spark://host:port, mesos://host:port, yarn, or local
--deploy-mode DEPLOY_MODE client/cluster
--class Your application's main class
--name A name of your application
--jars Comma-separated list of local jars to include on the driver and executor classpaths 以逗号分隔
--packages Comma-separated list of maven coordinates of jars to include on the driver and executor classpaths The format for the coordinates should be groupId:artifactId:version.
--exclude-packages Comma-separated list of groupId:artifactId, to exclude while resolving the dependencies provided in --packages to avoid dependency conflicts.
--repositories Comma-separated list of additional remote repositories to search for the maven coordinates given with --packages.
--py-files Comma-separated list of .zip, .egg, or .py files to place on the PYTHONPATH for Python apps.
--files Comma-separated list of files to be placed in the working directory of each executor.
--conf Arbitrary Spark configuration property.
--properties-file Path to a file from which to load extra properties. If not specified, this will look for conf/spark-defaults.conf.
--driver-memory Memory for driver (e.g. 1000M, 2G)
--driver-java-options Extra Java options to pass to the driver.
--driver-library-path Extra library path entries to pass to the driver.
--driver-class-path Extra class path entries to pass to the driver. Note that jars added with --jars are automatically included in the classpath.
--executor-memory Memory per executor (e.g. 1000M, 2G) Default: 1G
--proxy-user User to impersonate when submitting the application. This argument does not work with --principal / --keytab.
--help Show this help message and exit.
--verbose Print additional debug output.
--version Print the version of current Spark.
--driver-cores Cores for driver (Default: 1). Spark standalone with cluster deploy mode only
--supervise If given, restarts the driver on failure. Spark standalone or Mesos with cluster deploy mode only
--kill If given, kills the driver specified. Spark standalone or Mesos with cluster deploy mode only
--status If given, requests the status of the driver specified. Spark standalone or Mesos with cluster deploy mode only
--total-executor-cores Total cores for all executors. Spark standalone and Mesos only
--executor-cores Number of cores per executor. (Default: 1 in YARN mode, or all available cores on the worker in standalone mode) Spark standalone and YARN only
--driver-cores Number of cores used by the driver, only in cluster mode(Default: 1). YARN-only
--queue The YARN queue to submit to (Default: "default"). YARN-only
--num-executors Number of executors to launch (Default: 2). YARN-only
--archives Comma separated list of archives to be extracted into the working directory of each executor. YARN-only
--principal Principal to be used to login to KDC, while running on secure HDFS. YARN-only
--keytab The full path to the file that contains the keytab for the principal specified above. YARN-only

相关文章

网友评论

      本文标题:Spark提交参数

      本文链接:https://www.haomeiwen.com/subject/gdteaqtx.html