1. --class:你的应用的启动类 (如 org.apache.spark.examples.SparkPi)
    2. --queue
    3. --master:指定Master的地址,默认为Local
    4. --deploy-mode:是否发布你的驱动到worker节点(cluster) 或者作为一个本地客户端 (client) (default: client)
    5. --num-executors
    6. --executor-cores
    7. --executor-memory
    8. --driver-memory
    9. #任意的Spark配置属性, 格式key=value
    10. --conf "spark.yarn.appMasterEnv.JAVA_HOME=/home/iteblog/java/jdk1.8.0_25" \
    11. --conf "spark.executorEnv.JAVA_HOME=/home/iteblog/java/jdk1.8.0_25" \
    12. #spark-shell提交Spark Application如何解决依赖库
    13. --driver-class-path 多个jar包之间用冒号分割(windows是分号 linux是冒号表示结束)
    14. #依赖的jar
    15. --jars
    16. file:/home/hadoop/SparkSqlTest-1.0-SNAPSHOT.jar
    17. 参数一 参数二 。。。。。。
    1. spark-submit \
    2. --master yarn \
    3. --deploy-mode cluster \
    4. --queue default \
    5. --num-executors 30 \
    6. --executor-memory 1G \
    7. --executor-cores 1 \
    8. --driver-memory 1G \
    9. --py-files spark_dist/data_processing.zip \
    10. --jars hdfs://apm/jar-file/ip_udf-0.0.2-SNAPSHOT-jar-with-dependencies.jar \
    11. driver.py prod data_processing.apm_overseas.main
    1. #!/bin/bash
    2. #定时提交
    3. APP_HOME=/home/spark/sparkjob
    4. JAR_HOME=/home/spark/pack
    5. yesterday=`date +"%Y-%m-%d" -d "-1 days"`
    6. for i in $JAR_HOME/*.jar;
    7. do
    8. app_CLASSPATH=$i,${app_CLASSPATH}
    9. done
    10. len=${#app_CLASSPATH}-1
    11. JAR_PATH=${app_CLASSPATH:0:len}
    12. /usr/hdp/2.6.1.0-129/spark2/bin/spark-submit \
    13. --class com.kd.tonze.userPortrait.UsersLabelJob \
    14. --master yarn \
    15. --deploy-mode client \
    16. --num-executors 6 \
    17. --jars $JAR_PATH \
    18. $APP_HOME/kgraphx_2.11-0.0.1.jar $yesterday