spark restful api 启动任务,运行完之后到historyserver

首先

1、 spark-env.sh中添加

export SPARK_HISTORY_OPTS="-Dspark.history.ui.port=18080 -Dspark.history.retainedApplications=3 -Dspark.history.

fs.logDirectory=hdfs://xiaoqi0:9000/sparkeventlog"

2、spark-defaults.conf

spark.eventLog.enabled=true
spark.eventLog.dir=hdfs://xiaoqi0:9000/sparkeventlog
spark.eventLog.compress=true

3、

curl -X POST http://xiaoqi0:6066/v1/submissions/create --header "Content-Type:application/json;charset=UTF-8" --data '{

  "action" : "CreateSubmissionRequest",
  "appArgs" : [ "100" ], 
  "appResource" : "file:/usr/local/app/spark-2.2.0-bin-hadoop2.7/examples/jars/spark-examples_2.11-2.2.0.jar", 
  "clientSparkVersion" : "2.2.0",
  "environmentVariables" : {
    "SPARK_ENV_LOADED" : "1"
  },
  "mainClass" : "org.apache.spark.examples.SparkPi",
  "sparkProperties" : {
    "spark.jars" : "file:/usr/local/app/spark-2.2.0-bin-hadoop2.7/examples/jars/spark-examples_2.11-2.2.0.jar",
    "spark.driver.supervise" : "false",
    "spark.app.name" : "MyJob",
    "spark.eventLog.enabled": "true",
    "spark.eventLog.dir":"hdfs://xiaoqi0:9000/sparkeventlog",
    "spark.submit.deployMode" : "client",
    "spark.master" : "spark://xiaoqi0:6066"
  }

}'

获取状态

curl http://spark-cluster-ip:6066/v1/submissions/status/driver-20151008145126-0000

杀死任务

curl -X POST http://spark-cluster-ip:6066/v1/submissions/kill/driver-20151008145126-0000

经测试,虽然配置文件中已经指定了目录sparkeventlog,但是如果curl请求提交任务时如果不配置spark.eventLog.dir,日志依然会输出到本地默认的/tmp/spark-events目录。配置了就会生成到指定的目录,这样sparksubmit提交的和restful提交的都制定到同一个目录,historyserver会都识别出来



评论 2
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值