概述:本文主要讲述如何将本地Spark工程打包放到YARN中运行。
1、基本案例
(1)spark-shell位置
cd /root/app/spark-2.4.2-bin-hadoop2.6/bin
(2)设置HADOOP_CONF_DIR位置
export HADOOP_CONF_DIR=/root/app/hadoop-2.6.0-cdh5.7.0/etc/hadoop
#或写入spark-env.sh
vi /root/app/spark-2.4.2-bin-hadoop2.6/conf/spark-env.sh
#内容为
HADOOP_CONF_DIR=/root/app/hadoop-2.6.0-cdh5.7.0/etc/hadoop
(3)启动yarn
./app/hadoop-2.6.0-cdh5.7.0/sbin/start-all.sh
(4)执行spark自带实例
#命令参考http://spark.apache.org/docs/latest/submitting-applications.html
./app/spark-2.4.2-bin-hadoop2.6/bin/spark-submit \
--class org.apache.spark.examples.SparkPi \
--master yarn \
--executor-memory 1G \
--total-executor-cores 1 \
/root/app/spark-2