下载 spark1.6.0解压 tar -xvf spark-1.6.0-bin-hadoop2.6.tgz
复制到/usr/local sudo cp -r spark-1.6.0-bin-hadoop2.6 /usr/local/spark
修改目录归属为hadoop sudo chown -R hadoop:users /usr/local/spark/
配置spark-env.sh
export JAVA_HOME= /usr/java/
export HADOOP_HOME= /usr/local/hadoop
export HADOOP_CONF= /usr/local/hadoop/etc/hadoop
export SPARK_MASTER_IP= master
export SPARK_WORKER_MEMORY= 4g
export SPARK_EXECUTOR_MEMORY= 2g
export SPARK_DRIVER_MEMORY= 2g
export SPARK_WORKER_CORES= 4
配置slaves
slave1
slave2
配置spark-defaults.conf
spark.eventLog.enabled true
spark.eventLog.dir hdfs://master:9000/historyserverforSpark
spark.yarn.historyServer.address master:18080
spark.history.fs.logDirectory hdfs://master:9000/historyserverforSpark
spark.executor.extraJavaOptions -XX:+PrintGCDetails -Dkey= value -Dnumbers= "one two three"
配置~/.bashrc中环境变量
SPARK_HOME= /usr/local/spark
PATH= $JAVA_HOME /bin:$PATH : $HADOOP_HOME /bin:$HADOOP_HOME /sbin:$SPARK_HOME /bin:$SPARK_HOME /sbin
export JAVA_HOME CLASSPATH SPARK_HOME
将spark目录以及.bashrc同步到从节点
hadoop@master:~> scp -r /usr/local/spark/ root@slave1:/usr/local
hadoop@master:~> scp -r /usr/local/spark/ root@slave2:/usr/local
hadoop@master:~> scp ~/.bashrc hadoop@slave1:/home/hadoop
hadoop@master:~> scp ~/.bashrc hadoop@slave2:/home/hadoop
修改从节点hadoop目录权限
hadoop@slave1:~> sudo chown -R hadoop:users /usr/local/spark/
hadoop@slave2:~> sudo chown -R hadoop:users /usr/local/spark/
启动spark,执行sh start-all.sh
启动日志历史服务,首次启动前须创建hdfs路径
hadoop@master:/usr/local/spark/sbin> hadoop fs -mkdir hdfs://master:9000/historyserverforSpark
20/07/11 11:34:31 WARN util.NativeCodeLoader: Unable to load native-hadoop library for your platform.. . using builtin-java classes where applicable
hadoop@master:/usr/local/spark/sbin> sh start-history-server.sh
starting org.apache.spark.deploy.history.HistoryServer, logging to /usr/local/spark/logs/spark-hadoop-org.apache.spark.deploy.history.HistoryServer-1-master.out
hadoop@master:/usr/local/spark/sbin>
测试样例程序
hadoop@master:/usr/local/spark/bin> ./spark-submit \
> --class org.apache.spark.examples.SparkPi \
> --master spark://master:7077 \
> .. /lib/spark-examples-1.6.0-hadoop2.6.0.jar
Using Spark's default log4j profile: org/apache/spark/log4j-defaults.properties
.. .
停止命令
/usr/local/spark/sbin/stop-history-server.sh
/usr/local/spark/sbin/stop-all.sh