JDK1.8 + Spark2.3 + Hadoop2.7.1
- 安装JDK1.8
- 安装Spark2.3.1(http://spark.apache.org/downloads.html)
解压路径:D:\Tools\spark-2.3.1-bin-hadoop2.7
配置环境变量:1)系统变量:SPARK_HOME;变量值:D:\Tools\spark-2.3.1-bin-hadoop2.7
2)Path环境变量增加:%SPARK_HOME%\bin,%SPARK_HOME%\sbin - 安装Hadoop2.7.1(https://archive.apache.org/dist/hadoop/common/hadoop-2.7.1)
解压路径:D:\Tools\hadoop-2.7.1
配置环境变量:1)系统变量:HADOOP_HOME;变量值:D:\Tools\hadoop-2.7.1
2)Path环境变量增加:%HADOOP_HOME%\bin - 下载 winutils.exe 并复制到 hadoop的bin目录 https://github.com/steveloughran/winutils/blob/master/hadoop-2.7.1/bin/winutils.exe
- 启动Spark
- CMD切换到spark-2.3.1-bin-hadoop2.7\bin目录
- 运行spark-shell