一、windows下spark环境搭建
1.安装JDK和scala
2.安装spark和hadoop
http://spark.apache.org/downloads.html
https://archive.apache.org/dist/hadoop/common/
慢的话可以去国内镜像eg:
http://mirror.bit.edu.cn/apache/hadoop/common/hadoop-2.7.7/
注意spark、hadoop和scala版本问题,我用的是scala2.13.1(换成2.12.11了,不然pom文件中dependency会红),spark2.4.5,hadoop2.7.7
3.下winutils
https://github.com/steveloughran/winutils
4.在C盘建目录tmp/hive然后在CMD输入C:\你的安装目录\hadoop\bin\winutils.exe chmod 777 C:/tmp/hive
测试spark-shell