## download environment:
# 1. java: https://www.oracle.com/java/technologies/downloads/#jdk8
# 2. winutils-hadoop: https://github.com/cdarlint/winutils
# 3. spark: https://spark.apache.org/downloads.html
## uncompress and install:
# 1、将java安装到`C:\Install\Java`,设置环境变量`JAVA_HOME=C:\Install\Java\jdk-x-xxx`,且添加`%JAVA_HOME%\bin;`到path环境变量
# 2、将spark安装包解压到`C:\Install\SparkInstall\Spark`,
# 设置环境变量`SPARK_HOME=C:\Install\SparkInstall\Spark\spark-x.x.x-bin-hadoopx-scala2.xx`,且添加`%SPARK_HOME%\bin;`到path环境变量
# 3、创建hadoop路径`C:\Install\SparkInstall\Hadoop\hadoopx.x\bin`,并把下载的winutils.exe放到此处
# 4、设置环境变量`HADOOP_HOME=C:\Install\SparkInstall\Hadoop\hadoopx.x`,且添加`%HADOOP_HOME%\bin;`到path环境变量
# 5、创建hive路径`C:\Install\SparkInstall\Hive`,在cmd窗口执行`winutils.exe chmod -R 777 C:\Install\SparkInstall\Hive`
# 6、复制`C:\Install\SparkInstall\Spark\spark-x.x.x-bin-hadoopx-scala2.xx\python\pyspark`到`C:\Install\anaconda3\Lib\site-packages`。
# 7、设置环境变量`PYSPARK_PYTHON=C:\Install\anaconda3\python.exe`、`PYSPARK_DRIVER_PYTHON=C:\Install\anaconda3\python.exe`
# 8、测试:在cmd中执行`pyspark`或`spark-shell`等
# 大功告成!
spark在windows安装
于 2017-12-28 16:41:03 首次发布