1. 入門指南
http://blog.jobbole.com/86232/
2. ipython 使用spark 環境設置
http://ramhiser.com/2015/02/01/configuring-ipython-notebook-support-for-pyspark/
(錯誤處理)
Q1 沒有讀到設置檔
http://stackoverflow.com/questions/31974797/ipython-notebook-wont-read-the-configuration-file
Q2 py4j.java error
D:\spark-1.6.2-bin-hadoop2.6\python\lib\py4j-0.9-src.zip
http://stackoverflow.com/questions/26533169/why-cant-pyspark-find-py4j-java-gateway
Q3 import error no module name pyspark
http://stackoverflow.com/questions/23256536/importing-pyspark-in-python-shell
Q4 進入方式
配置环境变量Path,添加E:\spark-1.3.0-bin-hadoop2.4\bin。这时,你可以利用打开cmd,输入pySpark
http://blog.csdn.net/a819825294/article/details/51782773
3. LINUX 環境 需添加
http://www.cnblogs.com/onetwo/p/5424377.html
MASTER_IP = "127.0.0.1"
LOCAL_IP = "127.0.0.1"