公司三台测试服务器,测试spark的集群模式是否正常运行遇到的问题:
1.spark运行spark任务遇到的,
SparkContext did not initialize after waiting for 100000 ms. Please check earlier log output for errors. Failing the application.
参考博客:
https://www.cnblogs.com/huanongying/archive/2017/10/12/7655598.html
运行脚本有问题:
sudo -u hdfs /usr/hdp/2.6.5.0-292/spark2/bin/spark-submit \ spark的bin目录下执行
–master yarn \ spark的集群模式
–deploy-mode cluster \ yarn 的模式
–class com.amhy.test.Sprk01 \ 类的全路径
–num-executors 3 \ executor的数量
–driver-memory 512m \ driver的内存
–executor-memory 1g \ executor的内存
–executor-cores 1
/bigdata/jars/scala-yarn.jar \
将submit.py文件修改成可执行文件:
chmod +x 文件名
执行方法: .