问题1:
hadoop2.7+spark1.2: class not found com.hadoop.compression.lzo.LzoCodec
测试环境,上传jar包,spark-submit运行报上述错误,主要原因是:
hadoop集群的core-site.xml有:
<property><name>io.compression.codec.lzo.class</name>
<value>com.hadoop.compression.lzo.LzoCodec</value></property>
spark on yarn会默认使用集群的编码方式,但是相关的jar包不会自动的添加,需要自己配置,
在 $spark_home/conf/spark-env.sh ,添加 hadoop-lzo.jar(具体路径依据具体情况改变):
问题2:
WARN cluster.YarnSchedulerBackend$YarnSchedulerEndpoint: Executor for container container_1588340787997_0007_01_000002 exited because of a YARN event (e.g., pre-emption) and not because of an error in the running job.<