Hadoop的伪分布式搭建及配置
备注:重新搭建hadoop环境前需要删除以下文件
cd /opt/apache_hadoop/hadoop-2.7.3
rm -rf ./data/tmp
rm -rf ./logs
# hadoop-env.sh
export JAVA_HOME=/opt/jdk1.8.0_221
# mapred-env.sh
export JAVA_HOME=/opt/jdk1.8.0_221
# yarn-env.sh
export JAVA_HOME=/opt/jdk1.8.0_221
# core-site.xml
<configuration>
<!-- 指定namenode所在的机器节点 -->
<property>
<name>fs.defaultFS</name>
<value>hdfs://xpit:8020</value>
</property>
<!-- 指定工作目录 -->
<property>
<name>hadoop.tmp.dir</name>
<value>/opt/apache_hadoop/hadoop-2.7.3/data/tmp</value>
</property>
</configuration>
# hdfs-site.xml
<configuration>
<!-- 指定HDFS文件存储的副本数。默认3。单节点伪分布式环境配置1即可 -->
<property>
<name>dfs.replication</name>
<value>1</value>
</property>
</configuration>
# mapred-site.xml
<configuration>
<!-- 指定mapreduce程序运行在yarn