- 创建目录
- 安装目录下创建数据存放的文件夹,/home/hadoop/hadoop-2.7.1/tmp、hdfs、hdfs/data、hdfs/name
- core-site.xml
<property> <name>fs.defaultFS</name> <value>hdfs://centos01:9000</value> </property> <property> <name>hadoop.tmp.dir</name> <value>file:/home/hadoop/hadoop-2.7.1/tmp</value> </property> <property> <name>io.file.buffer.size</name> <value>131702</value> </property>
- hdfs-site.xml
<property> <name>dfs.namenode.name.dir</name> <value>file:/home/hadoop/hadoop-2.7.1/dfs/name</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/home/hadoop/hadoop-2.7.1/dfs/data</value> </property> <property> <name>dfs.replication</name> <value>2</value> </property> <property> <name>dfs.namenode.secondary.http-address</name> <value>centos01:9001</value> </property> <property> <name>dfs.webhdfs.enabled</name> <value>true</value> </property>
- mapred-site.xml
<property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> <property> <name>mapreduce.jobhistory.address</name> <value>centos01:10020</value> </property> <property> <name>mapreduce.jobhistory.webapp.address</name> <value>centos01:19888</value> </property>
- yarn-site.xml
<property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.auxservices.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> <property> <name>yarn.resourcemanager.address</name> <value>centos01:8032</value> </property> <property> <name>yarn.resourcemanager.scheduler.address</name> <value>centos01:8030</value> </property> <property> <name>yarn.resourcemanager.resource-tracker.address</name> <value>centos01:8031</value> </property> <property> <name>yarn.resourcemanager.admin.address</name> <value>centos01:8033</value> </property> <property> <name>yarn.resourcemanager.webapp.address</name> <value>centos01:8088</value> </property> <property> <name>yarn.nodemanager.resource.memory-mb</name> <value>768</value> </property>
- hadoop-env.sh、yarn-env.sh 加入JAVA_HOME环境变量
export JAVA_HOME=/usr/java/jdk1.7.0_80
- slaves
#localhost centos02
scp -r /home/hadoop/hadoop-2.7.1 root@centos02:/home/hadoop/
初始化hadoop
bin/hdfs namenode -format
启动hadoop
sbin/start-all.sh
访问web页面查看结果
http://centos01:8088/
成功页面
常见问题
- 如果是虚机安装在windows下,scp找不到主机,请在windows/system32/dirver/etc/hosts下把相关主机名添加上
192.168.222.130 centos01 192.168.222.131 centos02
- 如果web页面访问不到,请关闭防火墙和iptable
systemctl stop firewalld.service systemctl stop iptables.service
- 禁用防火墙自动启动
systemctl disable firewalld.service