hadoop集群规划
HDFS中有NN和DN
YARN中有RM和NM
假设有三台机器:
hadoop000 192.168.199.234 部署:NN DN RM NM
hadoop001 192.168.199.235 部署:DN NM
hadoop002 192.168.199.236 部署:DN NM
- 首先在每台机器中修改
/etc/hostname
对应的hadoop000 hadoop001 hadoop002 - 修改hostname与ip的对应关系。修改
/etc/hosts
192.168.199.234 hadoop000
192.168.199.235 hadoop001
192.168.199.236 hadoop002
192.168.199.236 localhost
- 每台机器安装ssh,并且设置免密码登录
在hadoop00中进行操作:
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop000
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop001
ssh-copy-id -i ~/.ssh/id_rsa.pub hadoop002
- 每个节点上安装jdk
- 安装hadoop
配置hadoop-env.sh中的java_home。
配置core-site.xml,内容如下:
<property>
<name>fs.default.name</name>
<value>hdfs://hadoop000:9000</value>
</property>
配置hdfs-site.xml,内容如下:
<property>
<name>hadoop.tmp.dir</name>
<value>/home/hadoop/app/tmp/hadoop-tmp</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/home/hadoop/app/tmp/dfs/name</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/home/hadoop/app/tmp/dfs/data</value>
</property>
配置yarn-site.xml,内容如下:
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.resourcemanager.hostname</name>
<value>hadoop000</value>
</property>
配置mapred-site.xml,内容如下:
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
- 在slaves文件中,修改内容如下:
hadoop000
hadoop001
hadoop002
- 分发hadoop目录到其他机器
- NN格式化:
hadoop namenode -format
- 启动HDFS
- 启动YARN