1. 集群规划
3. 设置ip地址 (如果有图形界面,可以直接通过图形界面进行设置,三台机器)
主机 | cpu核数 | 内存 | ip | 每个节点安装的服务 |
master | 1 | 3g | 192.168.108.10 | namenode, secondnamenode,datanode,resourcemanager,nodemanager,JobHistoryServer |
slave1 | 1 | 2g | 192.168.108.11 | datanode, nodemanager |
slave2 | 1 | 2g | 192.168.108.12 | datanode, nodemanager |
(1)centos7使用如下命令修改主机名(centos6需要配置/etc/sysconfig/network)
hostnamectl set-hostname master
- 编辑网卡配置文件,需要设置正确的网关地址,本例按照NAT模式进行设置,需要设置NAT的网关地址,否则上不了外网
vi /etc/sysconfig/network-script/ens-33
TYPE=Ethernet
PROXY_METHOD=none
BROWSER_ONLY=no
BOOTPROTO=static
DEFROUTE=yes
IPV4_FAILURE_FATAL=no
IPV6INIT=yes
IPV6_AUTOCONF=yes
IPV6_DEFROUTE=yes
IPV6_FAILURE_FATAL=no
IPV6_ADDR_GEN_MODE=stable-privacy
NAME=ens33
UUID=ccc1cf88-6881-4eec-9ec1-9c8539dec83d
DEVICE=ens33
ONBOOT=yes
GATEWAY=192.168.108.2
DNS1=192.168.108.2
DNS2=8.8.8.8
NETMASK=255.255.255.0
(2)设置完成后需要重启网络服务
service network restart
vi /etc/hosts
192.168.108.10 master
192.168.108.11 slave1
192.168.108.12 slave2
(1)关闭防火墙(centos6不同)
systemctl stop firewalld
systemctl disable firewalld
systemctl status firewalld
(2)关闭linux的子安全系统
vi /etc/sysconfig/selinux
SELINUX=disabled
reboot
- ssh-keygen -t rsa
- ssh-copy-id master
- scp authorized_keys slave1:/root/.ssh
其余机器都做同样的操作
https://hadoop.apache.org/releases.html
tar -zxvf hadoop-2.7.7.tar.gz -C /home/wangping/app
vi /etc/profile
export HADOOP_HOME=/home/wangping/app/hadoop-2.7.7
export PATH=$HADOOP_HOME:PATH
cd /home/wangping/app/hadoop-2.7.7/etc/hadoop
vi /hadoop-env.sh
export JAVA_HOME=/home/wangping/app/jdk1.8.0_144
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://master:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>/home/wangping/app/hadoop-2.7.7/tmp</value>
</property>
</configuration>
<configuration>
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>master:50090</value>
</property>
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>/home/wangping/app/hadoop-2.7.7/tmp/dfs/name</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>/home/wangping/app/hadoop-2.7.7/tmp/dfs/data</value>
</property>
</configuration>
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<property>
<name>mapreduce.jobhistory.address</name>
<value>master:10020</value>
</property>
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>master:19888</value>
</property>
</configuration>
<configuration>
<!-- Site specific YARN configuration properties -->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>master</value>
</property>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
</configuration>
scp -r hadoop-2.7.7/ slave1:/home/wangping/app
scp -r hadoop-2.7.7/ slave2:/home/wangping/app
vi slaves
master
slave1
slave2
cd /home/wangping/app/hadoop-2.7.7/bin
./hadoop namenode -format
./hadoop datanode -format
cd sbin
./start-all.sh
./start-all.sh ./mr-jobhistory-daemon.sh start historyserver
jps
13153 NodeManager
13554 JobHistoryServer
12584 NameNode
12872 SecondaryNameNode
13049 ResourceManager
12700 DataNode
cd sbin
./stop all
黄色为需要编辑的文本内容
灰色为shell命令