hadoop安装及配置
一、解压
tar -zxvf hadoop-2.8.2.tar.gz
二、修改Hadoop配置文件
cd hadoop-2.8.2/etc/hadoop
hadoop-env.sh文件添加以下内容:
export JAVA_HOME=/usr/local/src/jdk1.8.0_152
yarn-env.sh文件添加以下内容:
export JAVA_HOME=/usr/local/src/jdk1.8.0_152
slaves文件添加以下内容:
slave1
slave2
core-site.xml文件:
<configuration>
<property>
<name>fs.defaultFS</name>
<value>hdfs://172.16.11.97:9000</value>
</property>
<property>
<name>hadoop.tmp.dir</name>
<value>file:/usr/local/src/hadoop-2.8.2/tmp/</value>
</property>
</configuration>
hdfs-site.xml文件:
<configuration>
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>master:9001</value>
</property>
<property>
<name>dfs.namenode.name.dir</name>
<value>file:/usr/local/src/hadoop-2.8.2/dfs/name</value>
</property>
<property>
<name>dfs.datanode.data.dir</name>
<value>file:/usr/local/src/hadoop-2.8.2/dfs/data</value>
</property>
<property>
<name>dfs.replication</name>
<value>3</value>
</property>
</configuration>
mapred-site.xml文件:
<configuration>
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
</configuration>
yarn-site.xml文件:
<configuration>
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<property>
<name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
<value>org.apache.hadoop.mapred.ShuffleHandler</value>
</property>
<property>
<name>yarn.resourcemanager.address</name>
<value>master:8032</value>
</property>
<property>
<name>yarn.resourcemanager.scheduler.address</name>
<value>master:8030</value>
</property>
<property>
<name>yarn.resourcemanager.resource-tracker.address</name>
<value>master:8035</value>
</property>
<property>
<name>yarn.resourcemanager.admin.address</name>
<value>master:8033</value>
</property>
<property>
<name>yarn.resourcemanager.webapp.address</name>
<value>master:8088</value>
</property>
</configuration>
创建临时目录和文件目录:
mkdir /usr/local/src/hadoop-2.8.2/tmp
mkdir -p /usr/local/src/hadoop-2.8.2/dfs/name
mkdir -p /usr/local/src/hadoop-2.8.2/dfs/data
4. 配置环境变量
vim ~/.bashrc
HADOOP_HOME=/usr/local/src/hadoop-2.8.2
export PATH=$PATH:$HADOOP_HOME/bin
刷新环境变量
source ~/.bashrc
5. 拷贝安装包
scp -r /usr/local/src/hadoop-2.8.2 root@slave1:/usr/local/src/hadoop-2.8.2
scp -r /usr/local/src/hadoop-2.8.2 root@slave2:/usr/local/src/hadoop-2.8.2
6. 启动集群
Master 节点上初始化并启动集群
初始化Namenode:
hadoop namenode -format
启动集群:
./sbin/start-all.sh
7. 集群状态
jps
8. 监控页面
NameNode:
http://master:50070/dfshealth.jsp
SecondaryNameNode:
http://master:50090/status.jsp
DataNode:
http://slave1:50075/
http://slave2:50075/
JobTracker:
http://master:50030/jobtracker.jsp
TaskTracker:
http://slave1:50060/tasktracker.jsp
http://slave2:50060/tasktracker.jsp
9. 关闭集群
./sbin/hadoop stop-all.sh