hadoop环境的搭建

在搭建hadoop环境之前,我们要先配置好每台机器的java jdk环境,没有配置的小伙伴可以看我这篇博文进行配置哦:hadoop环境之java jdk的配置

hadoop环境的搭建

  1. 上传linux版本下的hadoop的安装包
rz命令可以从本地上传文件到虚拟机中
[root@node04 SoftWare]# rz

出现hadoop安装包,上传成功
[root@node04 SoftWare]# ls
hadoop-2.7.7.tar.gz  
  1. 解压
[root@node04 SoftWare]# tar -zxvf hadoop-2.7.7.tar.gz

出现hadoop-2.7.7,解压成功
[root@node04 SoftWare]# ls
hadoop-2.7.7  hadoop-2.7.7.tar.gz

3.修改配置文件

① 修改hadoop-env.sh文件

[root@node04 SoftWare]# cd hadoop-2.7.7/etc/hadoop/
[root@node04 hadoop]# vi hadoop-env.sh
修改内容
export JAVA_HOME=/opt/SoftWare/jdk1.8.0_212

② 修改core-site.xml文件

hadoop的核心文件
[root@node04 hadoop]# vi core-site.xml<configuration></configuration>之间添加以下内容:

<!--定义hadoop hdfs中namenode的URI和端口-->
<property>
<name>fs.defaultFS</name>
<value>hdfs://node04:9000</value>
</property>
<!--hadoop运行时临时的存储目录-->
<property>
<name>hadoop.tmp.dir</name>
<value>file://opt/SoftWare/hadoop-2.7.7/tmp</value>
</property>
<!--用作序列化文件处理时读写buffer的大小-->
<property>
<name>io.file.buffer.size</name>
<value>131702</value>
</property>

③ 修改hdfs-site.xml文件

hdfs的核心配置文件
[root@node04 hadoop]# vi hdfs-site.xml<configuration></configuration>之间添加以下内容:

<!--配置namenode节点上的元数据的存储目录-->
<property>
<name>dfs.namenode.name.dir</name>
<value>file:/opt/SoftWare/hadoop-2.7.7/dfs/name</value>
</property>
<!--配置DataNode节点上真正的数据存储目录-->
<property>
<name>dfs.datanode.data.dir</name>
<value>file:/opt/SoftWare/hadoop-2.7.7/dfs/data</value>
</property>
<!--指定DataNode存储block的副本数量,不大于DataNode的个数就行,默认为3-->
<proeprty>
<name>dfs.replication</name>
<value>2</value>
</property>
<!--指定SecondaryNamenode的工作目录-->
<property>
<name>dfs.namenode.checkpoint.dir</name>
<value>file:/opt/SoftWare/hadoop-2.7.7/dfs/namesecondary</value>
</property>
<!--指定SecondaryNamenode的http协议访问地址-->
<property>
<name>dfs.namenode.secondary.http-address</name>
<value>node05:9868</value>
</property>

④ 修改yarn-site.xml文件

yarn的核心配置文件
[root@node04 hadoop]# vi yarn-site.xml<configuration></configuration>之间添加以下内容:

<!--reduce获取数据的方式-->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>
<!--resourcemanager的主机名,配置后其他的address就不用配置了-->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>node04</value>
</property>
<!--nodemanager节点的内存大小,单位为MB-->
<property>
<name>yarn.nodemanager.resource.memory-mb</name>
<value>2048</value>
</property>

⑤ 修改mapred-site.xml文件
进入/hadoop-2.7.7/etc/hadoop目录没有mapred-site.xml文件,发现有mapred-site.xml.template文件,mapred-site.xml.template文件就是我们要找的mapred-site.xml文件

使用cp命令将mapred-site.xml.template文件复制一份,并且重命名为mapred-site.xml
[root@node04 hadoop]# cp mapred-site.xml.template mapred-site.xml
修改
[root@node04 hadoop]# vi mapred-site.xml<configuration></configuration>之间添加以下内容:

<!--使用yarn运行的MapReduce程序-->
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>
<!--mapreduce jobhistory Server界面地址-->
<property>
<name>mapreduce.jobhistory.webapp.address</name>
<value>node04:19888</value>
</property>

⑥ 修改slaves文件
slaves文件里面配置的是所有datanode节点的主机名,所以我slaves里面的内容是我的三台节点的名字node04、node05、node06

[root@node04 hadoop]# vi slaves
node04
node05
node06
  1. 文件分发
    上面的操作另外两台节点也要进行同样的操作,因此我们可以通过命令直接拷贝减少工作量
[root@node04 SoftWare]# scp -r /opt/SoftWare/ root@node05:/opt/SoftWare/
[root@node04 SoftWare]# scp -r /opt/SoftWare/ root@node06:/opt/SoftWare/
  1. 设置环境变量
    每台机器都需要操作
[root@node04 SoftWare]# vi /etc/profile
<!--添加以下内容-->
export HADOOP_HOME=/opt/SoftWare/hadoop-2.7.7
export HADOOP_LOG_DIR=$HADOOP_HOME/logs
export YARN_LOG_DIR=$HADOOP_LOG_DIR
export PATH=$PATH:$HADOOP_HOME/bin:$HADOOP_HOME/sbin
<!--使设置立即生效-->
[root@node04 SoftWare]# source /etc/profile
  1. 格式化hadoop
    该操作主要在首次启动前进行,格式化本质是进行文件系统的初始化操作,创建一些hadoop自己所需要的文件,格式化之后且集群启动成功后,后续就不需要再进行格式化,该操作主要在hdfs集群的主角色(namenode)所在机器上操作
[root@node04 SoftWare]# hdfs namenode -format

<!--出现这样一大串的信息,说明格式化成功-->

20/09/23 18:52:48 INFO namenode.NameNode: STARTUP_MSG: 
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = node04/192.168.110.101
STARTUP_MSG:   args = [-format]
STARTUP_MSG:   version = 2.7.7
STARTUP_MSG:   classpath = /opt/SoftWare/hadoop-2.7.7/etc/hadoop:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/activation-1.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/hadoop-annotations-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-configuration-1.6.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/xz-1.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/junit-4.11.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-httpclient-3.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/stax-api-1.0-2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/apacheds-i18n-2.0.0-M15.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/hadoop-auth-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/httpclient-4.2.5.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/mockito-all-1.8.5.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jackson-jaxrs-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-logging-1.1.3.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/curator-recipes-2.7.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jetty-sslengine-6.1.26.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jersey-json-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/avro-1.7.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/log4j-1.2.17.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-cli-1.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-digester-1.8.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/servlet-api-2.5.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/xmlenc-0.52.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jetty-util-6.1.26.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/guava-11.0.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-compress-1.4.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-io-2.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jackson-core-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jersey-core-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jsp-api-2.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-codec-1.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/netty-3.6.2.Final.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jetty-6.1.26.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jersey-server-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/paranamer-2.3.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/zookeeper-3.4.6.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-collections-3.2.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jettison-1.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/asm-3.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/api-asn1-api-1.0.0-M20.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/hamcrest-core-1.3.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jsch-0.1.54.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/api-util-1.0.0-M20.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/curator-framework-2.7.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-net-3.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/gson-2.2.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jets3t-0.9.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-lang-2.6.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jackson-mapper-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/commons-math3-3.1.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/httpcore-4.2.5.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/htrace-core-3.1.0-incubating.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/jsr305-3.0.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/slf4j-api-1.7.10.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/lib/curator-client-2.7.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/hadoop-common-2.7.7-tests.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/hadoop-nfs-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/common/hadoop-common-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/xercesImpl-2.9.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/xml-apis-1.3.04.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/guava-11.0.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/commons-io-2.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/jackson-core-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/asm-3.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/leveldbjni-all-1.8.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/jackson-mapper-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/netty-all-4.0.23.Final.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/htrace-core-3.1.0-incubating.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/lib/jsr305-3.0.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/hadoop-hdfs-2.7.7-tests.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/hadoop-hdfs-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/hdfs/hadoop-hdfs-nfs-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/activation-1.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/zookeeper-3.4.6-tests.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/aopalliance-1.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/xz-1.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jackson-jaxrs-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jersey-json-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/log4j-1.2.17.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/commons-cli-1.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/servlet-api-2.5.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jackson-xc-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/guava-11.0.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/commons-io-2.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jackson-core-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jersey-core-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/commons-codec-1.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/netty-3.6.2.Final.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jetty-6.1.26.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jersey-server-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/guice-3.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jersey-client-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/zookeeper-3.4.6.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/commons-collections-3.2.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jettison-1.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/asm-3.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/commons-lang-2.6.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jackson-mapper-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/javax.inject-1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/lib/jsr305-3.0.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-server-common-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-client-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-common-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-server-tests-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-api-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-registry-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-server-sharedcachemanager-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/hadoop-annotations-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/xz-1.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/junit-4.11.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/jackson-core-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/guice-3.0.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/asm-3.2.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/hamcrest-core-1.3.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/leveldbjni-all-1.8.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/lib/javax.inject-1.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.7.7-tests.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.7.7.jar:/opt/SoftWare/hadoop-2.7.7/contrib/capacity-scheduler/*.jar
STARTUP_MSG:   build = Unknown -r c1aad84bd27cd79c3d1a7dd58202a8c3ee1ed3ac; compiled by 'stevel' on 2018-07-18T22:47Z
STARTUP_MSG:   java = 1.8.0_212
************************************************************/
20/09/23 18:52:48 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
20/09/23 18:52:48 INFO namenode.NameNode: createNameNode [-format]
Formatting using clusterid: CID-6f04c362-f8ce-40bf-9449-81b1f09b3d9f
20/09/23 18:52:50 INFO namenode.FSNamesystem: No KeyProvider found.
20/09/23 18:52:50 INFO namenode.FSNamesystem: fsLock is fair: true
20/09/23 18:52:50 INFO namenode.FSNamesystem: Detailed lock hold time metrics enabled: false
20/09/23 18:52:50 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000
20/09/23 18:52:50 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
20/09/23 18:52:50 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
20/09/23 18:52:50 INFO blockmanagement.BlockManager: The block deletion will start around 2020 九月 23 18:52:50
20/09/23 18:52:50 INFO util.GSet: Computing capacity for map BlocksMap
20/09/23 18:52:50 INFO util.GSet: VM type       = 64-bit
20/09/23 18:52:50 INFO util.GSet: 2.0% max memory 966.7 MB = 19.3 MB
20/09/23 18:52:50 INFO util.GSet: capacity      = 2^21 = 2097152 entries
20/09/23 18:52:51 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false
20/09/23 18:52:51 INFO blockmanagement.BlockManager: defaultReplication         = 2
20/09/23 18:52:51 INFO blockmanagement.BlockManager: maxReplication             = 512
20/09/23 18:52:51 INFO blockmanagement.BlockManager: minReplication             = 1
20/09/23 18:52:51 INFO blockmanagement.BlockManager: maxReplicationStreams      = 2
20/09/23 18:52:51 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000
20/09/23 18:52:51 INFO blockmanagement.BlockManager: encryptDataTransfer        = false
20/09/23 18:52:51 INFO blockmanagement.BlockManager: maxNumBlocksToLog          = 1000
20/09/23 18:52:51 INFO namenode.FSNamesystem: fsOwner             = root (auth:SIMPLE)
20/09/23 18:52:51 INFO namenode.FSNamesystem: supergroup          = supergroup
20/09/23 18:52:51 INFO namenode.FSNamesystem: isPermissionEnabled = true
20/09/23 18:52:51 INFO namenode.FSNamesystem: HA Enabled: false
20/09/23 18:52:51 INFO namenode.FSNamesystem: Append Enabled: true
20/09/23 18:52:53 INFO util.GSet: Computing capacity for map INodeMap
20/09/23 18:52:53 INFO util.GSet: VM type       = 64-bit
20/09/23 18:52:53 INFO util.GSet: 1.0% max memory 966.7 MB = 9.7 MB
20/09/23 18:52:53 INFO util.GSet: capacity      = 2^20 = 1048576 entries
20/09/23 18:52:53 INFO namenode.FSDirectory: ACLs enabled? false
20/09/23 18:52:53 INFO namenode.FSDirectory: XAttrs enabled? true
20/09/23 18:52:53 INFO namenode.FSDirectory: Maximum size of an xattr: 16384
20/09/23 18:52:53 INFO namenode.NameNode: Caching file names occuring more than 10 times
20/09/23 18:52:53 INFO util.GSet: Computing capacity for map cachedBlocks
20/09/23 18:52:53 INFO util.GSet: VM type       = 64-bit
20/09/23 18:52:53 INFO util.GSet: 0.25% max memory 966.7 MB = 2.4 MB
20/09/23 18:52:53 INFO util.GSet: capacity      = 2^18 = 262144 entries
20/09/23 18:52:53 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
20/09/23 18:52:53 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
20/09/23 18:52:53 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension     = 30000
20/09/23 18:52:53 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10
20/09/23 18:52:53 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10
20/09/23 18:52:53 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
20/09/23 18:52:53 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
20/09/23 18:52:53 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
20/09/23 18:52:53 INFO util.GSet: Computing capacity for map NameNodeRetryCache
20/09/23 18:52:53 INFO util.GSet: VM type       = 64-bit
20/09/23 18:52:53 INFO util.GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
20/09/23 18:52:53 INFO util.GSet: capacity      = 2^15 = 32768 entries
20/09/23 18:52:53 INFO namenode.FSImage: Allocated new BlockPoolId: BP-1825808312-192.168.110.101-1600858373423
20/09/23 18:52:54 INFO common.Storage: Storage directory /opt/SoftWare/hadoop-2.7.7/dfs/name has been successfully formatted.
20/09/23 18:52:54 INFO namenode.FSImageFormatProtobuf: Saving image file /opt/SoftWare/hadoop-2.7.7/dfs/name/current/fsimage.ckpt_0000000000000000000 using no compression
20/09/23 18:52:54 INFO namenode.FSImageFormatProtobuf: Image file /opt/SoftWare/hadoop-2.7.7/dfs/name/current/fsimage.ckpt_0000000000000000000 of size 321 bytes saved in 0 seconds.
20/09/23 18:52:54 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
20/09/23 18:52:54 INFO util.ExitUtil: Exiting with status 0
20/09/23 18:52:55 INFO namenode.NameNode: SHUTDOWN_MSG: 
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at node04/192.168.110.101
************************************************************/
  1. 启动hadoop
[root@node04 ~]# start-dfs.sh
[root@node04 ~]# start-yarn.sh

<!--使用jps命令查看正在运行的进程-->
[root@node04 hadoop]# jps
10144 Jps
9734 ResourceManager
9354 NameNode
9851 NodeManager
9486 DataNode

[root@node05 hadoop]# jps
7840 Jps
7555 DataNode
7715 NodeManager
7644 SecondaryNameNode

[root@node06 hadoop]# jps
7704 Jps
7579 NodeManager
7484 DataNode

成功启动hadoop集群

  1. 浏览器访问
    在浏览器中输入http://192.168.110.101:50070(大家在写得时候要把192.168.110.101改成自己主节点所在的ip地址)出现以下界面,说明配置成功
    在这里插入图片描述
    如果出现不能访问的情况,关闭防火墙就可以了。
  • 0
    点赞
  • 3
    收藏
    觉得还不错? 一键收藏
  • 打赏
    打赏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包

打赏作者

疯狂的小强呀

你的鼓励将是我创作的最大动力

¥1 ¥2 ¥4 ¥6 ¥10 ¥20
扫码支付:¥1
获取中
扫码支付

您的余额不足,请更换扫码支付或充值

打赏作者

实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值