2 Hadoop 正式安装

1. 部署

1.1 安装hadoop

mkdir /opt/bigdata
tar xf hadoop-2.6.5.tar.gz -C /opt/bigdata/
cd /opt/bigdata/
ln -sf hadoop-2.6.5 hadoop
详细过程


[root@node01 ~]# mkdir /opt/bigdata
[root@node01 ~]# cd /usr/local/src/
[root@node01 src]# ls
hadoop-2.6.5.tar.gz  InstallJDK8.sh  JDK8_112.zip  jdk-8u112-linux-x64.tar.gz
[root@node01 src]# tar xf hadoop-2.6.5.tar.gz -C /opt/bigdata/
[root@node01 src]# cd /opt/bigdata/
[root@node01 bigdata]# ls
hadoop-2.6.5
[root@node01 bigdata]# ln -sf hadoop-2.6.5 hadoop
[root@node01 bigdata]# ll
total 0
lrwxrwxrwx 1 root root  12 Mar 11 21:14 hadoop -> hadoop-2.6.5
drwxrwxr-x 9 1000 1000 149 Oct  3  2016 hadoop-2.6.5

1.2 添加hadoop环境变量

cat>>/etc/profile<<EOF
# add by caimengzhi 
HADOOP_HOME=/opt/bigdata/hadoop
export PATH=\$PATH:\$HADOOP_HOME/bin
export PATH=\$PATH:\$HADOOP_HOME/sbin
EOF
source /etc/profile
which hadoop
详细操作


[root@node01 bigdata]# cat>>/etc/profile<<EOF
> # add by caimengzhi
> HADOOP_HOME=/opt/bigdata/hadoop
> export PATH=\$PATH:\$HADOOP_HOME/bin
> export PATH=\$PATH:\$HADOOP_HOME/sbin
> EOF
[root@node01 bigdata]# tail -4 /etc/profile
# add by caimengzhi
HADOOP_HOME=/opt/bigdata/hadoop
export PATH=$PATH:$HADOOP_HOME/bin
export PATH=$PATH:$HADOOP_HOME/sbin
[root@node01 bigdata]# source /etc/profile
[root@node01 bigdata]# which hadoop
/opt/bigdata/hadoop/bin/hadoop

1.3 指定JAVA_HOME

cd /opt/bigdata/hadoop/etc/hadoop/
sed -i 's@export JAVA_HOME=${JAVA_HOME}@export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_112@g' hadoop-env.sh
grep -irn '^export JAVA_HOME=' hadoop-env.sh

sed -i 's@# export JAVA_HOME=/home/y/libexec/jdk1.6.0/@export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_112@g' yarn-env.sh
grep -irn 'export JAVA_HOME=' yarn-env.sh
详细操作


[root@node01 hadoop]# sed -i 's@export JAVA_HOME=${JAVA_HOME}@export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_112@g' hadoop-env.sh
[root@node01 hadoop]# grep -irn 'export JAVA_HOME=' hadoop-env.sh
25:export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_112

[root@node01 hadoop]# sed -i 's@# export JAVA_HOME=/home/y/libexec/jdk1.6.0/@export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_112@g' yarn-env.sh
[root@node01 hadoop]# grep -irn 'export JAVA_HOME=' yarn-env.sh
23:export JAVA_HOME=/usr/lib/jvm/jdk1.8.0_112
指定的目的是 在于SSH远程操作的启动hadoop的时候,ssh不加载远程环境变量。

1.4 配置core-site.xml

添加RPC配置

core-site.xml


[root@node01 hadoop]# cat core-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->

<!-- Put site-specific property overrides in this file. -->

<configuration>
    <property>
        <name>fs.defaultFS</name>
        <value>hdfs://node01:9000</value>
    </property>
    <property>
        <name>hadoop.tmp.dir</name>
        <value>file:/opt/bigdata/hadoop/tmp</value>
    </property>
</configuration>
[root@node01 hadoop]# mkdir -p /opt/bigdata/hadoop/tmp

1.5 配置hdfs-site.xml

修改配置文件 添加HDFS配置

hdfs-site.xml


[root@node01 hadoop]# cat hdfs-site.xml
<?xml version="1.0" encoding="UTF-8"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->

<!-- Put site-specific property overrides in this file. -->
<configuration>
    <property>
        <name>dfs.namenode.secondary.http-address</name>
        <value>node02:50090</value>
    </property>
    <property>
        <name>dfs.namenode.name.dir</name>
        <value>file:/opt/bigdata/hadoop/dfs/name</value>
    </property>
    <property>
        <name>dfs.datanode.data.dir</name>
        <value>file:/opt/bigdata/hadoop/dfs/data</value>
    </property>
    <property>
        <name>dfs.replication</name>
        <value>3</value>
    </property>
    <property>
        <name>dfs.namenode.checkpoint.dir</name>
        <value>file:/opt/bigdata/hadoop/dfs/secondary</value>
    </property>
</configuration>
[root@node01 hadoop]# mkdir -p /opt/bigdata/hadoop/dfs/{name,data,secondary}

1.6 添加从节点

cat>slaves<<EOF
node01
node02
node03
node04
EOF
详细操作


[root@node01 hadoop]# cat slaves
node01
node02
node03
node04

1.7 配置mapred-site.xml

mapred-site.xml


[root@node01 hadoop]# cp mapred-site.xml.template mapred-site.xml
[root@node01 hadoop]# vim mapred-site.xml
[root@node01 hadoop]# cat mapred-site.xml
<?xml version="1.0"?>
<?xml-stylesheet type="text/xsl" href="configuration.xsl"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->

<!-- Put site-specific property overrides in this file. -->
<configuration>
    <property>
        <name>mapreduce.framework.name</name>
        <value>yarn</value>
    </property>
</configuration>

1.8 配置yarn-site.xml

[root@node01 hadoop]# cat yarn-site.xml
<?xml version="1.0"?>
<!--
  Licensed under the Apache License, Version 2.0 (the "License");
  you may not use this file except in compliance with the License.
  You may obtain a copy of the License at

    http://www.apache.org/licenses/LICENSE-2.0

  Unless required by applicable law or agreed to in writing, software
  distributed under the License is distributed on an "AS IS" BASIS,
  WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
  See the License for the specific language governing permissions and
  limitations under the License. See accompanying LICENSE file.
-->

<configuration>

<!-- Site specific YARN configuration properties -->

    <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce_shuffle</value>
    </property>
    <property>
        <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
        <value>org.apache.hadoop.mapred.ShuffleHandler</value>
    </property>
    <property>
        <name>yarn.resourcemanager.address</name>
        <value>node01:8032</value>
    </property>
    <property>
        <name>yarn.resourcemanager.scheduler.address</name>
        <value>node01:8030</value>
    </property>
    <property>
        <name>yarn.resourcemanager.resource-tracker.address</name>
        <value>node01:8035</value>
    </property>
    <property>
        <name>yarn.resourcemanager.admin.address</name>
        <value>node01:8033</value>
    </property>
    <property>
        <name>yarn.resourcemanager.webapp.address</name>
        <value>node01:8088</value>
    </property>
</configuration>

1.9 拷贝

拷贝环境变量配置文件到从

scp /etc/profile root@node02:/etc/profile
scp /etc/profile root@node03:/etc/profile
scp /etc/profile root@node04:/etc/profile
[root@node01 hadoop]# scp /etc/profile root@node02:/etc/profile
[root@node01 hadoop]# scp /etc/profile root@node03:/etc/profile
[root@node01 hadoop]# scp /etc/profile root@node04:/etc/profile
拷贝hadoop源码包到从

rsync -az /opt/bigdata/{hadoop,hadoop-2.6.5} root@node02:/opt/bigdata/
rsync -az /opt/bigdata/{hadoop,hadoop-2.6.5} root@node03:/opt/bigdata/
rsync -az /opt/bigdata/{hadoop,hadoop-2.6.5} root@node04:/opt/bigdata/

[root@node01 hadoop]# rsync -az /opt/bigdata/{hadoop,hadoop-2.6.5} root@node02:/opt/bigdata/
[root@node01 hadoop]# rsync -az /opt/bigdata/{hadoop,hadoop-2.6.5} root@node03:/opt/bigdata/
[root@node01 hadoop]# rsync -az /opt/bigdata/{hadoop,hadoop-2.6.5} root@node04:/opt/bigdata/
重新加载环境变量

[root@node02 ~]# source /etc/profile
[root@node03 ~]# source /etc/profile
[root@node04 ~]# source /etc/profile

1.0 格式化

hdfs 格式化[生成fsimage]


[root@node01 hadoop]# hadoop namenode -format
DEPRECATED: Use of this script to execute hdfs command is deprecated.
Instead use the hdfs command for it.

20/03/11 09:58:18 INFO namenode.NameNode: STARTUP_MSG:
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = node01/192.168.71.133
STARTUP_MSG:   args = [-format]
STARTUP_MSG:   version = 2.6.5
STARTUP_MSG:   classpath = /opt/bigdata/hadoop-2.6.5/etc/hadoop:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/activation-1.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jackson-mapper-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/java-xmlbuilder-0.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/protobuf-java-2.5.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/guava-11.0.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-net-3.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jackson-jaxrs-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/servlet-api-2.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/httpclient-4.2.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/xz-1.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-cli-1.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/slf4j-api-1.7.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jersey-server-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/stax-api-1.0-2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jersey-json-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/htrace-core-3.0.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/asm-3.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/hadoop-annotations-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-collections-3.2.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/curator-framework-2.6.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-configuration-1.6.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-math3-3.1.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-digester-1.8.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/apacheds-i18n-2.0.0-M15.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jasper-runtime-5.5.23.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/httpcore-4.2.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/curator-recipes-2.6.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/api-util-1.0.0-M20.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jsr305-1.3.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/junit-4.11.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jettison-1.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jets3t-0.9.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jsp-api-2.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/paranamer-2.3.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-io-2.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/zookeeper-3.4.6.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-el-1.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jersey-core-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jaxb-impl-2.2.3-1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jetty-util-6.1.26.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-lang-2.6.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/hadoop-auth-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/avro-1.7.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jsch-0.1.42.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jasper-compiler-5.5.23.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/apacheds-kerberos-codec-2.0.0-M15.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jackson-xc-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/netty-3.6.2.Final.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-beanutils-1.7.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/gson-2.2.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-codec-1.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jetty-6.1.26.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-httpclient-3.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-logging-1.1.3.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/curator-client-2.6.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/xmlenc-0.52.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-beanutils-core-1.8.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jackson-core-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/hamcrest-core-1.3.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/api-asn1-api-1.0.0-M20.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/log4j-1.2.17.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/commons-compress-1.4.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/jaxb-api-2.2.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/mockito-all-1.8.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/lib/snappy-java-1.0.4.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/hadoop-common-2.6.5-tests.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/hadoop-nfs-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/common/hadoop-common-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jackson-mapper-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/protobuf-java-2.5.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/guava-11.0.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/servlet-api-2.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/commons-cli-1.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jersey-server-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/htrace-core-3.0.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/asm-3.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jasper-runtime-5.5.23.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jsr305-1.3.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/xercesImpl-2.9.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jsp-api-2.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/commons-io-2.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/commons-daemon-1.0.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/commons-el-1.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jersey-core-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jetty-util-6.1.26.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/commons-lang-2.6.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/netty-3.6.2.Final.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/commons-codec-1.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jetty-6.1.26.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/commons-logging-1.1.3.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/xml-apis-1.3.04.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/xmlenc-0.52.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/jackson-core-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/lib/log4j-1.2.17.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/hadoop-hdfs-2.6.5-tests.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/hadoop-hdfs-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/hdfs/hadoop-hdfs-nfs-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/activation-1.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jackson-mapper-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/protobuf-java-2.5.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/guava-11.0.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jackson-jaxrs-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/servlet-api-2.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jline-0.9.94.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/javax.inject-1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/xz-1.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-cli-1.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jersey-server-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/stax-api-1.0-2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jersey-json-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/asm-3.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-collections-3.2.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/aopalliance-1.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jsr305-1.3.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jettison-1.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-io-2.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/zookeeper-3.4.6.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/guice-servlet-3.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jersey-core-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jaxb-impl-2.2.3-1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jersey-client-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jetty-util-6.1.26.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-lang-2.6.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jersey-guice-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jackson-xc-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/netty-3.6.2.Final.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/guice-3.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-codec-1.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jetty-6.1.26.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/leveldbjni-all-1.8.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-httpclient-3.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-logging-1.1.3.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jackson-core-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/log4j-1.2.17.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/commons-compress-1.4.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/lib/jaxb-api-2.2.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-common-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-server-resourcemanager-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-server-applicationhistoryservice-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-applications-distributedshell-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-api-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-server-tests-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-registry-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-server-nodemanager-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-server-web-proxy-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-client-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-server-common-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/yarn/hadoop-yarn-applications-unmanaged-am-launcher-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/jackson-mapper-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/protobuf-java-2.5.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/javax.inject-1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/xz-1.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/jersey-server-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/asm-3.2.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/hadoop-annotations-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/aopalliance-1.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/junit-4.11.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/paranamer-2.3.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/commons-io-2.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/guice-servlet-3.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/jersey-core-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/avro-1.7.4.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/jersey-guice-1.9.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/netty-3.6.2.Final.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/guice-3.0.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/leveldbjni-all-1.8.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/jackson-core-asl-1.9.13.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/hamcrest-core-1.3.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/log4j-1.2.17.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/commons-compress-1.4.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/lib/snappy-java-1.0.4.1.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-common-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-shuffle-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-core-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-app-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.6.5-tests.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-jobclient-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-2.6.5.jar:/opt/bigdata/hadoop-2.6.5/share/hadoop/mapreduce/hadoop-mapreduce-client-hs-plugins-2.6.5.jar:/contrib/capacity-scheduler/*.jar:/contrib/capacity-scheduler/*.jar
STARTUP_MSG:   build = https://github.com/apache/hadoop.git -r e8c9fe0b4c252caf2ebf1464220599650f119997; compiled by 'sjlee' on 2016-10-02T23:43Z
STARTUP_MSG:   java = 1.8.0_112
************************************************************/
20/03/11 09:58:18 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
20/03/11 09:58:18 INFO namenode.NameNode: createNameNode [-format]
Formatting using clusterid: CID-04102693-ae36-486e-a01b-a3df291bc13a
20/03/11 09:58:21 INFO namenode.FSNamesystem: No KeyProvider found.
20/03/11 09:58:21 INFO namenode.FSNamesystem: fsLock is fair:true
20/03/11 09:58:21 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000
20/03/11 09:58:21 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
20/03/11 09:58:21 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
20/03/11 09:58:21 INFO blockmanagement.BlockManager: The block deletion will start around 2020 Mar 11 09:58:21
20/03/11 09:58:21 INFO util.GSet: Computing capacity for map BlocksMap
20/03/11 09:58:21 INFO util.GSet: VM type       = 64-bit
20/03/11 09:58:21 INFO util.GSet: 2.0% max memory 966.7 MB = 19.3 MB
20/03/11 09:58:21 INFO util.GSet: capacity      = 2^21 = 2097152 entries
20/03/11 09:58:21 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false
20/03/11 09:58:21 INFO blockmanagement.BlockManager: defaultReplication         = 3
20/03/11 09:58:21 INFO blockmanagement.BlockManager: maxReplication             = 512
20/03/11 09:58:21 INFO blockmanagement.BlockManager: minReplication             = 1
20/03/11 09:58:21 INFO blockmanagement.BlockManager: maxReplicationStreams      = 2
20/03/11 09:58:21 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000
20/03/11 09:58:21 INFO blockmanagement.BlockManager: encryptDataTransfer        = false
20/03/11 09:58:21 INFO blockmanagement.BlockManager: maxNumBlocksToLog          = 1000
20/03/11 09:58:21 INFO namenode.FSNamesystem: fsOwner             = root (auth:SIMPLE)
20/03/11 09:58:21 INFO namenode.FSNamesystem: supergroup          = supergroup
20/03/11 09:58:21 INFO namenode.FSNamesystem: isPermissionEnabled = true
20/03/11 09:58:21 INFO namenode.FSNamesystem: HA Enabled: false
20/03/11 09:58:21 INFO namenode.FSNamesystem: Append Enabled: true
20/03/11 09:58:22 INFO util.GSet: Computing capacity for map INodeMap
20/03/11 09:58:22 INFO util.GSet: VM type       = 64-bit
20/03/11 09:58:22 INFO util.GSet: 1.0% max memory 966.7 MB = 9.7 MB
20/03/11 09:58:22 INFO util.GSet: capacity      = 2^20 = 1048576 entries
20/03/11 09:58:22 INFO namenode.NameNode: Caching file names occuring more than 10 times
20/03/11 09:58:22 INFO util.GSet: Computing capacity for map cachedBlocks
20/03/11 09:58:22 INFO util.GSet: VM type       = 64-bit
20/03/11 09:58:22 INFO util.GSet: 0.25% max memory 966.7 MB = 2.4 MB
20/03/11 09:58:22 INFO util.GSet: capacity      = 2^18 = 262144 entries
20/03/11 09:58:22 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
20/03/11 09:58:22 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
20/03/11 09:58:22 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension     = 30000
20/03/11 09:58:22 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
20/03/11 09:58:22 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
20/03/11 09:58:22 INFO util.GSet: Computing capacity for map NameNodeRetryCache
20/03/11 09:58:22 INFO util.GSet: VM type       = 64-bit
20/03/11 09:58:22 INFO util.GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
20/03/11 09:58:22 INFO util.GSet: capacity      = 2^15 = 32768 entries
20/03/11 09:58:22 INFO namenode.NNConf: ACLs enabled? false
20/03/11 09:58:22 INFO namenode.NNConf: XAttrs enabled? true
20/03/11 09:58:22 INFO namenode.NNConf: Maximum size of an xattr: 16384
20/03/11 09:58:22 INFO namenode.FSImage: Allocated new BlockPoolId: BP-218129651-192.168.71.133-1583935102537
20/03/11 09:58:22 INFO common.Storage: Storage directory /opt/bigdata/hadoop/dfs/name has been successfully formatted.
20/03/11 09:58:22 INFO namenode.FSImageFormatProtobuf: Saving image file /opt/bigdata/hadoop/dfs/name/current/fsimage.ckpt_0000000000000000000 using no compression
20/03/11 09:58:23 INFO namenode.FSImageFormatProtobuf: Image file /opt/bigdata/hadoop/dfs/name/current/fsimage.ckpt_0000000000000000000 of size 321 bytes saved in 0 seconds.
20/03/11 09:58:23 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
20/03/11 09:58:23 INFO util.ExitUtil: Exiting with status 0
20/03/11 09:58:23 INFO namenode.NameNode: SHUTDOWN_MSG:
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at node01/192.168.71.133
************************************************************/
提示下列内容即是成功完成格式化

20/03/11 09:58:22 INFO common.Storage: Storage directory /opt/bigdata/hadoop/dfs/name has been successfully formatted.


[root@node01 ~]# cd /opt/bigdata/hadoop/dfs/name/current/
[root@node01 current]# ls
fsimage_0000000000000000000  fsimage_0000000000000000000.md5  seen_txid  VERSION
注意


hadoop namenode -format做了什么
1. 创建目录,并初始化一个空的fsimage
2. VERSION CID

2. 启动集群

2.1 启动集群

[root@node01 hadoop]# start-all.sh
This script is Deprecated. Instead use start-dfs.sh and start-yarn.sh
Starting namenodes on [node01]
node01: starting namenode, logging to /opt/bigdata/hadoop-2.6.5/logs/hadoop-root-namenode-node01.out
node03: starting datanode, logging to /opt/bigdata/hadoop-2.6.5/logs/hadoop-root-datanode-node03.out
node02: starting datanode, logging to /opt/bigdata/hadoop-2.6.5/logs/hadoop-root-datanode-node02.out
node04: starting datanode, logging to /opt/bigdata/hadoop-2.6.5/logs/hadoop-root-datanode-node04.out
node01: starting datanode, logging to /opt/bigdata/hadoop-2.6.5/logs/hadoop-root-datanode-node01.out
Starting secondary namenodes [node02]
node02: starting secondarynamenode, logging to /opt/bigdata/hadoop-2.6.5/logs/hadoop-root-secondarynamenode-node02.out
starting yarn daemons
starting resourcemanager, logging to /opt/bigdata/hadoop-2.6.5/logs/yarn-root-resourcemanager-node01.out
node02: starting nodemanager, logging to /opt/bigdata/hadoop-2.6.5/logs/yarn-root-nodemanager-node02.out
node03: starting nodemanager, logging to /opt/bigdata/hadoop-2.6.5/logs/yarn-root-nodemanager-node03.out
node04: starting nodemanager, logging to /opt/bigdata/hadoop-2.6.5/logs/yarn-root-nodemanager-node04.out
node01: starting nodemanager, logging to /opt/bigdata/hadoop-2.6.5/logs/yarn-root-nodemanager-node01.out
start-all.sh,第一次:datanode和secondary角色初始化创建自己的数据目录

2.2 查看服务进程

node01

[root@node01 hadoop]# jps
6432 Jps
5734 DataNode
5991 ResourceManager
5640 NameNode
6089 NodeManager
node02

[root@node02 ~]# jps
2529 DataNode
2615 SecondaryNameNode
2682 NodeManager
node03

[root@node03 ~]# jps
2899 Jps
2678 DataNode
2776 NodeManager
node04

[root@node04 ~]# jps
2261 DataNode
2488 Jps
2361 NodeManager

2.3 测试hdfs

上传文件到hdfs文件系统中
[root@node01 ~]# hadoop fs -put /etc/passwd /

查看hdfs文件系统中的文件
[root@node01 ~]# hadoop fs -ls /
Found 1 items
-rw-r--r--   3 root supergroup       1161 2020-03-11 10:10 /passwd

读取文件内容
[root@node01 ~]# hadoop fs -text /passwd | tail -3
redis:x:996:995:Redis Database Server:/var/lib/redis:/sbin/nologin
mailnull:x:47:47::/var/spool/mqueue:/sbin/nologin
smmsp:x:51:51::/var/spool/mqueue:/sbin/nologin

2.3.2 web端测试

widows的hosts文件中添加三个主机的名和ip


192.168.71.133 node01
192.168.71.134 node02
192.168.71.135 node03
192.168.71.136 node04
windows C:\Windows\System32\drivers\etc\hosts

Linux /etc/hosts

要在测试的客户端机器上添加这个hosts解析

浏览器输入http://node01:50070 就可以看到后台hdfs数据信息

在这里插入图片描述

2.4 检查

[root@node01 ~]# for i in `seq 100000`;do echo "hello hadoop $i">>data.txt;done
[root@node01 ~]# ls -l -h data.txt
-rw-r--r-- 1 root root 1.6M Mar 11 23:00 data.txt

我已1M大小为单位上传data.txt
[root@node01 ~]# hdfs dfs -D dfs.blocksize=1048576 -put data.txt

在这里插入图片描述

data1.txt文件按照字节[1M=1024*1024=1048756]被切成了8块


[root@node01 subdir0]# pwd
/opt/bigdata/hadoop/dfs/data/current/BP-218129651-192.168.71.133-1583935102537/current/finalized/subdir0/subdir0
[root@node01 subdir0]# ll
total 1868
-rw-r--r-- 1 root root 1048576 Mar 11 23:26 blk_1073741829
-rw-r--r-- 1 root root    8199 Mar 11 23:26 blk_1073741829_1005.meta
-rw-r--r-- 1 root root  840319 Mar 11 23:26 blk_1073741830
-rw-r--r-- 1 root root    6575 Mar 11 23:26 blk_1073741830_1006.meta
[root@node01 subdir0]# tail -10 blk_1073741829
hello hadoop 55764
hello hadoop 55765
hello hadoop 55766
hello hadoop 55767
hello hadoop 55768
hello hadoop 55769
hello hadoop 55770
hello hadoop 55771
hello hadoop 55772
hello hadoop 5[root@node01 subdir0]#
[root@node01 subdir0]#
[root@node01 subdir0]# head -10 blk_1073741830
5773
hello hadoop 55774
hello hadoop 55775
hello hadoop 55776
hello hadoop 55777
hello hadoop 55778
hello hadoop 55779
hello hadoop 55780
hello hadoop 55781
hello hadoop 55782
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值