Hadoop基础教程-第2章 Hadoop快速入门(2.4 Hadoop单机运行)

第2章 Hadoop快速入门

2.4 Hadoop单机运行

紧接上一节内容,首先切换到Hadoop根目录

这里写图片描述 
或者cd /opt/hadoop-2.7.3进入Hadoop根目录 
通过pwd命令可以知道当前所在目录

[root@node1 hadoop-2.7.3]# pwd

这里写图片描述

注意:本节命令都将在/opt/hadoop-2.7.3目录下执行。

2.4.1 namenode格式化

执行bin/hadoop namenode -format命令,进行namenode格式化

[root@node1 hadoop-2.7.3]# bin/hadoop namenode -format

这里写图片描述 
输出信息(下半部分):

17/05/12 05:59:11 INFO namenode.NameNode: registered UNIX signal handlers for [TERM, HUP, INT]
17/05/12 05:59:11 INFO namenode.NameNode: createNameNode [-format]
Formatting using clusterid: CID-db9a34c9-661e-4fc0-a273-b554e0cfb32b
17/05/12 05:59:12 INFO namenode.FSNamesystem: No KeyProvider found.
17/05/12 05:59:12 INFO namenode.FSNamesystem: fsLock is fair:true
17/05/12 05:59:12 INFO blockmanagement.DatanodeManager: dfs.block.invalidate.limit=1000
17/05/12 05:59:12 INFO blockmanagement.DatanodeManager: dfs.namenode.datanode.registration.ip-hostname-check=true
17/05/12 05:59:12 INFO blockmanagement.BlockManager: dfs.namenode.startup.delay.block.deletion.sec is set to 000:00:00:00.000
17/05/12 05:59:12 INFO blockmanagement.BlockManager: The block deletion will start around 2017 May 12 05:59:12
17/05/12 05:59:12 INFO util.GSet: Computing capacity for map BlocksMap
17/05/12 05:59:12 INFO util.GSet: VM type       = 64-bit
17/05/12 05:59:12 INFO util.GSet: 2.0% max memory 966.7 MB = 19.3 MB
17/05/12 05:59:12 INFO util.GSet: capacity      = 2^21 = 2097152 entries
17/05/12 05:59:12 INFO blockmanagement.BlockManager: dfs.block.access.token.enable=false
17/05/12 05:59:12 INFO blockmanagement.BlockManager: defaultReplication         = 1
17/05/12 05:59:12 INFO blockmanagement.BlockManager: maxReplication             = 512
17/05/12 05:59:12 INFO blockmanagement.BlockManager: minReplication             = 1
17/05/12 05:59:12 INFO blockmanagement.BlockManager: maxReplicationStreams      = 2
17/05/12 05:59:12 INFO blockmanagement.BlockManager: replicationRecheckInterval = 3000
17/05/12 05:59:12 INFO blockmanagement.BlockManager: encryptDataTransfer        = false
17/05/12 05:59:12 INFO blockmanagement.BlockManager: maxNumBlocksToLog          = 1000
17/05/12 05:59:12 INFO namenode.FSNamesystem: fsOwner             = root (auth:SIMPLE)
17/05/12 05:59:12 INFO namenode.FSNamesystem: supergroup          = supergroup
17/05/12 05:59:12 INFO namenode.FSNamesystem: isPermissionEnabled = true
17/05/12 05:59:12 INFO namenode.FSNamesystem: HA Enabled: false
17/05/12 05:59:12 INFO namenode.FSNamesystem: Append Enabled: true
17/05/12 05:59:13 INFO util.GSet: Computing capacity for map INodeMap
17/05/12 05:59:13 INFO util.GSet: VM type       = 64-bit
17/05/12 05:59:13 INFO util.GSet: 1.0% max memory 966.7 MB = 9.7 MB
17/05/12 05:59:13 INFO util.GSet: capacity      = 2^20 = 1048576 entries
17/05/12 05:59:13 INFO namenode.FSDirectory: ACLs enabled? false
17/05/12 05:59:13 INFO namenode.FSDirectory: XAttrs enabled? true
17/05/12 05:59:13 INFO namenode.FSDirectory: Maximum size of an xattr: 16384
17/05/12 05:59:13 INFO namenode.NameNode: Caching file names occuring more than 10 times
17/05/12 05:59:13 INFO util.GSet: Computing capacity for map cachedBlocks
17/05/12 05:59:13 INFO util.GSet: VM type       = 64-bit
17/05/12 05:59:13 INFO util.GSet: 0.25% max memory 966.7 MB = 2.4 MB
17/05/12 05:59:13 INFO util.GSet: capacity      = 2^18 = 262144 entries
17/05/12 05:59:13 INFO namenode.FSNamesystem: dfs.namenode.safemode.threshold-pct = 0.9990000128746033
17/05/12 05:59:13 INFO namenode.FSNamesystem: dfs.namenode.safemode.min.datanodes = 0
17/05/12 05:59:13 INFO namenode.FSNamesystem: dfs.namenode.safemode.extension     = 30000
17/05/12 05:59:13 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.window.num.buckets = 10
17/05/12 05:59:13 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.num.users = 10
17/05/12 05:59:13 INFO metrics.TopMetrics: NNTop conf: dfs.namenode.top.windows.minutes = 1,5,25
17/05/12 05:59:13 INFO namenode.FSNamesystem: Retry cache on namenode is enabled
17/05/12 05:59:13 INFO namenode.FSNamesystem: Retry cache will use 0.03 of total heap and retry cache entry expiry time is 600000 millis
17/05/12 05:59:13 INFO util.GSet: Computing capacity for map NameNodeRetryCache
17/05/12 05:59:13 INFO util.GSet: VM type       = 64-bit
17/05/12 05:59:13 INFO util.GSet: 0.029999999329447746% max memory 966.7 MB = 297.0 KB
17/05/12 05:59:13 INFO util.GSet: capacity      = 2^15 = 32768 entries
17/05/12 05:59:13 INFO namenode.FSImage: Allocated new BlockPoolId: BP-2055065474-192.168.80.131-1494583153319
17/05/12 05:59:13 INFO common.Storage: Storage directory /tmp/hadoop-root/dfs/name has been successfully formatted.
17/05/12 05:59:13 INFO namenode.FSImageFormatProtobuf: Saving image file /tmp/hadoop-root/dfs/name/current/fsimage.ckpt_0000000000000000000 using no compression
17/05/12 05:59:13 INFO namenode.FSImageFormatProtobuf: Image file /tmp/hadoop-root/dfs/name/current/fsimage.ckpt_0000000000000000000 of size 351 bytes saved in 0 seconds.
17/05/12 05:59:13 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
17/05/12 05:59:13 INFO util.ExitUtil: Exiting with status 0
17/05/12 05:59:13 INFO namenode.NameNode: SHUTDOWN_MSG: 
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at node1/192.168.80.131
************************************************************/

这里写图片描述

在执行结果中可以找到has been successfully formatted,说明namenode格式化成功了!

2.4.2 启动HDFS

执行sbin/start-dfs.sh命令启动HDFS

[root@node1 hadoop-2.7.3]# sbin/start-dfs.sh
Starting namenodes on [node1]
node1: starting namenode, logging to /opt/hadoop-2.7.3/logs/hadoop-root-namenode-node1.out
localhost: starting datanode, logging to /opt/hadoop-2.7.3/logs/hadoop-root-datanode-node1.out
Starting secondary namenodes [0.0.0.0]
The authenticity of host '0.0.0.0 (0.0.0.0)' can't be established.
ECDSA key fingerprint is e2:9a:7d:70:25:24:45:11:97:12:35:e0:45:4c:64:31.
Are you sure you want to continue connecting (yes/no)? yes
0.0.0.0: Warning: Permanently added '0.0.0.0' (ECDSA) to the list of known hosts.
0.0.0.0: starting secondarynamenode, logging to /opt/hadoop-2.7.3/logs/hadoop-root-secondarynamenode-node1.out

在启动HDFS过程中,按照提示输入“yes” 
这里写图片描述

通过jps命令可以查看Java进程

[root@node1 hadoop-2.7.3]# jps

这里写图片描述

jps(Java Virtual Machine Process Status Tool)是JDK 1.5开始提供的一个显示当前所有Java进程pid的命令,简单实用,非常适合在Linux/unix平台上简单察看当前java进程的一些简单情况。 
jps -l输出应用程序main class的完整package名 或者 应用程序的jar文件完整路径名

[root@node1 ~]# jps -l
5752 sun.tools.jps.Jps
5342 org.apache.hadoop.hdfs.server.namenode.NameNode
5440 org.apache.hadoop.hdfs.server.datanode.DataNode
5630 org.apache.hadoop.hdfs.server.namenode.SecondaryNameNode

2.4.3 启动YARN

通过命令sbin/start-yarn.sh启动YARN

[root@node1 hadoop-2.7.3]# sbin/start-yarn.sh
starting yarn daemons
starting resourcemanager, logging to /opt/hadoop-2.7.3/logs/yarn-root-resourcemanager-node1.out
localhost: starting nodemanager, logging to /opt/hadoop-2.7.3/logs/yarn-root-nodemanager-node1.out

这里写图片描述 
然后通过jps查看YARN的进程

[root@node1 hadoop-2.7.3]# jps
5440 DataNode
7253 ResourceManager
7351 NodeManager
5342 NameNode
5630 SecondaryNameNode
7662 Jps

[root@node1 hadoop-2.7.3]# 

这里写图片描述

可以看到多了ResourceManager和NodeManager两个进程。

2.4.4 HDFS WEB界面

HDFS的Web界面默认端口号是50070。 
因为宿主机Windows的hosts文件没有配置虚拟机相关IP信息,所以需要通过IP地址来访问HDFS WEB界面,在浏览器中打开:http://192.168.80.131:50070

这里写图片描述
单击页面上部的导航栏中的“Datanodes” 
这里写图片描述

2.4.5 YARN WEB界面

YARN的Web界面默认端口号是8088。 
http://192.168.80.131:8088

这里写图片描述

单击左侧菜单栏的“Nodes”,可以查看NodeManager信息

这里写图片描述

灰常灰常感谢原博主的辛苦工作,为防止删博,所以转载,只供学习使用,不做其他任何商业用途。 https://blog.csdn.net/chengyuqiang/article/details/71699382
  • 0
    点赞
  • 1
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
好的,下面是Hadoop安装教程-单机-伪分布式配置-Hadoop2.6.0-Ubuntu14.04详细教程教学: 1. 安装Java Hadoop是基于Java开发的,因此需要先安装Java。可以在终端中输入以下命令进行安装: ```bash sudo apt-get update sudo apt-get install default-jdk ``` 2. 下载Hadoop 可以在Hadoop官网下载Hadoop压缩包,也可以在终端中输入以下命令进行下载: ```bash wget https://archive.apache.org/dist/hadoop/core/hadoop-2.6.0/hadoop-2.6.0.tar.gz ``` 3. 解压Hadoop 在终端中输入以下命令解压Hadoop压缩包: ```bash tar -xzvf hadoop-2.6.0.tar.gz ``` 4. 配置Hadoop环境变量 打开终端中的/etc/environment文件,并在其中添加以下内容: ```bash export HADOOP_HOME=/path/to/hadoop export PATH=$PATH:$HADOOP_HOME/bin ``` 其中,/path/to/hadoopHadoop所在的路径。保存文件后,输入以下命令使配置生效: ```bash source /etc/environment ``` 5. 配置Hadoop 进入Hadoop目录,编辑etc/hadoop/hadoop-env.sh文件。找到以下行: ```bash export JAVA_HOME=${JAVA_HOME} ``` 将其修改为: ```bash export JAVA_HOME=/usr/lib/jvm/default-java ``` 保存文件并关闭。 6. 配置HDFS 进入etc/hadoop目录,编辑hdfs-site.xml文件。添加以下内容: ```xml <configuration> <property> <name>dfs.replication</name> <value>1</value> </property> <property> <name>dfs.namenode.name.dir</name> <value>file:/path/to/hadoop/data/dfs/namenode</value> </property> <property> <name>dfs.datanode.data.dir</name> <value>file:/path/to/hadoop/data/dfs/datanode</value> </property> </configuration> ``` 其中,/path/to/hadoop/data为Hadoop数据存储路径。 7. 配置YARN 进入etc/hadoop目录,编辑yarn-site.xml文件。添加以下内容: ```xml <configuration> <property> <name>yarn.nodemanager.aux-services</name> <value>mapreduce_shuffle</value> </property> <property> <name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name> <value>org.apache.hadoop.mapred.ShuffleHandler</value> </property> </configuration> ``` 8. 配置MapReduce 进入etc/hadoop目录,编辑mapred-site.xml文件。添加以下内容: ```xml <configuration> <property> <name>mapreduce.framework.name</name> <value>yarn</value> </property> </configuration> ``` 9. 格式化HDFS 在终端中输入以下命令格式化HDFS: ```bash hdfs namenode -format ``` 10. 启动Hadoop 在终端中输入以下命令启动Hadoop: ```bash start-all.sh ``` 11. 验证Hadoop 在浏览器中输入http://localhost:50070,即可访问Hadoop Web UI。在该页面中可以查看HDFS的状态、数据块、节点等信息。 以上就是Hadoop安装教程-单机-伪分布式配置-Hadoop2.6.0-Ubuntu14.04详细教程教学。希望对你有所帮助!
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值