四、Hadoop初探:伪分布式模式

1、配置

    1.1 core-site.xml配置

        位置:$HADOOP_HOME/etc/hadoop/core-site.xml

<configuration>
    <property>
        <name>fs.defaultFS</name>
        <value>hdfs://localhost:9000</value>
    </property>
</configuration>

        PS:如提示无写权限,执行chmod -R 775 /usr/local/hadoop 添加读写权限

        1.1.1 配置参数

            详见 core-site.xml参数释义

    1.2 hdfs-site.xml配置

        位置:位置$HADOOP_HOME/etc/hadoop/hdfs-site.xml

<configuration>
    <property>
        <name>dfs.replication</name>
        <value>1</value>
    </property>
</configuration>

        1.2.1 配置参数

            详见hdfs-site.xml参数释义

    1.3 设置免密SSH

        首先尝试ssh localhost,如提示需要密码:

173552_RiM7_1866975.png

        则生成ssh密钥,执行如下操作:

$ ssh-keygen -t rsa -P '' -f ~/.ssh/id_rsa
$ cat ~/.ssh/id_rsa.pub >> ~/.ssh/authorized_keys
$ chmod 0600 ~/.ssh/authorized_keys

2、运行

    2.1 格式化文件系统

$ bin/hdfs namenode -format

            显示结果

[hadoop@localhost hadoop]$ bin/hdfs namenode -format
18/05/27 17:43:30 INFO namenode.NameNode: STARTUP_MSG: 
/************************************************************
STARTUP_MSG: Starting NameNode
STARTUP_MSG:   host = localhost/127.0.0.1
STARTUP_MSG:   args = [-format]
STARTUP_MSG:   version = 2.9.1
STARTUP_MSG:   classpath = 

......

18/05/27 17:43:33 INFO common.Storage: Storage directory /tmp/hadoop-hadoop/dfs/name has been successfully formatted.
18/05/27 17:43:33 INFO namenode.FSImageFormatProtobuf: Saving image file /tmp/hadoop-hadoop/dfs/name/current/fsimage.ckpt_0000000000000000000 using no compression
18/05/27 17:43:33 INFO namenode.FSImageFormatProtobuf: Image file /tmp/hadoop-hadoop/dfs/name/current/fsimage.ckpt_0000000000000000000 of size 323 bytes saved in 0 seconds .
18/05/27 17:43:33 INFO namenode.NNStorageRetentionManager: Going to retain 1 images with txid >= 0
18/05/27 17:43:33 INFO namenode.NameNode: SHUTDOWN_MSG: 
/************************************************************
SHUTDOWN_MSG: Shutting down NameNode at localhost/127.0.0.1
************************************************************/

    2.2 启动namenode 和 datanode 守护线程

$ sbin/start-dfs.sh

            提示如下:JAVA_HOME未识别到

[hadoop@localhost hadoop]$ sbin/start-dfs.sh
Starting namenodes on [localhost]
localhost: Error: JAVA_HOME is not set and could not be found.
localhost: Error: JAVA_HOME is not set and could not be found.
Starting secondary namenodes [0.0.0.0]
The authenticity of host '0.0.0.0 (0.0.0.0)' can't be established.
ECDSA key fingerprint is SHA256:KkgCrCih0ZSBx/61V6D30J7m6wAl6HNuD3K0Q/gQobw.
ECDSA key fingerprint is MD5:8c:07:a9:6d:85:66:e4:ca:c8:89:d5:3e:ae:b7:d5:70.
Are you sure you want to continue connecting (yes/no)? yes
0.0.0.0: Warning: Permanently added '0.0.0.0' (ECDSA) to the list of known hosts.
0.0.0.0: Error: JAVA_HOME is not set and could not be found.

            设置hadoop-env.sh文件的java环境变量,编辑$HADOOP_HOME/etc/haddop/hadoop-env.sh

export JAVA_HOME=/usr/local/apps/java

            再次执行 sbin/start-dfs.sh

[hadoop@localhost hadoop]$ sbin/start-dfs.sh
Starting namenodes on [localhost]
localhost: starting namenode, logging to /usr/local/apps/hadoop/logs/hadoop-hadoop-namenode-localhost.localdomain.out
localhost: starting datanode, logging to /usr/local/apps/hadoop/logs/hadoop-hadoop-datanode-localhost.localdomain.out
Starting secondary namenodes [0.0.0.0]
0.0.0.0: starting secondarynamenode, logging to /usr/local/apps/hadoop/logs/hadoop-hadoop-secondarynamenode-localhost.localdomain.out

        访问 http://host:50070

        因为博主采用的是虚拟机CentOS7,防火墙默认启动,所以需要关闭防火墙:

//临时关闭
# systemctl stop firewalld
//禁止开机启动
# systemctl disable firewalld

        关闭防火墙后,访问结果

175907_jx8U_1866975.png

    2.3 测试运行

#新建文件夹
$ bin/hdfs dfs -mkdir /usr
$ bin/hdfs dfs -mkdir /usr/hadoop

       191255_yEwJ_1866975.png

#复制文件
$ bin/hdfs dfs -put etc/hadoop input

        191540_b8ie_1866975.png

#运行测试样例
bin/hadoop jar share/hadoop/mapreduce/hadoop-mapreduce-examples-2.9.1.jar grep input output 'dfs[a-z.]+'
#查看样例运行结果
$ bin/hdfs dfs -cat output/*

    191849_dcln_1866975.png

当然,也可以把结果复制到本地系统中查看

#获取hdfs系统上output文件夹至本地output文件夹
$ bin/hdfs dfs -get output output
$ cat output/*

3 停止

#运行停止脚本,停止namenode 和datanode 守护线程
$ sbin/stop-dfs.sh

4 单机模式YARN配置

    4.1 配置

<configuration>
    <property>
        <name>mapreduce.framework.name</name>
        <value>yarn</value>
    </property>
</configuration>
<configuration>
    <property>
        <name>yarn.nodemanager.aux-services</name>
        <value>mapreduce_shuffle</value>
    </property>
</configuration>

    4.2 启动与停止

  • 启动
$HADOOP_HOME/sbin/start-yarn.sh
  • 停止
$HADOOP_HOME/sbin/stop-yarn.sh
  • 访问http://host:8088

170900_PKp8_1866975.png

转载于:https://my.oschina.net/u/1866975/blog/1819510

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值