Hadoop安装指南

安装环境

OS:Ubuntu Linux 8.0

Java:sun-java 1.6.0.20

Hadoop:hadoop-0.20.2

安装Hadoop之前需要做一些准备工作,系统应该先安装Java和SSH,并且保证sshd一直运行,以便使用Hadoop脚本管理器远端访问Hadoop守护进程。

 

第一步:安装SSH

使用OpenSSH,在配置好源的情况下,输入如下命令:

sudo apt-get install openssh-serveropenssh-client

结果如下:

       出现这种情况的可能原因是,客户端与服务器不匹配。首先删除客户端,运行如下命令:

       sudo apt-get remove openssh-client

       重新运行如下命令:

       sudo apt-get install openssh-server

       选择是继续完成安装。

       可以使用如下3条命令停止、启动、重启SSH。

       /etc/init.d/ssh stop

       /etc/init.d/ssh start

/etc/init.d/ssh restart

 

第二步 安装Java

       Ubuntu默认安装了Java,但那不是我们所需要的。因为它是BSD OpenJDK,而不是SUN JDK。

       可以通过如下命令查询:

       java –version

       使用如下命令来获得JDK:

       sudo apt-get install sun-java6-jdk

       安装过程中会出现DLJ许可协议,阅读同意选择OK回车,即可继续安装。安装完成后可以在/usr/lib/jvm/下发行新文件夹java-6-sun-1.6.0.20。

 

第三步 配置SSH

       Hadoop通过SSH来启动所有的slaves。使用命令$ssh localhost:

       选择yes:

       需要密码访问。重新设置,输入如下内容:

       sudo ssh-keygen –t dsa –P ‘’ –f~/.ssh/id_dsa

       输入如下命令创建一个新文件authorized_keys:

       cat ~/.ssh/id_dsa.pub >>~/.ssh/authorized_keys

       再次$ ssh localhost将不再需要密码。

 

第四步 配置Hadoop

       修改文件conf/hadoop-env.sh。找到Java路径一句:

       # export JAVA_HOME=/usr/lib/j2sdk1.5-sun

       将其改为:

       export JAVA_HOME=/usr/lib/jvm/java-6-sun-1.6.0.20

 

第四步 Hadoop单机模式

       尝试运行$bin/hadoop,将会显示Hadoop脚本的使用文档。

       新建一个文件夹,用于存放测试数据。

       $mkdir input

       将Hadoop配置文件夹的.xml文件放入新建的文件夹,作为测试数据。

       $cp conf/.xml input

       总共放入了5个xml文件。

       运行命令:

       $bin/hadoop jar hadoop-*-examples.jargrep input output ‘dfs[a-z.]+’

       产生的信息如下:

10/05/0419:28:37 INFO jvm.JvmMetrics: Initializing JVM Metrics withprocessName=JobTracker, sessionId=
10/05/04 19:28:37 INFO mapred.FileInputFormat: Total input paths to process : 5
10/05/04 19:28:38 INFO mapred.JobClient: Running job: job_local_0001
10/05/04 19:28:38 INFO mapred.FileInputFormat: Total input paths to process : 5
10/05/04 19:28:38 INFO mapred.MapTask: numReduceTasks: 1
10/05/04 19:28:38 INFO mapred.MapTask: io.sort.mb = 100
10/05/04 19:28:39 INFO mapred.JobClient:  map 0% reduce 0%
10/05/04 19:28:40 INFO mapred.MapTask: data buffer = 79691776/99614720
10/05/04 19:28:40 INFO mapred.MapTask: record buffer = 262144/327680
10/05/04 19:28:40 INFO mapred.MapTask: Starting flush of map output
10/05/04 19:28:40 INFO mapred.TaskRunner: Task:attempt_local_0001_m_000000_0 isdone. And is in the process of commiting
10/05/04 19:28:40 INFO mapred.LocalJobRunner:file:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/input/core-site.xml:0+178
10/05/04 19:28:40 INFO mapred.TaskRunner: Task 'attempt_local_0001_m_000000_0'done.
10/05/04 19:28:40 INFO mapred.MapTask: numReduceTasks: 1
10/05/04 19:28:40 INFO mapred.MapTask: io.sort.mb = 100
10/05/04 19:28:41 INFO mapred.JobClient:  map 100% reduce 0%
10/05/04 19:28:41 INFO mapred.MapTask: data buffer = 79691776/99614720
10/05/04 19:28:41 INFO mapred.MapTask: record buffer = 262144/327680
10/05/04 19:28:41 INFO mapred.MapTask: Starting flush of map output
10/05/04 19:28:41 INFO mapred.MapTask: Finished spill 0
10/05/04 19:28:41 INFO mapred.TaskRunner: Task:attempt_local_0001_m_000001_0 isdone. And is in the process of commiting
10/05/04 19:28:41 INFO mapred.LocalJobRunner:file:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/input/hadoop-policy.xml:0+4190
10/05/04 19:28:41 INFO mapred.TaskRunner: Task 'attempt_local_0001_m_000001_0'done.
10/05/04 19:28:41 INFO mapred.MapTask: numReduceTasks: 1
10/05/04 19:28:41 INFO mapred.MapTask: io.sort.mb = 100
10/05/04 19:28:42 INFO mapred.MapTask: data buffer = 79691776/99614720
10/05/04 19:28:42 INFO mapred.MapTask: record buffer = 262144/327680
10/05/04 19:28:42 INFO mapred.MapTask: Starting flush of map output
10/05/04 19:28:42 INFO mapred.TaskRunner: Task:attempt_local_0001_m_000002_0 isdone. And is in the process of commiting
10/05/04 19:28:42 INFO mapred.LocalJobRunner:file:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/input/capacity-scheduler.xml:0+3936
10/05/04 19:28:42 INFO mapred.TaskRunner: Task 'attempt_local_0001_m_000002_0'done.
10/05/04 19:28:42 INFO mapred.MapTask: numReduceTasks: 1
10/05/04 19:28:42 INFO mapred.MapTask: io.sort.mb = 100
10/05/04 19:28:42 INFO mapred.MapTask: data buffer = 79691776/99614720
10/05/04 19:28:42 INFO mapred.MapTask: record buffer = 262144/327680
10/05/04 19:28:42 INFO mapred.MapTask: Starting flush of map output
10/05/04 19:28:42 INFO mapred.TaskRunner: Task:attempt_local_0001_m_000003_0 isdone. And is in the process of commiting
10/05/04 19:28:42 INFO mapred.LocalJobRunner:file:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/input/mapred-site.xml:0+178
10/05/04 19:28:42 INFO mapred.TaskRunner: Task 'attempt_local_0001_m_000003_0'done.
10/05/04 19:28:42 INFO mapred.MapTask: numReduceTasks: 1
10/05/04 19:28:42 INFO mapred.MapTask: io.sort.mb = 100
10/05/04 19:28:43 INFO mapred.MapTask: data buffer = 79691776/99614720
10/05/04 19:28:43 INFO mapred.MapTask: record buffer = 262144/327680
10/05/04 19:28:43 INFO mapred.MapTask: Starting flush of map output
10/05/04 19:28:43 INFO mapred.TaskRunner: Task:attempt_local_0001_m_000004_0 isdone. And is in the process of commiting
10/05/04 19:28:43 INFO mapred.LocalJobRunner:file:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/input/hdfs-site.xml:0+178
10/05/04 19:28:43 INFO mapred.TaskRunner: Task 'attempt_local_0001_m_000004_0'done.
10/05/04 19:28:43 INFO mapred.LocalJobRunner:
10/05/04 19:28:43 INFO mapred.Merger: Merging 5 sorted segments
10/05/04 19:28:43 INFO mapred.Merger: Down to the last merge-pass, with 1segments left of total size: 21 bytes
10/05/04 19:28:43 INFO mapred.LocalJobRunner:
10/05/04 19:28:43 INFO mapred.TaskRunner: Task:attempt_local_0001_r_000000_0 isdone. And is in the process of commiting
10/05/04 19:28:43 INFO mapred.LocalJobRunner:
10/05/04 19:28:43 INFO mapred.TaskRunner: Task attempt_local_0001_r_000000_0 isallowed to commit now
10/05/04 19:28:43 INFO mapred.FileOutputCommitter: Saved output of task'attempt_local_0001_r_000000_0' tofile:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/grep-temp-1659003440
10/05/04 19:28:43 INFO mapred.LocalJobRunner: reduce > reduce
10/05/04 19:28:43 INFO mapred.TaskRunner: Task 'attempt_local_0001_r_000000_0'done.
10/05/04 19:28:44 INFO mapred.JobClient:  map 100% reduce 100%
10/05/04 19:28:44 INFO mapred.JobClient: Job complete: job_local_0001
10/05/04 19:28:44 INFO mapred.JobClient: Counters: 13
10/05/04 19:28:44 INFO mapred.JobClient:   FileSystemCounters
10/05/04 19:28:44 INFO mapred.JobClient:    FILE_BYTES_READ=974719
10/05/04 19:28:44 INFO mapred.JobClient:    FILE_BYTES_WRITTEN=1032450
10/05/04 19:28:44 INFO mapred.JobClient:   Map-Reduce Framework
10/05/04 19:28:44 INFO mapred.JobClient:     Reduce inputgroups=1
10/05/04 19:28:44 INFO mapred.JobClient:     Combine outputrecords=1
10/05/04 19:28:44 INFO mapred.JobClient:     Map inputrecords=219
10/05/04 19:28:44 INFO mapred.JobClient:     Reduce shufflebytes=0
10/05/04 19:28:44 INFO mapred.JobClient:     Reduce outputrecords=1
10/05/04 19:28:44 INFO mapred.JobClient:     SpilledRecords=2
10/05/04 19:28:44 INFO mapred.JobClient:     Map outputbytes=17
10/05/04 19:28:44 INFO mapred.JobClient:     Map inputbytes=8660
10/05/04 19:28:44 INFO mapred.JobClient:     Combine inputrecords=1
10/05/04 19:28:44 INFO mapred.JobClient:     Map outputrecords=1
10/05/04 19:28:44 INFO mapred.JobClient:     Reduce inputrecords=1
10/05/04 19:28:44 INFO jvm.JvmMetrics: Cannot initialize JVM Metrics withprocessName=JobTracker, sessionId= - already initialized
10/05/04 19:28:44 WARN mapred.JobClient: Use GenericOptionsParser for parsingthe arguments. Applications should implement Tool for the same.
10/05/04 19:28:44 INFO mapred.FileInputFormat: Total input paths to process : 1
10/05/04 19:28:45 INFO mapred.JobClient: Running job: job_local_0002
10/05/04 19:28:45 INFO mapred.FileInputFormat: Total input paths to process : 1
10/05/04 19:28:45 INFO mapred.MapTask: numReduceTasks: 1
10/05/04 19:28:45 INFO mapred.MapTask: io.sort.mb = 100
10/05/04 19:28:45 INFO mapred.MapTask: data buffer = 79691776/99614720
10/05/04 19:28:45 INFO mapred.MapTask: record buffer = 262144/327680
10/05/04 19:28:45 INFO mapred.MapTask: Starting flush of map output
10/05/04 19:28:45 INFO mapred.MapTask: Finished spill 0
10/05/04 19:28:45 INFO mapred.TaskRunner: Task:attempt_local_0002_m_000000_0 isdone. And is in the process of commiting
10/05/04 19:28:45 INFO mapred.LocalJobRunner: file:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/grep-temp-1659003440/part-00000:0+111
10/05/04 19:28:45 INFO mapred.TaskRunner: Task 'attempt_local_0002_m_000000_0'done.
10/05/04 19:28:45 INFO mapred.LocalJobRunner:
10/05/04 19:28:45 INFO mapred.Merger: Merging 1 sorted segments
10/05/04 19:28:45 INFO mapred.Merger: Down to the last merge-pass, with 1segments left of total size: 21 bytes
10/05/04 19:28:45 INFO mapred.LocalJobRunner:
10/05/04 19:28:46 INFO mapred.TaskRunner: Task:attempt_local_0002_r_000000_0 isdone. And is in the process of commiting
10/05/04 19:28:46 INFO mapred.LocalJobRunner:
10/05/04 19:28:46 INFO mapred.TaskRunner: Task attempt_local_0002_r_000000_0 isallowed to commit now
10/05/04 19:28:46 INFO mapred.FileOutputCommitter: Saved output of task 'attempt_local_0002_r_000000_0'to file:/home/sulliy/UESTC/Hadoop/Hadoop/hadoop-0.20.2/output
10/05/04 19:28:46 INFO mapred.LocalJobRunner: reduce > reduce
10/05/04 19:28:46 INFO mapred.TaskRunner: Task 'attempt_local_0002_r_000000_0'done.
10/05/04 19:28:46 INFO mapred.JobClient:  map 100% reduce 100%
10/05/04 19:28:46 INFO mapred.JobClient: Job complete: job_local_0002
10/05/04 19:28:46 INFO mapred.JobClient: Counters: 13
10/05/04 19:28:46 INFO mapred.JobClient:   FileSystemCounters
10/05/04 19:28:46 INFO mapred.JobClient:    FILE_BYTES_READ=640771
10/05/04 19:28:46 INFO mapred.JobClient:    FILE_BYTES_WRITTEN=685225
10/05/04 19:28:46 INFO mapred.JobClient:   Map-Reduce Framework
10/05/04 19:28:46 INFO mapred.JobClient:     Reduce inputgroups=1
10/05/04 19:28:46 INFO mapred.JobClient:     Combine outputrecords=0
10/05/04 19:28:46 INFO mapred.JobClient:     Map inputrecords=1
10/05/04 19:28:46 INFO mapred.JobClient:     Reduce shufflebytes=0
10/05/04 19:28:46 INFO mapred.JobClient:     Reduce outputrecords=1
10/05/04 19:28:46 INFO mapred.JobClient:     SpilledRecords=2
10/05/04 19:28:46 INFO mapred.JobClient:     Map outputbytes=17
10/05/04 19:28:46 INFO mapred.JobClient:     Map inputbytes=25
10/05/04 19:28:46 INFO mapred.JobClient:     Combine inputrecords=0
10/05/04 19:28:46 INFO mapred.JobClient:     Map outputrecords=1
10/05/04 19:28:46 INFO mapred.JobClient:     Reduce inputrecords=1

 

        结尾处给出了Map和Reduce的统计信息。

       使用命令查看结果:

       cat output/*

      

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值