搭建hadoop2.5.2/Ubuntu12集群(完全分布式)

一. 环境配置

 1. 用户(ts)获取root权限:vi /etc/passwd  (ts:x:1000:1000:crystal:/home/ts:/bin/bash   将两个1000改成0)

 2. 修改hostname:gedit /etc/hostname (分别为master(namenode),slave1,slave2(datanode))

 3. 设置hosts:gedit /ect/hosts

    

 4. 安装SSH:sudo apt-get install ssh

 5. 设置无密码登录:

     产生密钥

    ssh-keygen -t dsa -P '' -f ~/.ssh/id_dsa

    导入authorized_keys

    cat ~/.ssh/id_dsa.pub >> ~/.ssh/authorized_keys

    测试   ssh localhost

   

    把slave的id_dsa.pub,copy到master,并导入到authorized_keys

   

  6.  安装JDK

       解压: tar -zxvf  jdk-7u79-linux-x64.tar.gz 

       移动到安装目录:mv jdk /usr/lib/jvm/java-7-sun

       配置:gedit /etc/environment   添加--/usr/lib/jvm/java-7-sun/bin:

二.  安装Hadoop

  1.  创建文件            

        ~/dfs/name
        ~/dfs/data
        ~/tmp
  2. 解压文件
       tar zxvf hadoop-2.5.2.tar.gz
       mv hadoop /usr/lib/jvm/hadoop
           配置:gedit /etc/environment   添加--/usr/lib/jvm/hadoop/bin:/usr/lib/jvm/hadoop/sbin:

     

 3. 修改配置文件(目录:hadoop/etc/hadoop)

     1. hadoop-env.sh

         export JAVA_HOME=/usr/lib/jvm/java-7-sun

   2. yarn-env.sh

        export JAVA_HOME=/usr/lib/jvm/java-7-sun

    3. slaves

       

      4. core-site.xml       
 <configuration>
       <property>
                <name>fs.defaultFS</name>
                <value>hdfs://master:8020</value>
       </property>
       <property>
                <name>io.file.buffer.size</name>
                <value>131072</value>
        </property>
       <property>
               <name>hadoop.tmp.dir</name>
               <value>file:/home/ts/tmp</value>
               <description>Abase for other temporary   directories.</description>
       </property>
        <property>
               <name>hadoop.proxyuser.aboutyun.hosts</name>
               <value>*</value>
       </property>
       <property>
               <name>hadoop.proxyuser.aboutyun.groups</name>
               <value>*</value>
       </property>
</configuration>

     5. hdfs-site.xml     

<configuration>
       <property>
                <name>dfs.namenode.secondary.http-address</name>
               <value>master:9001</value>
       </property>
     <property>
             <name>dfs.namenode.name.dir</name>
             <value>file:/home/ts/dfs/name</value>
       </property>
      <property>
              <name>dfs.datanode.data.dir</name>
              <value>file:/home/ts/dfs/data</value>
       </property>
       <property>
               <name>dfs.replication</name>
               <value>2</value>
        </property>
        <property>
                 <name>dfs.webhdfs.enabled</name>
                  <value>true</value>
         </property>
</configuration>

      6. mapred-site.xml   
<configuration>
          <property>                                                                  <name>mapreduce.framework.name</name>
                <value>yarn</value>
           </property>
          <property>
                  <name>mapreduce.jobhistory.address</name>
                  <value>master:10020</value>
          </property>
          <property>
                <name>mapreduce.jobhistory.webapp.address</name>
                <value>master:19888</value>
       </property>
</configuration>

       7. yarn-site.xml   
<configuration>
        <property>
               <name>yarn.nodemanager.aux-services</name>
               <value>mapreduce_shuffle</value>
        </property>
        <property>                                                                
<name>yarn.nodemanager.aux-services.mapreduce.shuffle.class</name>
               <value>org.apache.hadoop.mapred.ShuffleHandler</value>
        </property>
        <property>
               <name>yarn.resourcemanager.address</name>
               <value>master:8032</value>
       </property>
       <property>
               <name>yarn.resourcemanager.scheduler.address</name>
               <value>master:8030</value>
       </property>
       <property>
            <name>yarn.resourcemanager.resource-tracker.address</name>
             <value>master:8031</value>
      </property>
      <property>
              <name>yarn.resourcemanager.admin.address</name>
               <value>master:8033</value>
       </property>
       <property>
               <name>yarn.resourcemanager.webapp.address</name>
               <value>master:8088</value>
       </property>
</configuration>

  4. copy到其它节点 /

     scp -r /usr/lib/jvm/hadoop root@slave1:/usr/lib/jvm/

    scp -r/usr/lib/jvm/hadoop root@slave2:/usr/lib/jvm/

 5.  测试

      格式化namenode:hadoop namenode format

      启动:start-all.sh

      查看:

     

    

    

      

       

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值