[b]1. download hadoop from apache.org[/b]
[url]http://archive.apache.org/dist/hadoop/core/[/url]
选一个稳定的版本下载,例如:hadoop-2.7.1.tar.gz
[b]2. 把hadoop上传到linux[/b]
makdir /host01
还是FTP到/host01
[b]3. 解压到/host01[/b]
tar -zxvf /host01/hadoop-2.7.1.tar.gz -C /host01
[b]4. 修改hadoop2.x的配置文件$HADOOP_HOME/etc/hadoop[/b]
1)hadoop-env.sh
vim hadoop-env.sh
#设置JAVA_HOME
[quote]export JAVA_HOME=/usr/java/jdk1.6.0_45[/quote]
2)core-site.xml
[quote]<!-- config HDFS namenode address -->
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost.localdomain:9000</value>
</property>
<!-- cinfig hadoop runtime directory -->
<property>
<name>hadoop.tmp.dir</name>
<value>/host01/hadoop-2.7.1/tmp</value>
</property>[/quote]
3) hdfs-site.xml
[quote] <!-- config HDFS backup count as 1 -->
<property>
<name>dfs.replication</name>
<value>1</value>
</property>[/quote]
4) mapred-site.xml
mv mapred-site.xml.template mapred-site.xml
vim mapred-site.xml
[quote]<!-- config mapreduce running on yarn -->
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>[/quote]
5) yarn-site.xml
[quote]<!-- configure the boss of yarn (ResourceManager) address -->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>localhost.localdomain</value>
</property>
<!-- the way reducer get the data -->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>[/quote]
[b]5. 添加java,hadoop到环境变量 [/b]
vim /etc/proflie
[quote]export JAVA_HOME=/usr/java/jdk1.6.0_45
export HADOOP_HOME=/host01/hadoop-2.7.1
export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin[/quote]
source /etc/profile
[b]6.启动hadoop[/b]
1)格式化namenode
hdfs namenode -format
INFO common.Storage: Storage directory /host01/hadoop-2.7.1/tmp/dfs/name has been successfully formatted.
2)启动
先启动HDFS
sbin/start-dfs.sh
再启动YARN
sbin/start-yarn.sh
3) jps验证
11836 SecondaryNameNode
11598 DataNode
12229 NodeManager
12533 Jps
11470 NameNode
12118 ResourceManager
http://hostip:50070 (HDFS管理界面)
[url]http://archive.apache.org/dist/hadoop/core/[/url]
选一个稳定的版本下载,例如:hadoop-2.7.1.tar.gz
[b]2. 把hadoop上传到linux[/b]
makdir /host01
还是FTP到/host01
[b]3. 解压到/host01[/b]
tar -zxvf /host01/hadoop-2.7.1.tar.gz -C /host01
[b]4. 修改hadoop2.x的配置文件$HADOOP_HOME/etc/hadoop[/b]
1)hadoop-env.sh
vim hadoop-env.sh
#设置JAVA_HOME
[quote]export JAVA_HOME=/usr/java/jdk1.6.0_45[/quote]
2)core-site.xml
[quote]<!-- config HDFS namenode address -->
<property>
<name>fs.defaultFS</name>
<value>hdfs://localhost.localdomain:9000</value>
</property>
<!-- cinfig hadoop runtime directory -->
<property>
<name>hadoop.tmp.dir</name>
<value>/host01/hadoop-2.7.1/tmp</value>
</property>[/quote]
3) hdfs-site.xml
[quote] <!-- config HDFS backup count as 1 -->
<property>
<name>dfs.replication</name>
<value>1</value>
</property>[/quote]
4) mapred-site.xml
mv mapred-site.xml.template mapred-site.xml
vim mapred-site.xml
[quote]<!-- config mapreduce running on yarn -->
<property>
<name>mapreduce.framework.name</name>
<value>yarn</value>
</property>[/quote]
5) yarn-site.xml
[quote]<!-- configure the boss of yarn (ResourceManager) address -->
<property>
<name>yarn.resourcemanager.hostname</name>
<value>localhost.localdomain</value>
</property>
<!-- the way reducer get the data -->
<property>
<name>yarn.nodemanager.aux-services</name>
<value>mapreduce_shuffle</value>
</property>[/quote]
[b]5. 添加java,hadoop到环境变量 [/b]
vim /etc/proflie
[quote]export JAVA_HOME=/usr/java/jdk1.6.0_45
export HADOOP_HOME=/host01/hadoop-2.7.1
export PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin[/quote]
source /etc/profile
[b]6.启动hadoop[/b]
1)格式化namenode
hdfs namenode -format
INFO common.Storage: Storage directory /host01/hadoop-2.7.1/tmp/dfs/name has been successfully formatted.
2)启动
先启动HDFS
sbin/start-dfs.sh
再启动YARN
sbin/start-yarn.sh
3) jps验证
11836 SecondaryNameNode
11598 DataNode
12229 NodeManager
12533 Jps
11470 NameNode
12118 ResourceManager
http://hostip:50070 (HDFS管理界面)