1 hadoop版本
2 安装机器
-
master:hadoop2-m1
-
slaver:hadoop2-s1
3 配置
-
core-site.xml
-
hdfs-site.xml
-
mapred-site.xml
-
yarn-site.xml
3.1 设置JAVA_HOME环境变量
1
|
export
JAVA_HOME=
/your/path/to/jdkdir
|
3.2 设置core-site.xml
1
2
3
4
5
6
7
8
9
10
|
<
configuration
>
<
property
>
<
name
>fs.defaultFS</
name
>
<
value
>hdfs://hadoop2-m1:8020</
value
>
</
property
>
<
property
>
<
name
>hadoop.tmp.dir</
name
>
<
value
>/home/tmp/hadoop2.0</
value
>
</
property
>
</
configuration
>
|
3.3 设置hdfs-site.xml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
|
<
configuration
>
<
property
>
<
name
>dfs.replication</
name
>
<
value
>1</
value
>
</
property
>
<
property
>
<
name
>dfs.namenode.name.dir</
name
>
<
value
>/home/dfs/name</
value
>
</
property
>
<
property
>
<
name
>dfs.datanode.data.dir</
name
>
<
value
>/home/dfs/data</
value
>
</
property
>
<
property
>
<
name
>dfs.permissions</
name
>
<
value
>false</
value
>
</
property
>
</
configuration
>
|
3.4 设置mapred-site.xml
1
2
3
4
5
6
|
<
configuration
>
<
property
>
<
name
>mapreduce.framework.name</
name
>
<
value
>yarn</
value
>
</
property
>
</
configuration
>
|
3.5 yarn-site.xml
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
|
<
configuration
>
<
property
>
<
name
>yarn.resourcemanager.address</
name
>
<
value
>hadoop2-m1:8032</
value
>
</
property
>
<
property
>
<
name
>yarn.resourcemanager.scheduler.address</
name
>
<
value
>hadoop2-m1:8030</
value
>
</
property
>
<
property
>
<
name
>yarn.resourcemanager.resource-tracker.address</
name
>
<
value
>hadoop2-m1:8031</
value
>
</
property
>
<
property
>
<
name
>yarn.resourcemanager.admin.address</
name
>
<
value
>hadoop2-m1:8033</
value
>
</
property
>
<
property
>
<
name
>yarn.resourcemanager.webapp.address</
name
>
<
value
>hadoop2-m1:8088</
value
>
</
property
>
<
property
>
<
name
>yarn.resourcemanager.scheduler.class</
name
>
<
value
>org.apache.hadoop.yarn.server.resourcemanager.scheduler.capacity.CapacityScheduler</
value
>
</
property
>
<
property
>
<
name
>yarn.nodemanager.aux-services</
name
>
<
value
>mapreduce_shuffle</
value
>
</
property
>
<
property
>
<
name
>yarn.nodemanager.aux-services.mapreduce.shuffle.class</
name
>
<
value
>org.apache.hadoop.mapred.ShuffleHandler</
value
>
</
property
>
</
configuration
>
|
4 启动
4.1 格式化
1
2
|
cd
$HADOOP_HOME
bin
/hdfs
namenode -
format
|
4.2 在hadoop2-m1中启动 namenode 和 resourcemanager
1
2
3
4
5
|
sbin
/hadoop-daemon
.sh start namenode
sbin
/yarn-daemon
.sh start resourcemanager
#停止服务的命令
sbin
/hadoop-daemon
.sh stop namenode
sbin
/yarn-daemon
.sh stop resourcemanager
|
4.3 在hadoop2-s1中启动 datanode 和 nodemanager
1
2
3
4
5
|
sbin
/hadoop-daemon
.sh start datanode
sbin
/yarn-daemon
.sh start nodemanager
#停止服务的命令
sbin
/hadoop-daemon
.sh stop datanode
sbin
/yarn-daemon
.sh stop nodemanager
|
4.4 在hadoop2-m1中启动 proxyserver 和 historyserver
1
2
3
4
5
|
sbin
/yarn-daemon
.sh start proxyserver
sbin
/mr-jobhistory-daemon
.sh start historyserver
#停止服务的命令
sbin
/yarn-daemon
.sh stop proxyserver
sbin
/mr-jobhistory-daemon
.sh stop historyserver
|
5 检查和测试
5.1 检查以下两个页面是否能打开
5.2 运行以下hdfs命令
1
2
3
4
5
|
bin
/hdfs
dfs -
mkdir
/test/input1
bin
/hdfs
dfs -put NOTICE.txt
/test/input1/
bin
/hdfs
dfs -put README.txt
/test/input1/
bin
/hdfs
dfs -get
/test/input1/NOTICE
.txt
/tmp/
bin
/hdfs
dfs -
cat
/test/input1/NOTICE
.txt
|
5.3 运行mapreduce job
1
|
bin
/hadoop
jar share
/hadoop/mapreduce/hadoop-mapreduce-examples-2
.2.0.jar
grep
/test/input1
/test/output1
'code'
|