1. Zookeeper Dump
访问HBase的web页面:http://192.168.26.140:16030/zk.jsp
HBase is rooted at /hbase
Active master address: hadoop.master,16020,1428562242762
Backup master addresses:
Region server holding hbase:meta: hadoop.master,16201,1428562244095
Region servers:
hadoop.master,16201,1428562244095
/hbase/replication:
/hbase/replication/peers:
/hbase/replication/rs:
/hbase/replication/rs/hadoop.master,16201,1428562244095:
Quorum Server Statistics:
localhost:2181
Zookeeper version: 3.4.6-1569965, built on 02/20/2014 09:09 GMT
Clients:
/127.0.0.1:47168[1](queued=0,recved=2916,sent=2920)
/127.0.0.1:47251[1](queued=0,recved=2752,sent=2752)
/0:0:0:0:0:0:0:1:46915[1](queued=0,recved=2753,sent=2753)
/0:0:0:0:0:0:0:1:46907[1](queued=0,recved=2763,sent=2763)
/0:0:0:0:0:0:0:1:47060[1](queued=0,recved=2717,sent=2717)
/127.0.0.1:47173[1](queued=0,recved=2754,sent=2754)
/127.0.0.1:47174[1](queued=0,recved=3310,sent=3310)
/127.0.0.1:47167[1](queued=0,recved=7394,sent=7465)
/127.0.0.1:54521[1](queued=0,recved=33,sent=33)
/127.0.0.1:47172[1](queued=0,recved=2753,sent=2753)
/127.0.0.1:54583[0](queued=0,recved=1,sent=0)
Latency min/avg/max: 0/0/790
Received: 30227
Sent: 30301
Connections: 11
Outstanding: 0
Zxid: 0xe5
Mode: standalone
Node count: 45
1.包含的信息:HBase在Zookeeper中的根目录:/hbase
2.Master地址:hadoop.master,16020,1428562242762(域名,端口和启动的时间戳)
3. meta表所在的Region Server:hadoop.master,16201,1428562244095(域名,端口和启动的时间戳)
4. HBase集群包含的Region Servers:这里是一个Standalone集群,因此只有一个:hadoop.master,16201,1428562244095(同时也是meta表所在的Region Server)
2. meta表包含的信息
hbase命令行上执行如下命令,打印meta表的所有的数据
hbase(main):021:0> scan 'hbase:meta'
ROW COLUMN+CELL
hbase:namespace,,1426129053122.2e6 column=info:regioninfo, timestamp=1426129054266, value={ENCODED => 2e6738351085a34e0b83933c9dbe98b6, NA
738351085a34e0b83933c9dbe98b6. ME => 'hbase:namespace,,1426129053122.2e6738351085a34e0b83933c9dbe98b6.', STARTKEY => '', ENDKEY => ''}
hbase:namespace,,1426129053122.2e6 column=info:seqnumDuringOpen, timestamp=1428562261802, value=\x00\x00\x00\x00\x00\x00\x00\x0C
738351085a34e0b83933c9dbe98b6.
hbase:namespace,,1426129053122.2e6 column=info:server, timestamp=1428562261802, value=hadoop.master:16201
738351085a34e0b83933c9dbe98b6.
hbase:namespace,,1426129053122.2e6 column=info:serverstartcode, timestamp=1428562261802, value=1428562244095
738351085a34e0b83933c9dbe98b6.
//手工添加注释:行table3,,1428632308630.bb738e209c3b4d5833bb123549b324c3. 列族info 列:regioninfo, seqnumDuringOpen,server,serverstartcode
//关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围
//table3的第一个Region(0-300)
table3,,1428632308630.bb738e209c3b column=info:regioninfo, timestamp=1428632310195, value={ENCODED => bb738e209c3b4d5833bb123549b324c3, NA
4d5833bb123549b324c3. ME => 'table3,,1428632308630.bb738e209c3b4d5833bb123549b324c3.', STARTKEY => '', ENDKEY => 'row-300'}
table3,,1428632308630.bb738e209c3b column=info:seqnumDuringOpen, timestamp=1428632313497, value=\x00\x00\x00\x00\x00\x00\x00\x02
4d5833bb123549b324c3.
table3,,1428632308630.bb738e209c3b column=info:server, timestamp=1428632313497, value=hadoop.master:16201
4d5833bb123549b324c3.
table3,,1428632308630.bb738e209c3b column=info:serverstartcode, timestamp=1428632313497, value=1428562244095
4d5833bb123549b324c3.
//关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围
//table3的第二个Region(300-500)
table3,row-300,1428632308630.2cf83 column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 2cf837baf289fb122c82d77214562c0a, NA
7baf289fb122c82d77214562c0a. ME => 'table3,row-300,1428632308630.2cf837baf289fb122c82d77214562c0a.', STARTKEY => 'row-300', ENDKEY =
> 'row-500'}
table3,row-300,1428632308630.2cf83 column=info:seqnumDuringOpen, timestamp=1428632313425, value=\x00\x00\x00\x00\x00\x00\x00\x02
7baf289fb122c82d77214562c0a.
table3,row-300,1428632308630.2cf83 column=info:server, timestamp=1428632313425, value=hadoop.master:16201
7baf289fb122c82d77214562c0a.
table3,row-300,1428632308630.2cf83 column=info:serverstartcode, timestamp=1428632313425, value=1428562244095
7baf289fb122c82d77214562c0a.
//关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围
//table3的第三个Region(500-700)
table3,row-500,1428632308630.1fa2e column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 1fa2e49c7404d3cd39afc39a99cc1c26, NA
49c7404d3cd39afc39a99cc1c26. ME => 'table3,row-500,1428632308630.1fa2e49c7404d3cd39afc39a99cc1c26.', STARTKEY => 'row-500', ENDKEY =
> 'row-700'}
table3,row-500,1428632308630.1fa2e column=info:seqnumDuringOpen, timestamp=1428632312856, value=\x00\x00\x00\x00\x00\x00\x00\x02
49c7404d3cd39afc39a99cc1c26.
table3,row-500,1428632308630.1fa2e column=info:server, timestamp=1428632312856, value=hadoop.master:16201
49c7404d3cd39afc39a99cc1c26.
table3,row-500,1428632308630.1fa2e column=info:serverstartcode, timestamp=1428632312856, value=1428562244095
49c7404d3cd39afc39a99cc1c26.
//关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围
//table3的第四个Region(700-900)
table3,row-700,1428632308630.1d3d6 column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 1d3d6e64741683e551e5b8d0722d7ff9, NA
e64741683e551e5b8d0722d7ff9. ME => 'table3,row-700,1428632308630.1d3d6e64741683e551e5b8d0722d7ff9.', STARTKEY => 'row-700', ENDKEY =
> 'row-900'}
table3,row-700,1428632308630.1d3d6 column=info:seqnumDuringOpen, timestamp=1428632312817, value=\x00\x00\x00\x00\x00\x00\x00\x02
e64741683e551e5b8d0722d7ff9.
table3,row-700,1428632308630.1d3d6 column=info:server, timestamp=1428632312817, value=hadoop.master:16201
e64741683e551e5b8d0722d7ff9.
table3,row-700,1428632308630.1d3d6 column=info:serverstartcode, timestamp=1428632312817, value=1428562244095
e64741683e551e5b8d0722d7ff9.
//关注regioninfo列,STARTKEY和ENDKEY,表示该Region的起止RowKey范围
//table3的第五个Region(900-正无穷)
table3,row-900,1428632308630.4e048 column=info:regioninfo, timestamp=1428632310195, value={ENCODED => 4e048f1c6a7a9a58994629ae957d4ddf, NA
f1c6a7a9a58994629ae957d4ddf. ME => 'table3,row-900,1428632308630.4e048f1c6a7a9a58994629ae957d4ddf.', STARTKEY => 'row-900', ENDKEY =
> ''}
table3,row-900,1428632308630.4e048 column=info:seqnumDuringOpen, timestamp=1428632312795, value=\x00\x00\x00\x00\x00\x00\x00\x02
f1c6a7a9a58994629ae957d4ddf.
table3,row-900,1428632308630.4e048 column=info:server, timestamp=1428632312795, value=hadoop.master:16201
f1c6a7a9a58994629ae957d4ddf.
table3,row-900,1428632308630.4e048 column=info:serverstartcode, timestamp=1428632312795, value=1428562244095
f1c6a7a9a58994629ae957d4ddf.
3. -ROOT-表、meta表以及Region Server、Region的关系
4. Hadoop生态系统与HBase的集成
- MapReduce为HBase提供了高性能的计算能力
- Zookeeper为HBase提供了稳定服务和failover机制。
- Pig和Hive还为HBase提供了高层语言支持,使得在HBase上进行数据统计处理变的非常简单;
- Sqoop则为HBase提供了方便的RDBMS数据导入功能,使得传统数据库数据向HBase中迁移变的非常方便。