hbase 使用disruptor_hbase安装时HRegionServer启动后自动退出

Thu Dec  3 14:01:45 CST 2015 Starting

regionserver on slave1

core file size  (blocks,

-c) 0

data seg size  (kbytes,

-d) unlimited

scheduling priority  (-e) 0

file size  (blocks, -f) unlimited

pending signals  (-i)

8192

max locked memory  (kbytes, -l) 32

max memory size  (kbytes, -m) unlimited

open files  (-n)

1024

pipe size  (512 bytes, -p) 8

POSIX message queues  (bytes, -q) 819200

real-time priority  (-r) 0

stack size  (kbytes, -s) 10240

cpu time  (seconds, -t) unlimited

max user processes  (-u) 8192

virtual memory  (kbytes,

-v) unlimited

file locks  (-x)

unlimited

2015-12-03 14:02:03,994 INFO  [main]

util.VersionInfo: HBase 1.0.2

2015-12-03 14:02:03,995 INFO  [main]

util.VersionInfo: Source code repository

git://HW10676.local/Users/enis/projects/git-repos/hbase

revision=76745a2cbffe08b812be16e0e19e637a23a923c5

2015-12-03 14:02:03,995 INFO  [main]

util.VersionInfo: Compiled by enis on Tue Aug 25 15:59:49 PDT

2015

2015-12-03 14:02:03,995 INFO  [main]

util.VersionInfo: From source with checksum

0bddb2fdbd66bdb14c0742f71c680f16

2015-12-03 14:02:06,699 INFO  [main]

util.ServerCommandLine:

env:JAVA_HOME=/home/hadoop/jdk1.7.0_67/

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine: env:SHLVL=3

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine:

env:HBASE_LOG_DIR=/home/hadoop/hbase-1.0.2/bin/../logs

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine:

env:HBASE_HOME=/home/hadoop/hbase-1.0.2/bin/..

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine: env:HBASE_OPTS=-XX:+UseConcMarkSweepGC

-Dhbase.log.dir=/home/hadoop/hbase-1.0.2/bin/../logs

-Dhbase.log.file=hbase-hadoop-regionserver-slave1.log

-Dhbase.home.dir=/home/hadoop/hbase-1.0.2/bin/..

-Dhbase.id.str=hadoop -Dhbase.root.logger=INFO,RFA

-Dhbase.security.logger=INFO,RFAS

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine: env:HBASE_ENV_INIT=true

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine:

env:XFILESEARCHPATH=/usr/dt/app-defaults/%L/Dt

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine: env:SSH_CLIENT=192.168.50.58 49929 22

2015-12-03 14:02:06,700 INFO  [main]

util.ServerCommandLine: env:MAIL=/var/mail/hadoop

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine:

env:HBASE_ZNODE_FILE=/tmp/hbase-hadoop-regionserver.znode

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:PWD=/home/hadoop/hbase-1.0.2

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:LOGNAME=hadoop

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:HBASE_MANAGES_ZK=false

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:G_BROKEN_FILENAMES=1

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:HBASE_NICENESS=0

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:HBASE_REST_OPTS=

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine:

env:NLSPATH=/usr/dt/lib/nls/msg/%L/%N.cat

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:MALLOC_ARENA_MAX=4

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:SSH_CONNECTION=192.168.50.58 49929

192.168.50.59 22

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:SHELL=/bin/bash

2015-12-03 14:02:06,710 INFO  [main]

util.ServerCommandLine: env:HBASE_ROOT_LOGGER=INFO,RFA

2015-12-03 14:02:06,711 INFO  [main]

util.ServerCommandLine:

env:CLASSPATH=/home/hadoop/hbase-1.0.2/conf:/home/hadoop/jdk1.7.0_67//lib/tools.jar:/home/hadoop/hbase-1.0.2/bin/..:/home/hadoop/hbase-1.0.2/bin/../lib/activation-1.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/aopalliance-1.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/apacheds-i18n-2.0.0-M15.jar:/home/hadoop/hbase-1.0.2/bin/../lib/apacheds-kerberos-codec-2.0.0-M15.jar:/home/hadoop/hbase-1.0.2/bin/../lib/api-asn1-api-1.0.0-M20.jar:/home/hadoop/hbase-1.0.2/bin/../lib/api-util-1.0.0-M20.jar:/home/hadoop/hbase-1.0.2/bin/../lib/asm-3.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/avro-1.7.4.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-beanutils-1.7.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-beanutils-core-1.8.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-cli-1.2.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-codec-1.9.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-collections-3.2.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-compress-1.4.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-configuration-1.6.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-daemon-1.0.13.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-digester-1.8.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-el-1.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-httpclient-3.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-io-2.4.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-lang-2.6.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-logging-1.2.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-math-2.2.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-math3-3.1.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/commons-net-3.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/disruptor-3.3.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/findbugs-annotations-1.3.9-1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/guava-12.0.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/guice-3.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/guice-servlet-3.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-annotations-2.6.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-auth-2.6.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-client-2.5.1.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-common-2.6.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-hdfs-2.6.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-mapreduce-client-app-2.6.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-mapreduce-client-common-2.6.0.jar:/home/hadoop/hbase-1.0.2/bin/../lib/hadoop-mapreduce-client-c

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值