Hive的Shell里hive> 执行操作时,出现FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec....

 

  

  不多说,直接上干货!

 

    

  这个问题,得非 你的hive和hbase是不是同样都是CDH版本,还是一个是apache版本,一个是CDH版本。

 

 

 

   问题详情

[kfk@bigdata-pro01 apache-hive-1.0.0-bin]$ bin/hive

Logging initialized using configuration in file:/opt/modules/apache-hive-1.0.0-bin/conf/hive-log4j.properties
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/modules/hadoop-2.6.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/modules/apache-hive-1.0.0-bin/lib/hive-jdbc-1.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
hive> create external table weblogs(id string,datatime string,userid string,searchname string,retorder string,cliorder string,cliurl string)  STORED BY  'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES("hbase.columns.mapping" = ":key,info:datatime,info:userid,info:searchname,info:retorder,info:cliorder,info:cliurl") TBLPROPERTIES("hbase.table.name" = "weblogs");
FAILED: Execution Error, return code 1 from org.apache.hadoop.hive.ql.exec.DDLTask. MetaException(message:MetaException(message:HBase table weblogs doesn't exist while the table is declared as an external table.)
    at org.apache.hadoop.hive.hbase.HBaseStorageHandler.preCreateTable(HBaseStorageHandler.java:202)
    at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:608)
    at org.apache.hadoop.hive.metastore.HiveMetaStoreClient.createTable(HiveMetaStoreClient.java:601)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:606)
    at org.apache.hadoop.hive.metastore.RetryingMetaStoreClient.invoke(RetryingMetaStoreClient.java:90)
    at com.sun.proxy.$Proxy7.createTable(Unknown Source)
    at org.apache.hadoop.hive.ql.metadata.Hive.createTable(Hive.java:671)
    at org.apache.hadoop.hive.ql.exec.DDLTask.createTable(DDLTask.java:3973)
    at org.apache.hadoop.hive.ql.exec.DDLTask.execute(DDLTask.java:295)
    at org.apache.hadoop.hive.ql.exec.Task.executeTask(Task.java:160)
    at org.apache.hadoop.hive.ql.exec.TaskRunner.runSequential(TaskRunner.java:85)
    at org.apache.hadoop.hive.ql.Driver.launchTask(Driver.java:1604)
    at org.apache.hadoop.hive.ql.Driver.execute(Driver.java:1364)
    at org.apache.hadoop.hive.ql.Driver.runInternal(Driver.java:1177)
    at org.apache.hadoop.hive.ql.Driver.run(Driver.java:1004)
    at org.apache.hadoop.hive.ql.Driver.run(Driver.java:994)
    at org.apache.hadoop.hive.cli.CliDriver.processLocalCmd(CliDriver.java:201)
    at org.apache.hadoop.hive.cli.CliDriver.processCmd(CliDriver.java:153)
    at org.apache.hadoop.hive.cli.CliDriver.processLine(CliDriver.java:364)
    at org.apache.hadoop.hive.cli.CliDriver.executeDriver(CliDriver.java:712)
    at org.apache.hadoop.hive.cli.CliDriver.run(CliDriver.java:631)
    at org.apache.hadoop.hive.cli.CliDriver.main(CliDriver.java:570)
    at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
    at sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:57)
    at sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
    at java.lang.reflect.Method.invoke(Method.java:606)
    at org.apache.hadoop.util.RunJar.run(RunJar.java:221)
    at org.apache.hadoop.util.RunJar.main(RunJar.java:136)
)

 

 

 

 

    说明没创建成功。

 

 

 

 

 

 Hive和HBase同是CDH版本解决办法

 

 

 

  首先通过用下面的命令,重新启动hive
  ./hive -hiveconf hive.root.logger=DEBUG,console 进行debug
  查看到错误原因

 

 

 

 

  调遣hbase包错误,配置文件不能加载。

  将hbase目录下jar包拷贝到hadoop/lib下,(注意了我这里为了保险起见3个节点都做了这一步)

[kfk@bigdata-pro01 lib]$ cp  hbase-client-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-common-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-common-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-examples-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar /opt/modules/h
hadoop-2.6.0/          hbase-0.98.6-cdh5.3.0/ hive-0.13.1-cdh5.3.0/  
[kfk@bigdata-pro01 lib]$ cp hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-hadoop2-compat-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-hadoop-compat-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-hadoop-compat-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-it-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-it-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-prefix-tree-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-protocol-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-server-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-server-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-shell-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp  hbase-testing-util-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ cp hbase-thrift-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro01 lib]$ pwd
/opt/modules/hbase-0.98.6-cdh5.3.0/lib
[kfk@bigdata-pro01 lib]$ 

 

 

 

 

 

 

 

 

[kfk@bigdata-pro02 lib]$ cp  hbase-client-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-common-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-common-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-examples-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-hadoop2-compat-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-hadoop-compat-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-hadoop-compat-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-it-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-it-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-prefix-tree-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-protocol-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-server-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-server-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-shell-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ 
[kfk@bigdata-pro02 lib]$ cp  hbase-testing-util-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp hbase-thrift-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ cp  hbase-testing-util-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro02 lib]$ pwd
/opt/modules/hbase-0.98.6-cdh5.3.0/lib
[kfk@bigdata-pro02 lib]$ 

 

 

 

 

 

 

 

[kfk@bigdata-pro03 ~]$ cd /opt/modules/hbase-0.98.6-cdh5.3.0/lib/
[kfk@bigdata-pro03 lib]$ cp  hbase-client-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-common-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-common-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-examples-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-hadoop2-compat-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-hadoop-compat-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-hadoop-compat-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-it-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-it-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-prefix-tree-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-protocol-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-server-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-server-0.98.6-cdh5.3.0-tests.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ cp hbase-shell-0.98.6-cdh5.3.0.jar /opt/modules/hadoop-2.6.0/lib
[kfk@bigdata-pro03 lib]$ pwd
/opt/modules/hbase-0.98.6-cdh5.3.0/lib
[kfk@bigdata-pro03 lib]$ 

 

 

 

  然后,修改hive的配置文件,使hbase的lib包生效

 

 

 

 

 

   重启一下hive,再次建表

[kfk@bigdata-pro01 hive-0.13.1-cdh5.3.0]$ bin/hive

Logging initialized using configuration in file:/opt/modules/hive-0.13.1-cdh5.3.0/conf/hive-log4j.properties
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/modules/hadoop-2.6.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/modules/hbase-0.98.6-cdh5.3.0/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
hive (default)> create external table weblogs(
              > id string,
              > datatime string,
              > userid string,
              > searchname string,
              > retorder string,
              > cliorder string,
              > cliurl string
              > )
              > STORED BY  'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
              > WITH SERDEPROPERTIES("hbase.columns.mapping" = ":key,info:datatime,info:userid,info:searchname,info:retorder,info:cliorder,info:cliurl") 
              > TBLPROPERTIES("hbase.table.name" = "weblogs");
OK
Time taken: 2.084 seconds
hive (default)> 

 

 

 

 

 

 

 

 

 Hive和HBase不同是CDH版本解决办法

  (1) 退出,重新进hive

     (2) 重启Hbase服务试试

   (3)别光检查你的Hive,还看看你的HBase进程啊。

    从时间同步,防火墙,MySQL启动没、hive-site.xml配置好了没,以及授权是否正确等方面去排查。

java.lang.RuntimeException: HRegionServer Aborted的问题

 

分布式集群HBase启动后某节点的HRegionServer自动消失问题

 

 

     (4) 少包了,是Hive和HBase集成这过程出错了。

   官网 https://cwiki.apache.org/confluence/display/Hive/HBaseIntegration

 

 

 

 

 

 

 

   

   一闪一闪就是软连接有错误,资源找不到,

 

   

 

 

 

export HBASE_HOME=/opt/modules/hbase-0.98.6-cdh5.3.0
export HIVE_HOME=/opt/modules/apache-hive-1.0.0-bin



ln -s $HBASE_HOME/lib/hbase-server-0.98.6-cdh5.3.0.jar $HIVE_HOME/lib/hbase-server-0.98.6-cdh5.3.0.jar
ln -s $HBASE_HOME/lib/hbase-client-0.98.6-cdh5.3.0.jar $HIVE_HOME/lib/hbase-client-0.98.6-cdh5.3.0.jar
ln -s $HBASE_HOME/lib/hbase-protocol-0.98.6-cdh5.3.0.jar $HIVE_HOME/lib/hbase-protocol-0.98.6-cdh5.3.0.jar 
ln -s $HBASE_HOME/lib/hbase-it-0.98.6-cdh5.3.0.jar $HIVE_HOME/lib/hbase-it-0.98.6-cdh5.3.0.jar 
ln -s $HBASE_HOME/lib/htrace-core-2.04.jar $HIVE_HOME/lib/htrace-core-2.04.jar
ln -s $HBASE_HOME/lib/hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar $HIVE_HOME/lib/hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar 
ln -s $HBASE_HOME/lib/hbase-hadoop-compat-0.98.6-cdh5.3.0.jar $HIVE_HOME/lib/hbase-hadoop-compat-0.98.6-cdh5.3.0.jar 
ln -s $HBASE_HOME/lib/high-scale-lib-1.1.1.jar $HIVE_HOME/lib/high-scale-lib-1.1.1.jar 
ln -s $HBASE_HOME/lib/hbase-common-0.98.6-cdh5.3.0.jar $HIVE_HOME/lib/hbase-common-0.98.6-cdh5.3.0.jar 

 

 

 

 

 

 

[kfk@bigdata-pro01 lib]$ ll
total 82368
-rw-rw-r--  1 kfk kfk  4368200 Jan  7 17:35 accumulo-core-1.6.0.jar
-rw-rw-r--  1 kfk kfk   102069 Jan  7 17:35 accumulo-fate-1.6.0.jar
-rw-rw-r--  1 kfk kfk    57420 Jan  7 17:35 accumulo-start-1.6.0.jar
-rw-rw-r--  1 kfk kfk   117409 Jan  7 17:35 accumulo-trace-1.6.0.jar
-rw-rw-r--  1 kfk kfk    62983 Jan  7 17:35 activation-1.1.jar
-rw-rw-r--  1 kfk kfk  1997485 Jan  7 17:35 ant-1.9.1.jar
-rw-rw-r--  1 kfk kfk    18336 Jan  7 17:35 ant-launcher-1.9.1.jar
-rw-rw-r--  1 kfk kfk   445288 Jan  7 17:35 antlr-2.7.7.jar
-rw-rw-r--  1 kfk kfk   164368 Jan  7 17:35 antlr-runtime-3.4.jar
-rw-rw-r--  1 kfk kfk    32693 Jan  7 17:35 asm-commons-3.1.jar
-rw-rw-r--  1 kfk kfk    21879 Jan  7 17:35 asm-tree-3.1.jar
-rw-rw-r--  1 kfk kfk   400680 Jan  7 17:35 avro-1.7.5.jar
-rw-rw-r--  1 kfk kfk   110600 Jan  7 17:35 bonecp-0.8.0.RELEASE.jar
-rw-rw-r--  1 kfk kfk    89044 Jan  7 17:35 calcite-avatica-0.9.2-incubating.jar
-rw-rw-r--  1 kfk kfk  3156089 Jan  7 17:35 calcite-core-0.9.2-incubating.jar
-rw-rw-r--  1 kfk kfk   188671 Jan  7 17:35 commons-beanutils-1.7.0.jar
-rw-rw-r--  1 kfk kfk   206035 Jan  7 17:35 commons-beanutils-core-1.8.0.jar
-rw-rw-r--  1 kfk kfk    41123 Jan  7 17:35 commons-cli-1.2.jar
-rw-rw-r--  1 kfk kfk    58160 Jan  7 17:35 commons-codec-1.4.jar
-rw-rw-r--  1 kfk kfk   575389 Jan  7 17:35 commons-collections-3.2.1.jar
-rw-rw-r--  1 kfk kfk    30595 Jan  7 17:35 commons-compiler-2.7.6.jar
-rw-rw-r--  1 kfk kfk   241367 Jan  7 17:35 commons-compress-1.4.1.jar
-rw-rw-r--  1 kfk kfk   298829 Jan  7 17:35 commons-configuration-1.6.jar
-rw-rw-r--  1 kfk kfk   160519 Jan  7 17:35 commons-dbcp-1.4.jar
-rw-rw-r--  1 kfk kfk   143602 Jan  7 17:35 commons-digester-1.8.jar
-rw-rw-r--  1 kfk kfk   279781 Jan  7 17:35 commons-httpclient-3.0.1.jar
-rw-rw-r--  1 kfk kfk   185140 Jan  7 17:35 commons-io-2.4.jar
-rw-rw-r--  1 kfk kfk   284220 Jan  7 17:35 commons-lang-2.6.jar
-rw-rw-r--  1 kfk kfk    62050 Jan  7 17:35 commons-logging-1.1.3.jar
-rw-rw-r--  1 kfk kfk   832410 Jan  7 17:35 commons-math-2.1.jar
-rw-rw-r--  1 kfk kfk    96221 Jan  7 17:35 commons-pool-1.5.4.jar
-rw-rw-r--  1 kfk kfk   415578 Jan  7 17:35 commons-vfs2-2.0.jar
-rw-rw-r--  1 kfk kfk    68866 Jan  7 17:35 curator-client-2.6.0.jar
-rw-rw-r--  1 kfk kfk   185245 Jan  7 17:35 curator-framework-2.6.0.jar
-rw-rw-r--  1 kfk kfk   339666 Jan  7 17:35 datanucleus-api-jdo-3.2.6.jar
-rw-rw-r--  1 kfk kfk  1890075 Jan  7 17:35 datanucleus-core-3.2.10.jar
-rw-rw-r--  1 kfk kfk  1809447 Jan  7 17:35 datanucleus-rdbms-3.2.9.jar
-rw-rw-r--  1 kfk kfk  2831358 Jan  7 17:35 derby-10.10.1.1.jar
-rw-rw-r--  1 kfk kfk    18402 Jan  7 17:35 eigenbase-properties-1.1.4.jar
-rw-rw-r--  1 kfk kfk    12452 Jan  7 17:35 geronimo-annotation_1.0_spec-1.1.1.jar
-rw-rw-r--  1 kfk kfk    30548 Jan  7 17:35 geronimo-jaspic_1.0_spec-1.0.jar
-rw-rw-r--  1 kfk kfk    16030 Jan  7 17:35 geronimo-jta_1.1_spec-1.1.1.jar
-rw-rw-r--  1 kfk kfk  6377448 Jan  7 17:35 groovy-all-2.1.6.jar
-rw-rw-r--  1 kfk kfk  1648200 Jan  7 17:35 guava-11.0.2.jar
-rw-rw-r--  1 kfk kfk    76643 Jan  7 17:35 hamcrest-core-1.1.jar
lrwxrwxrwx  1 kfk kfk       71 Jan  7 17:41 hbase-client-0.98.6-cdh5.3.0.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/hbase-client-0.98.6-cdh5.3.0.jar
lrwxrwxrwx  1 kfk kfk       71 Jan  7 17:41 hbase-common-0.98.6-cdh5.3.0.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/hbase-common-0.98.6-cdh5.3.0.jar
lrwxrwxrwx  1 kfk kfk       79 Jan  7 17:41 hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/hbase-hadoop2-compat-0.98.6-cdh5.3.0.jar
lrwxrwxrwx  1 kfk kfk       78 Jan  7 17:41 hbase-hadoop-compat-0.98.6-cdh5.3.0.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/hbase-hadoop-compat-0.98.6-cdh5.3.0.jar
lrwxrwxrwx  1 kfk kfk       67 Jan  7 17:41 hbase-it-0.98.6-cdh5.3.0.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/hbase-it-0.98.6-cdh5.3.0.jar
lrwxrwxrwx  1 kfk kfk       73 Jan  7 17:41 hbase-protocol-0.98.6-cdh5.3.0.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/hbase-protocol-0.98.6-cdh5.3.0.jar
lrwxrwxrwx  1 kfk kfk       71 Jan  7 17:41 hbase-server-0.98.6-cdh5.3.0.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/hbase-server-0.98.6-cdh5.3.0.jar
lrwxrwxrwx  1 kfk kfk       63 Jan  7 17:41 high-scale-lib-1.1.1.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/high-scale-lib-1.1.1.jar
-rw-rw-r--  1 kfk kfk   112948 Jan  7 17:35 hive-accumulo-handler-1.0.0.jar
-rw-rw-r--  1 kfk kfk    42337 Jan  7 17:35 hive-ant-1.0.0.jar
-rw-rw-r--  1 kfk kfk   128193 Jan  7 17:35 hive-beeline-1.0.0.jar
-rw-rw-r--  1 kfk kfk    35030 Jan  7 17:35 hive-cli-1.0.0.jar
-rw-rw-r--  1 kfk kfk   234200 Jan  7 17:35 hive-common-1.0.0.jar
-rw-rw-r--  1 kfk kfk   120929 Jan  7 17:35 hive-contrib-1.0.0.jar
-rw-rw-r--  1 kfk kfk 17429823 Jan  7 17:35 hive-exec-1.0.0.jar
-rw-rw-r--  1 kfk kfk   107720 Jan  7 17:35 hive-hbase-handler-1.0.0.jar
-rw-rw-r--  1 kfk kfk    27378 Jan  7 17:35 hive-hwi-1.0.0.jar
-rw-rw-r--  1 kfk kfk    91218 Jan  7 17:35 hive-jdbc-1.0.0.jar
-rw-rw-r--  1 kfk kfk 15002397 Jan  7 17:35 hive-jdbc-1.0.0-standalone.jar
-rw-rw-r--  1 kfk kfk  5125725 Jan  7 17:35 hive-metastore-1.0.0.jar
-rw-rw-r--  1 kfk kfk   813184 Jan  7 17:35 hive-serde-1.0.0.jar
-rw-rw-r--  1 kfk kfk  1866002 Jan  7 17:35 hive-service-1.0.0.jar
-rw-rw-r--  1 kfk kfk    32386 Jan  7 17:35 hive-shims-0.20-1.0.0.jar
-rw-rw-r--  1 kfk kfk    29032 Jan  7 17:35 hive-shims-0.20S-1.0.0.jar
-rw-rw-r--  1 kfk kfk    48341 Jan  7 17:35 hive-shims-0.23-1.0.0.jar
-rw-rw-r--  1 kfk kfk     7467 Jan  7 17:35 hive-shims-1.0.0.jar
-rw-rw-r--  1 kfk kfk    48305 Jan  7 17:35 hive-shims-common-1.0.0.jar
-rw-rw-r--  1 kfk kfk    79215 Jan  7 17:35 hive-shims-common-secure-1.0.0.jar
-rw-rw-r--  1 kfk kfk    14525 Jan  7 17:35 hive-testutils-1.0.0.jar
lrwxrwxrwx  1 kfk kfk       59 Jan  7 17:41 htrace-core-2.04.jar -> /opt/modules/hbase-0.98.6-cdh5.3.0/lib/htrace-core-2.04.jar
-rw-rw-r--  1 kfk kfk   433368 Jan  7 17:35 httpclient-4.2.5.jar
-rw-rw-r--  1 kfk kfk   227708 Jan  7 17:35 httpcore-4.2.5.jar
-rw-rw-r--  1 kfk kfk   611863 Jan  7 17:35 janino-2.7.6.jar
-rw-rw-r--  1 kfk kfk   114415 Jan  7 17:35 jansi-1.11.jar
-rw-rw-r--  1 kfk kfk    60527 Jan  7 17:35 jcommander-1.32.jar
-rw-rw-r--  1 kfk kfk   201124 Jan  7 17:35 jdo-api-3.0.1.jar
-rw-rw-r--  1 kfk kfk  1681148 Jan  7 17:35 jetty-all-7.6.0.v20120127.jar
-rw-rw-r--  1 kfk kfk  1683027 Jan  7 17:35 jetty-all-server-7.6.0.v20120127.jar
-rw-rw-r--  1 kfk kfk    87325 Jan  7 17:35 jline-0.9.94.jar
-rw-rw-r--  1 kfk kfk    12131 Jan  7 17:35 jpam-1.1.jar
-rw-rw-r--  1 kfk kfk    33015 Jan  7 17:35 jsr305-1.3.9.jar
-rw-rw-r--  1 kfk kfk    15071 Jan  7 17:35 jta-1.1.jar
-rw-rw-r--  1 kfk kfk   245039 Jan  7 17:35 junit-4.11.jar
-rw-rw-r--  1 kfk kfk   275186 Jan  7 17:35 libfb303-0.9.0.jar
-rw-rw-r--  1 kfk kfk   347531 Jan  7 17:35 libthrift-0.9.0.jar
-rw-rw-r--  1 kfk kfk   489213 Jan  7 17:35 linq4j-0.4.jar
-rw-rw-r--  1 kfk kfk   481535 Jan  7 17:35 log4j-1.2.16.jar
-rw-rw-r--  1 kfk kfk   447676 Jan  7 17:35 mail-1.4.1.jar
-rw-rw-r--  1 kfk kfk    94421 Jan  7 17:35 maven-scm-api-1.4.jar
-rw-rw-r--  1 kfk kfk    40066 Jan  7 17:35 maven-scm-provider-svn-commons-1.4.jar
-rw-rw-r--  1 kfk kfk    69858 Jan  7 17:35 maven-scm-provider-svnexe-1.4.jar
-rw-rw-r--  1 kfk kfk   827942 Jan  7 17:36 mysql-connector-java-5.1.21.jar
-rw-rw-r--  1 kfk kfk    19827 Jan  7 17:35 opencsv-2.3.jar
-rw-rw-r--  1 kfk kfk    65261 Jan  7 17:35 oro-2.0.8.jar
-rw-rw-r--  1 kfk kfk    29555 Jan  7 17:35 paranamer-2.3.jar
-rw-rw-r--  1 kfk kfk    48528 Jan  7 17:35 pentaho-aggdesigner-algorithm-5.1.3-jhyde.jar
drwxrwxr-x  6 kfk kfk     4096 Jan  7 17:35 php
-rw-rw-r--  1 kfk kfk   250546 Jan  7 17:35 plexus-utils-1.5.6.jar
drwxrwxr-x 10 kfk kfk     4096 Jan  7 17:35 py
-rw-rw-r--  1 kfk kfk    37325 Jan  7 17:35 quidem-0.1.1.jar
-rw-rw-r--  1 kfk kfk    25429 Jan  7 17:35 regexp-1.3.jar
-rw-rw-r--  1 kfk kfk   105112 Jan  7 17:35 servlet-api-2.5.jar
-rw-rw-r--  1 kfk kfk  1251514 Jan  7 17:35 snappy-java-1.0.5.jar
-rw-rw-r--  1 kfk kfk   236660 Jan  7 17:35 ST4-4.0.4.jar
-rw-rw-r--  1 kfk kfk    26514 Jan  7 17:35 stax-api-1.0.1.jar
-rw-rw-r--  1 kfk kfk   148627 Jan  7 17:35 stringtemplate-3.2.1.jar
-rw-rw-r--  1 kfk kfk    93210 Jan  7 17:35 super-csv-2.2.0.jar
-rw-rw-r--  1 kfk kfk    55953 Jan  7 17:35 tempus-fugit-1.1.jar
-rw-rw-r--  1 kfk kfk   392124 Jan  7 17:35 velocity-1.5.jar
-rw-rw-r--  1 kfk kfk    94672 Jan  7 17:35 xz-1.0.jar
-rw-rw-r--  1 kfk kfk   779974 Jan  7 17:35 zookeeper-3.4.5.jar
[kfk@bigdata-pro01 lib]$ 

 

 

 

 

 

 

    (4)如果你遇到更多的Hive问题,见如下,也许能帮助到你

 

 

 

 

 

 

 

 

 

   然后,再来执行

[kfk@bigdata-pro01 apache-hive-1.0.0-bin]$ bin/hive

Logging initialized using configuration in file:/opt/modules/apache-hive-1.0.0-bin/conf/hive-log4j.properties
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/modules/hadoop-2.6.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/modules/apache-hive-1.0.0-bin/lib/hive-jdbc-1.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
hive> create external table weblogs(
    > id string,
    > datatime string,
    > userid string,
    > searchname string,
    > retorder string,
    > cliorder string,
    > cliurl string
    > )  
    > STORED BY  'org.apache.hadoop.hive.hbase.HBaseStorageHandler' 
    > WITH SERDEPROPERTIES("hbase.columns.mapping" = ":key,info:datatime,info:userid,info:searchname,info:retorder,info:cliorder,info:cliurl") 
    > TBLPROPERTIES("hbase.table.name" = "weblogs");

 

 

 

 

  或者

[kfk@bigdata-pro01 apache-hive-1.0.0-bin]$ bin/hive

Logging initialized using configuration in file:/opt/modules/apache-hive-1.0.0-bin/conf/hive-log4j.properties
SLF4J: Class path contains multiple SLF4J bindings.
SLF4J: Found binding in [jar:file:/opt/modules/hadoop-2.6.0/share/hadoop/common/lib/slf4j-log4j12-1.7.5.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: Found binding in [jar:file:/opt/modules/apache-hive-1.0.0-bin/lib/hive-jdbc-1.0.0-standalone.jar!/org/slf4j/impl/StaticLoggerBinder.class]
SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.
SLF4J: Actual binding is of type [org.slf4j.impl.Log4jLoggerFactory]
hive> create external table weblogs(id string,datatime string,userid string,searchname string,retorder string,cliorder string,cliurl string)  STORED BY  'org.apache.hadoop.hive.hbase.HBaseStorageHandler' WITH SERDEPROPERTIES("hbase.columns.mapping" = ":key,info:datatime,info:userid,info:searchname,info:retorder,info:cliorder,info:cliurl") TBLPROPERTIES("hbase.table.name" = "weblogs");

     以上是创建与HBase集成的Hive的外部表。

     你创建这个只是告诉hive 去读Hbase的表。
     你Hbase对应的表的结构要和你创建的一样。
     你flume采集的数据有没有进入到hbase到呀。
 
 

 

 

 

 

 

 

 

欢迎大家,加入我的微信公众号:大数据躺过的坑        人工智能躺过的坑
 
 
 

同时,大家可以关注我的个人博客

   http://www.cnblogs.com/zlslch/   和     http://www.cnblogs.com/lchzls/      http://www.cnblogs.com/sunnyDream/   

   详情请见:http://www.cnblogs.com/zlslch/p/7473861.html

 

  人生苦短,我愿分享。本公众号将秉持活到老学到老学习无休止的交流分享开源精神,汇聚于互联网和个人学习工作的精华干货知识,一切来于互联网,反馈回互联网。
  目前研究领域:大数据、机器学习、深度学习、人工智能、数据挖掘、数据分析。 语言涉及:Java、Scala、Python、Shell、Linux等 。同时还涉及平常所使用的手机、电脑和互联网上的使用技巧、问题和实用软件。 只要你一直关注和呆在群里,每天必须有收获

 

      对应本平台的讨论和答疑QQ群:大数据和人工智能躺过的坑(总群)(161156071) 

 

 

 

 

 

 

 

 

 

 

 

 

转载于:https://www.cnblogs.com/zlslch/p/8228781.html

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论

“相关推荐”对你有帮助么?

  • 非常没帮助
  • 没帮助
  • 一般
  • 有帮助
  • 非常有帮助
提交
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值