Hive的安装和使用

1 .安装Hive:

sudo tar xvfz apache-hive-2.1.0-bin.tar.gz

sudo cp -R apache-hive-2.1.0-bin /usr/local/hive

sudo chown -R hadoop:hadoop /usr/local/hive

2. 修改/etc/profile加入HIVE_HOME的变量:

export HIVE_HOME=/usr/local/hive

export PATH=$PATH:$HIVE_HOME/bin

export CLASSPATH=.:${JAVA_HOME}/lib:${JRE_HOME}/lib:/usr/local/hive/lib

$source /etc/profile

3 .修改hive/conf下的几个template模板并重命名为其他:

cp hive-env.sh.template hive-env.sh

cp hive-default.xml.template hive-site.xml

配置hive-env.sh文件,指定HADOOP_HOME

HADOOP_HOME=/usr/local/hadoop

4 .修改hive-site.xml文件,指定MySQL数据库驱动、数据库名、用户名及密码,修改的内容如下所示

<property>

  <name>javax.jdo.option.ConnectionURL</name>

<value>jdbc:mysql://192.168.111.16:3306/hive?createDatabaseIfNotExist=true</value>

  <description>JDBC connect string for a JDBC metastore</description>

</property>

<property>

  <name>javax.jdo.option.ConnectionDriverName</name>

  <value>com.mysql.jdbc.Driver</value>

  <description>Driver class name for a JDBC metastore</description>

</property>

<property>

  <name>javax.jdo.option.ConnectionUserName</name>

  <value>hadoop</value>

  <description>username to use against metastore database</description>

</property>

<property>

  <name>javax.jdo.option.ConnectionPassword</name>

  <value>1234</value>

  <description>password to use against metastore database</description>

</property>

其中:

javax.jdo.option.ConnectionURL参数指定的是Hive连接数据库的连接字符串;

javax.jdo.option.ConnectionDriverName参数指定的是驱动的类入口名称;

javax.jdo.option.ConnectionUserName参数指定了数据库的用户名;

javax.jdo.option.ConnectionPassword参数指定了数据库的密码。

5 .缓存目录的问题,如果不配置也会出错的

 <property>

 <name>hive.exec.local.scratchdir</name>

 <value>/home/hadoop/iotmp</value>

 <description>Local scratch space for Hive jobs</description>

 </property>

 <property>

 <name>hive.downloaded.resources.dir</name>

 <value>/home/hadoop/iotmp</value>

 <description>Temporary local directory for added resources in the remote file system.</description>

 </property>

并且需要对目录进行权限设定:

mkdir -p /home/hadoop/iotmp

chmod -R 775 /home/hadoop/iotmp

6.修改hive/bin下的hive-config.sh文件,设置JAVA_HOME,HADOOP_HOME

export JAVA_HOME=/usr/lib/jvm

export HADOOP_HOME=/usr/local/hadoop

export HIVE_HOME=/usr/local/hive

7.下载mysql-connector-java-5.1.27-bin.jar文件,并放到$HIVE_HOME/lib目录下可以从Mysql的官方网站下载

8.HDFS中创建/tmp/user/hive/warehouse并设置权限

hadoop fs -mkdir /tmp

hadoop fs -mkdir -p /user/hive/warehouse

hadoop fs -chmod g+w /tmp

hadoop fs -chmod g+w /user/hive/warehouse

9.初始化meta数据库

进入之前需要初始化数据库

schematool -initSchema -dbType mysql

 

hadoop@hadoopmaster:/usr/local/hive/lib$ schematool -initSchema -dbType mysql

SLF4J: Class path contains multiple SLF4J bindings.

SLF4J: Found binding in [jar:file:/usr/local/hive/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.

SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]

Metastore connection URL: jdbc:mysql://192.168.1.166:3306/hive?createDatabaseIfNotExist=true

Metastore Connection Driver : com.mysql.jdbc.Driver

Metastore connection User: hive

Starting metastore schema initialization to 2.1.0

Initialization script hive-schema-2.1.0.mysql.sql

Initialization script completed

schemaTool completed

 

 

10.测试hive shell

hive

show databases;

show tables;

11.可以在hadoop中查看hive生产的文件

hadoop dfs -ls /user/hive/warehouse

12. Hive shell使用实例

(1) 创建数据(文本以tab分隔)

~ vi /home/cos/demo/t_hive.txt

16      2       3

61      12      13

41      2       31

17      21      3

71      2       31

1       12      34

11      2       34

(2 )创建新表

hive> CREATE TABLE t_hive (a int, b int, c int) ROW FORMAT DELIMITED FIELDS TERMINATED BY '\t';

OK

Time taken: 0.121 seconds

(3 )导入数据t_hive.txtt_hive

hive> LOAD DATA LOCAL INPATH '/tmp/t_hive.txt' OVERWRITE INTO TABLE t_hive ;

Loading data to table default.t_hive

OK

Time taken: 0.609 seconds

(4 )查看表

hive> show tables;

OK

t_hive

Time taken: 0.099 seconds

(5 )正则匹配表名

hive>show tables '*t*';

OK

t_hive

Time taken: 0.065 seconds

(6 )查看表数据

hive> select * from t_hive;

OK

16      2       3

61      12      13

41      2       31

17      21      3

71      2       31

1       12      34

11      2       34

Time taken: 0.264 seconds

(7) 查看表结构

hive> desc t_hive;

OK

a       int

b       int

c       int

Time taken: 0.1 seconds

(8 )增加一个字段

hive> ALTER TABLE t_hive ADD COLUMNS (new_col String);

OK

Time taken: 0.186 seconds

hive> desc t_hive;

OK

a       int

b       int

c       int

new_col string

Time taken: 0.086 seconds

(9 )重命令表名

~ ALTER TABLE t_hive RENAME TO t_hadoop;

OK

Time taken: 0.45 seconds

hive> show tables;

OK

t_hadoop

Time taken: 0.07 seconds

(10 )删除表

hive> DROP TABLE t_hadoop;

OK

Time taken: 0.767 seconds

hive> show tables;

OK

Time taken: 0.064 seconds

FAQ

1.hadoop@hadoopmaster:/usr/local/hive/conf$ hive

SLF4J: Class path contains multiple SLF4J bindings.

SLF4J: Found binding in [jar:file:/usr/local/hive/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.

SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]

原因:

没有执行schematool -initSchema -dbType mysql  执行之后搞定

 

2. hadoop@hadoopmaster:/usr/local/hive/lib$ hive

SLF4J: Class path contains multiple SLF4J bindings.

SLF4J: Found binding in [jar:file:/usr/local/hive/lib/log4j-slf4j-impl-2.4.1.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: Found binding in [jar:file:/usr/local/hadoop/share/hadoop/common/lib/slf4j-log4j12-1.7.10.jar!/org/slf4j/impl/StaticLoggerBinder.class]

SLF4J: See http://www.slf4j.org/codes.html#multiple_bindings for an explanation.

SLF4J: Actual binding is of type [org.apache.logging.slf4j.Log4jLoggerFactory]

Logging initialized using configuration in jar:file:/usr/local/hive/lib/hive-common-2.1.0.jar!/hive-log4j2.properties Async: true

Exception in thread "main" java.lang.IllegalArgumentException: java.net.URISyntaxException: Relative path in absolute URI: ${system:java.io.tmpdir%7D/$%7Bsystem:user.name%7D

 

需要对目录进行权限设定

mkdir -p /home/hadoop/iotmp

chmod -R 775 /home/hadoop/iotmp

 

3.hadoop命令报错:

 

解决方法:

 

 

4.启动hive:

 

 

5.导入hive数据为NULL

 

解决出现NULL:

直接copy 过来的,使用TAB键,可以正常显示

 

  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值