1.下载hive软件
wget https://mirrors.tuna.tsinghua.edu.cn/apache/hive/hive-2.1.1/apache-hive-2.1.1-bin.tar.gz
3.~/.bash_profile
HIVE_HOME=/usr/local/apache-hive-2.1.1-bin
export HADOOP_CLASSPATH=$HIVE_HOME/conf:$HIVE_HOME/lib/*
PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$HIVE_HOME/bin:
#cp hive-default.xml.template hive-site.xml
hive-site.xml
<property>
<name>javax.jdo.option.ConnectionDriverName</name>
<value>com.mysql.jdbc.Driver</value>
<description>Driver class name for a JDBC metastore</description>
</property>
<property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc:mysql://localhost:3306/hive?createDatabaseIfNotExist=true</value>
<description>JDBC connect string for a JDBC metastore</description>
</property>
<property>
<name>javax.jdo.option.ConnectionUserName</name>
<value>hadoop</value>
<description>Username to use against metastore database</description>
</property>
<property>
<name>javax.jdo.option.ConnectionPassword</name>
<value>neusoft</value>
<description>password to use against metastore database</description>
</property>
<property>
<name>hive.metastore.uris</name>
<value>thrift:xslave3:9000</value>
<description/>
</property>
<property>
<name>hive.metastore.warehouse.dir</name>
<value>/user/hive/warehouse</value> (创建该目录并赋权)
<description>location of default database for the warehouse</description>
</property>
<property>
<name>hive.exec.scratchdir</name>
<value>/tmp/hive</value> (创建该目录并赋权)
<description>HDFS root scratch dir for Hive jobs which gets created with write all (733) permission. For each connecting user, an HDFS scratch dir: ${hive.exec.scratchdir}/<username> is created, with ${hive.scratch.dir.permission}.</description>
</property>
4.apache-hive-2.0.0-bin/conf/hive-env.sh(按需要添加)
export HADOOP_HOME=/usr/local/hadoop-2.6.4 ##Hadoop安装路径
export HIVE_CONF_DIR=/usr/local/apache-hive-2.1.1-bin/conf ##Hive配置文件路径
export JAVA_HOME=/opt/java ##Java路径
export HIVE_HOME=/opt/hive ##Hive安装路径
5./sqoop-1.4.6-cdh5.8.2/conf/sqoop-env.sh
export HIVE_HOME=/usr/local/apache-hive-2.0.0-bin
export PATH=$HIVE_HOME/bin:$PATH
export HIVE_CONF_DIR=/usr/local/apache-hive-2.0.0-bin/conf
6.启动metastore
service mysqld start
schematool -dbType mysql -initSchema 初始化
nohup hive --service metastore &
hive --service metastore 启动metastore
hive --service hiveserver2 启动hive_service 用beeline hive
-----------------------
wget https://mirrors.tuna.tsinghua.edu.cn/apache/hive/hive-2.1.1/apache-hive-2.1.1-bin.tar.gz
按需下载
2.解压
tar -zxvf apache-hive-2.1.1-bin.tar.gz3.~/.bash_profile
HIVE_HOME=/usr/local/apache-hive-2.1.1-bin
export HADOOP_CLASSPATH=$HIVE_HOME/conf:$HIVE_HOME/lib/*
PATH=$PATH:$JAVA_HOME/bin:$HADOOP_HOME/bin:$HADOOP_HOME/sbin:$HIVE_HOME/bin:
#source .bash_profile(使其生效)
4.在apache-hive-2.1.1-bin/lib目录下添加mysql-connector-java-5.1.38-bin.jar包
5.apache-hive-2.1.1-bin/conf#cp hive-default.xml.template hive-site.xml
hive-site.xml
<property>
<name>javax.jdo.option.ConnectionDriverName</name>
<value>com.mysql.jdbc.Driver</value>
<description>Driver class name for a JDBC metastore</description>
</property>
<property>
<name>javax.jdo.option.ConnectionURL</name>
<value>jdbc:mysql://localhost:3306/hive?createDatabaseIfNotExist=true</value>
<description>JDBC connect string for a JDBC metastore</description>
</property>
<property>
<name>javax.jdo.option.ConnectionUserName</name>
<value>hadoop</value>
<description>Username to use against metastore database</description>
</property>
<property>
<name>javax.jdo.option.ConnectionPassword</name>
<value>neusoft</value>
<description>password to use against metastore database</description>
</property>
<property>
<name>hive.metastore.uris</name>
<value>thrift:xslave3:9000</value>
<description/>
</property>
<property>
<name>hive.metastore.warehouse.dir</name>
<value>/user/hive/warehouse</value> (创建该目录并赋权)
<description>location of default database for the warehouse</description>
</property>
<property>
<name>hive.exec.scratchdir</name>
<value>/tmp/hive</value> (创建该目录并赋权)
<description>HDFS root scratch dir for Hive jobs which gets created with write all (733) permission. For each connecting user, an HDFS scratch dir: ${hive.exec.scratchdir}/<username> is created, with ${hive.scratch.dir.permission}.</description>
</property>
4.apache-hive-2.0.0-bin/conf/hive-env.sh(按需要添加)
export HADOOP_HOME=/usr/local/hadoop-2.6.4 ##Hadoop安装路径
export HIVE_CONF_DIR=/usr/local/apache-hive-2.1.1-bin/conf ##Hive配置文件路径
export JAVA_HOME=/opt/java ##Java路径
export HIVE_HOME=/opt/hive ##Hive安装路径
5./sqoop-1.4.6-cdh5.8.2/conf/sqoop-env.sh
export HIVE_HOME=/usr/local/apache-hive-2.0.0-bin
export PATH=$HIVE_HOME/bin:$PATH
export HIVE_CONF_DIR=/usr/local/apache-hive-2.0.0-bin/conf
6.启动metastore
service mysqld start
schematool -dbType mysql -initSchema 初始化
nohup hive --service metastore &
hive --service metastore 启动metastore
hive --service hiveserver2 启动hive_service 用beeline hive
-----------------------