kerberos认证_Zeppelin带有Kerberos认证的Hive解释器的配置

a72d81a7f8767781ce660761c77e00fc.png

2.zeppelin连接Hive安装配置

zeppelin 版本0.8.2 ,hive版本:3.0.0

2.1.安装启动hive3

2.1.配置hiveserver2

如果需要配置zeppelin与hive的集成,我们需要启动hive的metastore服务以及hiveserver2服务。

首先为hive配置hiveserver2

conf/hive-site.xml

hive.server2.thrift.port

10005

hive.server2.thrift.bind.host

hadoop01

hive.server2.long.polling.timeout

5000

hive.metastore.uris

thrift://hadoop01:9083

启动hiveserver2:

nohup bin/hive --service hiveserver2 &

2.2.beeline 连接:

beeline -u "jdbc:hive2://hadoop01:10005/feedback;principal=hive/_HOST@YOUR_REALM;auth=kerberos"

2.3.[Zeppelin] Kerberos认证下Hive解释器的配置

default.driver = org.apache.hive.jdbc.HiveDriver

# thrift端口是10000,http端口是10001

# 如果hive开启了负载均衡,则不能使用这些端口直连,只能连接负载均衡器的地址和端口

# YOUR_REALM表示kerberos所属域,需要替换成实际的域,不清楚可以查看/etc/krb5.conf

default.url = jdbc:hive2://localhost:10000/default;principal=hive/_HOST@YOUR_REALM

default.user = hive

zeppelin.jdbc.auth.type = KERBEROS

# 运行Zeppelin的用户需要对hive.keytab具有读权限

zeppelin.jdbc.keytab.location = /home/xxxxx/hive.keytab

# 需要替换

zeppelin.jdbc.principal = hive@YOUR_REALM

default.proxy.user.property = hive.server2.proxy.user

2.4.添加依赖包

可以直接把jar 拷贝到interpreter/jdbc/目录下,也可以在zeppelin界面上添加。

为什么是拷贝到 interpreter/jdbc 目录呢?因为我们配置的hive解释器属于JDBC组,启动解释器时 interpreter/jdbc 目录会被加到CLASSPATH里。

2.5.配置时报错内容:

org.apache.zeppelin.interpreter.InterpreterException: Error in doAs at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:479) at org.apache.zeppelin.jdbc.JDBCInterpreter.executeSql(JDBCInterpreter.java:692) at org.apache.zeppelin.jdbc.JDBCInterpreter.interpret(JDBCInterpreter.java:820) at org.apache.zeppelin.interpreter.LazyOpenInterpreter.interpret(LazyOpenInterpreter.java:103) at org.apache.zeppelin.interpreter.remote.RemoteInterpreterServer$InterpretJob.jobRun(RemoteInterpreterServer.java:632) at org.apache.zeppelin.scheduler.Job.run(Job.java:188) at org.apache.zeppelin.scheduler.ParallelScheduler$JobRunner.run(ParallelScheduler.java:162) at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511) at java.util.concurrent.FutureTask.run(FutureTask.java:266) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:180) at java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:293) at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149) at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624) at java.lang.Thread.run(Thread.java:748) Caused by: java.lang.reflect.UndeclaredThrowableException at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1700) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnection(JDBCInterpreter.java:471) ... 13 more Caused by: java.sql.SQLException: Could not open client transport with JDBC Uri: jdbc:hive2://10.16.0.83:10005/feedback;principal=mcloud/bj-jd-appsubmit-node-prod-0002.tendcloud.com@HADOOP.COM;auth=kerberos: GSS initiate failed at org.apache.hive.jdbc.HiveConnection.(HiveConnection.java:256) at org.apache.hive.jdbc.HiveDriver.connect(HiveDriver.java:107) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:208) at org.apache.commons.dbcp2.DriverManagerConnectionFactory.createConnection(DriverManagerConnectionFactory.java:79) at org.apache.commons.dbcp2.PoolableConnectionFactory.makeObject(PoolableConnectionFactory.java:205) at org.apache.commons.pool2.impl.GenericObjectPool.create(GenericObjectPool.java:861) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:435) at org.apache.commons.pool2.impl.GenericObjectPool.borrowObject(GenericObjectPool.java:363) at org.apache.commons.dbcp2.PoolingDriver.connect(PoolingDriver.java:129) at java.sql.DriverManager.getConnection(DriverManager.java:664) at java.sql.DriverManager.getConnection(DriverManager.java:270) at org.apache.zeppelin.jdbc.JDBCInterpreter.getConnectionFromPool(JDBCInterpreter.java:425) at org.apache.zeppelin.jdbc.JDBCInterpreter.access$000(JDBCInterpreter.java:91) at org.apache.zeppelin.jdbc.JDBCInterpreter$2.run(JDBCInterpreter.java:474) at org.apache.zeppelin.jdbc.JDBCInterpreter$2.run(JDBCInterpreter.java:471) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682) ... 14 more Caused by: org.apache.thrift.transport.TTransportException: GSS initiate failed at org.apache.thrift.transport.TSaslTransport.sendAndThrowMessage(TSaslTransport.java:232) at org.apache.thrift.transport.TSaslTransport.open(TSaslTransport.java:316) at org.apache.thrift.transport.TSaslClientTransport.open(TSaslClientTransport.java:37) at org.apache.hadoop.hive.metastore.security.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:51) at org.apache.hadoop.hive.metastore.security.TUGIAssumingTransport$1.run(TUGIAssumingTransport.java:48) at java.security.AccessController.doPrivileged(Native Method) at javax.security.auth.Subject.doAs(Subject.java:422) at org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1682) at org.apache.hadoop.hive.metastore.security.TUGIAssumingTransport.open(TUGIAssumingTransport.java:48) at org.apache.hive.jdbc.HiveConnection.openTransport(HiveConnection.java:343) at org.apache.hive.jdbc.HiveConnection.(HiveConnection.java:228) ... 32 more

解决:解释器中添加如下配置完美解决:

default.proxy.user.property=hive.server2.proxy.user

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值