根据网上教程的安装和将MySQL的数据导入到hive里面一直出现说java.lang.ClassNotFoundException: Class Job_fill not found
我的导入语句开始是这样的:
bin/sqoop import \
--connect jdbc:mysql://namenode:3306/test \
--username root \
--password 远程连接密码 \
--table Job_fill \
--num-mappers 1 \
--hive-import \
--hive-overwrite \
--fields-terminated-by ',' \
--hive-database test \
--hive-table 'Job_fill'
但上面的语句在sqoop执行一直会报
java.lang.Exception: java.lang.RuntimeException: java.lang.ClassNotFoundException: Class Job_fill not found
at org.apache.hadoop.mapred.LocalJobRunner$Job.runTasks(LocalJobRunner.java:492)
at org.apache.hadoop.mapred.LocalJobRunner$Job.run(LocalJobRunner.java:552)
Caused by: java.lang.RuntimeException: java.lang.ClassNotFoundException: Class Job_fill not found
at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2638)
at org.apache.sqoop.mapreduce.db.DBConfiguration.getInputClass(DBConfiguration.java:403)
at org.apache.sqoop.mapreduce.db.DataDrivenDBInputFormat.createDBRecordReader(DataDrivenDBInputFormat.java:237)
at org.apache.sqoop.mapreduce.db.DBInputFormat.createRecordReader(DBInputFormat.java:263)
at org.apache.hadoop.mapred.MapTask$NewTrackingRecordReader.<init>(MapTask.java:527)
at org.apache.hadoop.mapred.MapTask.runNewMapper(MapTask.java:770)
at org.apache.hadoop.mapred.MapTask.run(MapTask.java:347)
at org.apache.hadoop.mapred.LocalJobRunner$Job$MapTaskRunnable.run(LocalJobRunner.java:271)
at java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:511)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1149)
at java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:624)
at java.lang.Thread.run(Thread.java:748)
Caused by: java.lang.ClassNotFoundException: Class Job_fill not found
at org.apache.hadoop.conf.Configuration.getClassByName(Configuration.java:2542)
at org.apache.hadoop.conf.Configuration.getClass(Configuration.java:2636)
... 12 more
后面我找到解决办法是将机器名改成我们安装sqoop的机器的IP地址,这样就可以很好的解决这个问题,我认为出现这个问题原因是sqoop可以读取mysql文件也能保存到hdfs上作为临时文件,但连接hive可能会有问题,hive不能直接识别机器名称,要用机器的IP地址才能精确的识别到,当然这个也是我的猜测,改的语句如下:
bin/sqoop import \
--connect jdbc:mysql://192.168.**.***:3306/test \
--username root \
--password 自己MySQL远程连接密码\
--table Job_fill \
--num-mappers 1 \
--hive-import \
--hive-overwrite \
--fields-terminated-by ',' \
--hive-database test \
--hive-table 'Job_fill'
如果你出现以上问题用我这个方法还不能解决可以留言看看是什么问题