报错
喜欢到处游荡的家伙
hello world
展开
-
Hive support is required to CREATE Hive TABLE (AS SELECT)
idea中,Spark SQL连接到了一个部署好的Hive,尝试在 包含Hive支持的Spark SQL中CREATE TABLE我的代码:package spark.hiveimport org.apache.spark.sql.SparkSessionobject SparkOnHive extends App { case class Record(key: Int, value: String) //val warehouseLocation = new File("/use原创 2020-07-21 23:02:01 · 3875 阅读 · 0 评论 -
Illegal pattern component: XXX
IDEA创建SparkSQL程序报错我的代码:package spark.sparksqlimport org.apache.spark.sql.SparkSessionobject SparkSql { def main(args: Array[String]): Unit = { //创建sparkconf val spark = SparkSession.builder().appName("Spark Sql lm") .config("spark.some原创 2020-07-21 22:39:40 · 964 阅读 · 1 评论 -
spark的错误
spark的错误idea中通过反射获取Scheam报错我的原始代码:package sparkimport org.apache.spark.sql.SparkSessionimport org.apache.spark.{SparkConf, SparkContext}object Rdd2DF { def main(args: Array[String]): Unit = { //创建SparkSession val spark = SparkSession原创 2020-07-16 14:44:36 · 672 阅读 · 0 评论 -
报错
类型不匹配mismatchmismatch报错如下:INFO mapreduce.Job:Task Id:attempt_1589445736284_0017_m_000000_0, Status : FAILEDError: java.io.IOException:Type mismatch in key from map:expected org.apache.hadoop.io.Text, received org.example.sort2.FlowBean at org.apache.h原创 2020-07-11 13:17:41 · 4398 阅读 · 1 评论