Exception in thread "main" org.apache.spark.sql.AnalysisException: unresolved operator 'InsertIntoStatement HiveTableRelation [`test_bigdata`.`test_shd_atlas_spline`, org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe, Data Cols: [id#23, lineage#24, fetchedtime#25], Partition Cols: [ds#26]], [ds=None], false, false; 'InsertIntoStatement HiveTableRelation [`test_bigdata`.`test_shd_atlas_spline`, org.apache.hadoop.hive.ql.io.parquet.serde.ParquetHiveSerDe, Data Cols: [id#23, lineage#24, fetchedtime#25], Partition Cols: [ds#26]], [ds=None], false, false +- Project [ansi_cast(_1#15 as string) AS id#27, ansi_cast(_2#16 as string) AS lineage#28, ansi_cast(_3#17 as int) AS fetchedtime#29, ansi_cast(_4#18 as string) AS ds#30] +- Project [_1#15, _2#16, _3#17, _4#18] +- SubqueryAlias aa +- SerializeFromObject [staticinvoke(class org.apache.spark.unsafe.types.UTF8String, StringType, fromString, knownnotnull(assertnotnull(input[0, scala.Tuple4, true]))._1, true, false) AS _1#15, staticinvoke(class org.apache.spark.unsafe.types.UTF8String, StringType, fromString, knownnotnull(assertnotnull(input[0, scala.Tuple4, true]))._2, true, false) AS _2#16, knownnotnull(assertnotnull(input[0, scala.Tuple4, true]))._3.intValue AS _3#17, staticinvoke(class org.apache.spark.unsafe.types.UTF8String, StringType, fromString, knownnotnull(assertnotnull(input[0, scala.Tuple4, true]))._4, true, false) AS _4#18] +- ExternalRDD [obj#14] at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.failAnalysis(CheckAnalysis.scala:50) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.failAnalysis$(CheckAnalysis.scala:49) at org.apache.spark.sql.catalyst.analysis.Analyzer.failAnalysis(Analyzer.scala:155) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$47(CheckAnalysis.scala:702) at org.apache.spark.sql.catalyst.analysis.CheckAnalysis.$anonfun$checkAnalysis$47$adapted(CheckAnalysis.scala:700) at org.apache.spark.sql.catalyst.tr