import java.util
import org.apache.spark.sql.types.{DataTypes, StructField}
import org.apache.spark.sql.{Row, SQLContext}
import org.apache.spark.{SparkConf, SparkContext}
object TelephoneData13 {
def main(args: Array[String]): Unit = {
val conf = new SparkConf().setMaster("local").setAppName("TelephoneData13")
val sc = new SparkContext(conf)
val sQLContext = new SQLContext(sc)
val rdd = sQLContext.read.format("com.databricks.spark.csv")
.option("sep", ",")
.option("header", "true")
.load("F:\\ideaWorkspace\\hello\\data\\yy.csv")
rdd.foreach(x => println(x))
rdd.printSchema()
rdd.show(1000,false)
val temp = rdd.rdd.map(line => {
Row(
line.getString(0).toString,
<