object SecondSort {
def main(args: Array[String]) {
val conf=new SparkConf().setAppName("SecondSort").setMaster("local[2]")
val sc=new SparkContext(conf)
val textFile=sc.textFile("/spark/second.txt")
val word=textFile
.map(_.split("\t")) //: RDD[Array[String]]
.map(tuple=>(tuple(0),tuple(1).toInt)) //: RDD[(String, Int)]
.groupByKey() //: RDD[(String, Iterable[Int])]
.map(tuple=>{
val key=tuple._1
val value=tuple._2
(key,value.toList.sorted)
}) //: RDD[(String, List[Int])]
.sortByKey() //: RDD[(String, List[Int])]
word.foreach(count=>println(count._1+":"+count._2.toString()))
}
}
def main(args: Array[String]) {
val conf=new SparkConf().setAppName("SecondSort").setMaster("local[2]")
val sc=new SparkContext(conf)
val textFile=sc.textFile("/spark/second.txt")
val word=textFile
.map(_.split("\t")) //: RDD[Array[String]]
.map(tuple=>(tuple(0),tuple(1).toInt)) //: RDD[(String, Int)]
.groupByKey() //: RDD[(String, Iterable[Int])]
.map(tuple=>{
val key=tuple._1
val value=tuple._2
(key,value.toList.sorted)
}) //: RDD[(String, List[Int])]
.sortByKey() //: RDD[(String, List[Int])]
word.foreach(count=>println(count._1+":"+count._2.toString()))
}
}