import org.apache.spark.rdd.RDD
import org.apache.spark.{
SparkConf, SparkContext}
/**
* Spark应用:自定义排序 + 二次排序
* 需求:对数据集按照产品价格降序排序,产品价格相同再按照产品数量降序排序
*/
object sortAPP {
def main(args: Array[String]): Unit = {
//初始化sc
val sparkConf = new SparkConf().setAppName("sortAPP").setMaster("local[2]")
val sc = new SparkContext(sparkConf)
//生成RDD
val rdd = sc.makeRDD(List(
"Huawei P40,4999,1000",
"Huawei P30 Pro,3999,5000",
"IPhone 8P,6999,6000",
"IPhone X,10999,3000",
"IPhone 7,4999,3000",
"Sunsum Galexy,5399,6000",
"OPPO R11,3999,8000"))
//功能实现
//方法一:通过元组实现自定义排序
val sortedRDD1 = rdd.map(line => {
val strings = line.split(",")
val proName = strings(0)
val proPrice =
Spark开发之四种方式实现Scala中的自定义排序和二次排序算法
最新推荐文章于 2020-08-15 21:06:18 发布