1、文件数据
Spark,100
Hadoop,62
Flink,77
Kafka,91
Hadoop,93
Spark,78
Hadoop,69
Spark,98
Hadoop,62
Spark,99
Hadoop,61
Spark,70
Hadoop,75
Spark,88
Hadoop,68
Spark,90
Hadoop,61
2、Scala代码:
package topN
import org.apache.spark.rdd.RDD
import org.apache.spark.{SparkConf, SparkContext}
object GroupTop {
def main(args: Array[String]): Unit = {
/**
* 判断参数的多少
*/
if (args.length < 2){
println(
"""
|topN.GroupTop<inputPath><outputPath>
|<inputPath> 文件输入目录
|<outputPath> 输出目录
""".stripMargin
)
System.exit(0)
}
/**
* 接收参数
*/
val Array(inputPath,outputPath) = args
/**
* 初始化程序入口
*/
val conf = new SparkConf()
conf.setAppName(s"${this.getClass.getSimpleName}")
conf.setMaster("local")