代码:
import org.apache.spark.{SparkConf, SparkContext}
import scala.util.Random
object MaxAndMin extends App {
System.setProperty("hadoop.home.dir","D:\\soft\\hadoop\\hadoop-2.7.3")
val conf=new SparkConf().setMaster("local[*]").setAppName("MaxAndMin ")
val sc=new SparkContext(conf)
val lines=sc.textFile(args(0))
//s"key${Random.nextInt(lines.getNumPartitions)} 线程
lines.filter(line=> line.trim.length>0).map(line => (s"key${Random.nextInt(lines.getNumPartitions)}",line.trim.toInt)).groupByKey()
.map(x=>{
var min =Integer.MAX_VALUE
var max =Integer.MIN_VALUE
for (num <- x._2){
if (num >max){
max=num
}
if (num< min){
min=num
}
}
(max,min)
}).collect().foreach(x => {
println("max" + x._1)
println("min "+ x._2)
})
sc.stop()
}