package com.ser
import java.net.InetAddress
import org.apache.spark.rdd.RDD
import org.apache.spark.{SparkConf, SparkContext}
object serTest {
def main(args: Array[String]): Unit = {
val infos: getComputerInfo.type = getComputerInfo
val conf: SparkConf = new SparkConf().setAppName("sertTest").setMaster("local[*]")
val sc: SparkContext = new SparkContext(conf)
val line: RDD[String] = sc.textFile("hdfs://hadoop01:9000/testdate/e.txt")
val r: RDD[(String, String, Any, String)] = line.map(word => {
val hostname: String = InetAddress.getLocalHost.getHostName
val threadName: String = Thread.currentThread().getName
(hostname, threadName, infos.resultMap.getOrElse(word, "have not this word"), infos.toString)
})
r.saveAsTextFile("hdfs://hadoop01:9000/testdate/output")
sc.stop()
}
}
package com.ser
import java.net.InetAddress
object getComputerInfo extends Serializable {
val resultMap: Map[String, Double] = Map("hadoop"->2.7,"spark"->2.2)
}