-----读取本地打印输出
val text=sc.textFile("hdfs://172.22.241.183:8020/user/spark/yzg_test.txt")
sc.textFile("hdfs://172.22.241.183:8020/user/spark/yzg_test.txt").flatMap(_.split(" ")).map((_,1)).reduceByKey(_+_).collect
--spark-shell实现文件流
import org.apache.spark.streaming._
val ssc = new StreamingContext(sc, Seconds(5))
val lines = ssc.textFileStream("hdfs://172.22.241.183:8020/user/spark/yzg_test.txt")
val Counts = lines.flatMap(_.split(" ")).map((_,1)).reduceByKey(_ + _)
Counts.saveAsTextFiles("hdfs://172.22.241.183:8020/user/spark/bendi-test")
ssc.start()
ssc.awaitTermination()
--spa