话不多说,直接上码
package com.myspark.wordcount.java_version;
import org.apache.spark.SparkConf;
import org.apache.spark.api.java.JavaPairRDD;
import org.apache.spark.api.java.JavaRDD;
import org.apache.spark.api.java.JavaSparkContext;
import org.apache.spark.api.java.function.FlatMapFunction;
import org.apache.spark.api.java.function.Function2;
import org.apache.spark.api.java.function.PairFunction;
import scala.Tuple2;
import java.util.Arrays;
import java.util.Iterator;
import java.util.List;
/**
*
* @Author:Yh.Huang Date:2019/4/26
* Desc:利用java语言开发spark wordcount程序
*/
public class WordCount_Java {
public static void main(String[] args) {
//1.创建sparkConf对象
SparkConf conf = new SparkConf().setAppName("WordCount_Java").setMaster("local[2]");
//2.创建sparkcontext对象
JavaSparkContext jsc = new JavaSparkContext(conf);
jsc.setLogLevel("WARN");
//3.读取数据文件
JavaRDD<String&