Java spark 调用spark 算子 parallelizePairs 报错:
Incompatible equality constraint: String and T1
List<Tuple2<String, Double>> arr2 = Arrays.asList(
new Tuple2<String, Double>("u1", 20.01),
new Tuple2<String, Double>("u2", 18.95),
new Tuple2<String, Double>("u3", 20.55),
new Tuple2<String, Double>("u4", 20.12),
new Tuple2<String, Double>("u5", 100.11)
);
JavaPairRDD<String, Double> rdd2 = jsc.parallelizePairs(arr2);
代码修改成如下:
JavaPairRDD<String, Double> rdd2 = jsc.<String, Double>parallelizePairs(arr2);