data数据源,请参考我的博客http://blog.csdn.net/hadoop_spark_storm/article/details/53412598
import org.apache.spark.sql.DataFrameStatFunctions
查看字段中频繁元素的集合
val colArray1=Array("affairs", "gender", "age", "yearsmarried")
data.stat.freqItems(colArray1).show(10,truncate=false)
+-------------------------------+----------------+------------------------------------------------------+-----------------------------------------------+
|affairs_freqItems |gender_freqItems|age_freqItems |yearsmarried_freqItems |
+-------------------------------+----------------+-------------------------------------------------