spark接hbase读取数据:
val sc = new SparkContext(sparkConf)
val conf = HBaseConfiguration.create()
conf.set("hbase.zookeeper.property.clientPort", "port..")
conf.set("hbase.zookeeper.quorum", "ip..")
conf.set(TableInputFormat.INPUT_TABLE,"table1..")
val hBaseRDD = sc.newAPIHadoopRDD(conf, classOf[TableInputFormat],
classOf[org.apache.hadoop.hbase.io.ImmutableBytesWritable],
classOf[org.apache.hadoop.hbase.client.Result])
val count = hBaseRDD.count()
println("HBase RDD Count:" + count)
hBaseRDD.cache()
//遍历输出
hBaseRDD.collect().foreach{ case (_,result) =>
val key = Bytes.toInt(result.getRow)
val name = Bytes.toString(result.getValue("pd".getBytes,"name".getBytes))
val age = Bytes.toInt(result.getValue("pd".getBytes,