package com.bawei
import org.apache.flink.configuration.Configuration
import org.apache.flink.streaming.api.functions.sink.{RichSinkFunction, SinkFunction}
import org.apache.flink.streaming.api.scala._
import org.apache.hadoop.conf
import org.apache.hadoop.hbase.{HBaseConfiguration, HConstants, TableName}
import org.apache.hadoop.hbase.client._
import org.apache.hadoop.hbase.util.Bytes
case class SensorReading(id: String, timeStamp: Long, temperature: Double)
object hbasesink {
def main(args: Array[String]): Unit = {
val env = StreamExecutionEnvironment.getExecutionEnvironment
env.setParallelism(1)
val stream = env.readTextFile(“C:\Intel\day0411\src\main\resources\sensor.txt”)
val dataStream: DataStream[SensorReading] = stream.map(data => {
val dataArr = data.split(",")
SensorReading(dataArr(0).trim, dataArr(1).trim.toLong, dataArr(2).trim.toDouble)
})
dataStream.addSink(new MyHBaseSink())
env.execute()
}
}
class MyHBaseSink()extends RichSinkFunction[SensorReading]{
//创建连接
var conn:Connection=_
//创建BufferedMutator
//作用类似put,可以实现批量异步
var mutator:BufferedMutator=null
//初始化
override def open(parameters: Configuration): Unit = {
val conf = HBaseConfiguration.create()
//设置zookeeper的主机名称
conf.set(HConstants.ZOOKEEPER_QUORUM,“hdp111,hdp222,hdp333”)
//设置端口号
conf.set(HConstants.ZOOKEEPER_CLIENT_PORT,“2181”)
//实例化
conn = ConnectionFactory.createConnection(conf)
val tableName = TableName.valueOf(“student”)
val params = new BufferedMutatorParams(tableName)
params.writeBufferSize(1024*1024)
mutator = conn.getBufferedMutator(params)
}
override def invoke(value: SensorReading, context: SinkFunction.Context[_]): Unit = {
val family=“info”
val put = new Put(Bytes.toBytes(value.id))
put.addColumn(Bytes.toBytes(family), Bytes.toBytes(“eventTime”), Bytes.toBytes(value.timeStamp.toString))
put.addColumn(Bytes.toBytes(family), Bytes.toBytes(“temp”), Bytes.toBytes(value.temperature.toString))
mutator.mutate(put)
}
override def close(): Unit = {
mutator.close()
conn.close()
}
}