package project2Storm; import org.apache.hadoop.hbase.client.Durability; import org.apache.storm.Config; import org.apache.storm.LocalCluster; import org.apache.storm.StormSubmitter; import org.apache.storm.generated.StormTopology; import org.apache.storm.hbase.bolt.mapper.HBaseProjectionCriteria; import org.apache.storm.hbase.bolt.mapper.HBaseValueMapper; import org.apache.storm.hbase.trident.mapper.SimpleTridentHBaseMapMapper; import org.apache.storm.hbase.trident.mapper.SimpleTridentHBaseMapper; import org.apache.storm.hbase.trident.mapper.TridentHBaseMapper; import org.apache.storm.hbase.trident.state.HBaseState; import org.apache.storm.hbase.trident.state.HBaseStateFactory; import org.apache.storm.hbase.trident.state.HBaseUpdater; import org.apache.storm.kafka.BrokerHosts; import org.apache.storm.kafka.StringScheme; import org.apache.storm.kafka.ZkHosts; import org.apache.storm.kafka.trident.OpaqueTridentKafkaSpout; import org.apache.storm.kafka.trident.TridentKafkaConfig; import org.apache.storm.spout.SchemeAsMultiScheme; import org.apache.storm.topology.base.BaseWindowedBolt; import org.apache.storm.trident.TridentTopology; import org.apache.storm.trident.operation.BaseAggregator; import org.apache.storm.trident.operation.BaseFunction; import org.apache.storm.trident.operation.TridentCollector; import org.apache.storm.trident.state.StateFactory; import org.apache.storm.trident.testing.Split; import org.apache.storm.trident.tuple.TridentTuple; import org.apache.storm.trident.windowing.config.SlidingDurationWindow; import org.apache.storm.trident.windowing.config.WindowConfig; import org.apache.storm.tuple.Fields; import org.apache.storm.tuple.Values; import java.sql.Connection; import java.sql.DriverManager; import java.sql.SQLException; import java.sql.Statement; import java.util.HashMap; import java.util.List; import java.util.Map; import java.util.concurrent.TimeUnit; import java.util.Date; import java.text.SimpleDateFormat; /** * Created by Dank 2017/7/17. */ public class wind { public static class Split2 extends BaseFunction { @Override public void execute(TridentTuple tuple, TridentCollector collector) { String line = tuple.getValueByField("str").toString(); String result[]=line.split(",&#
kafka+storm+Hbase流式数据处理
最新推荐文章于 2024-01-28 14:32:45 发布
该博客介绍了如何使用Kafka、Storm和HBase构建一个实时数据处理系统。通过Kafka作为消息中间件,Storm进行实时流处理,HBase作为数据存储,实现了数据的高效流转和存储。博客中详细展示了如何配置TridentKafkaSpout、TridentHBaseMapper和窗口操作,以确保数据的正确读取、处理和写入。
摘要由CSDN通过智能技术生成