storm简单实例+如何将storm任务迁移至flink处理研究

12 篇文章 1 订阅
4 篇文章 0 订阅
本文探讨了如何将基础的storm流处理程序转化为flink-storm和纯flink实现。首先展示了storm的基本操作,然后详细说明了如何将storm任务迁移到flink-storm,并替换相关依赖。最后,介绍了直接使用flink来处理相似任务的方法,包括从Kafka读取和处理数据。
摘要由CSDN通过智能技术生成

基础storm程序示例

Storm的流处理主要就是通过Spout和Bolt节点进行处理,可以继承这些类写自己的逻辑

public class FlinkStormDemo {

    public static void main(String[] args) {
        //1.创建执行环境
        LocalCluster stormCluster = new LocalCluster();
        TopologyBuilder builder = new TopologyBuilder();
        //2.创建一个初始的数据源
        builder.setSpout("word", new WordSpout());
        //3.对数据源进行第一次加工
        builder.setBolt("word-1",new WordBolt1(), 1).shuffleGrouping("word");
        //4.对数据源进行第二次加工
        builder.setBolt("word-2",new WordBolt2(), 1).shuffleGrouping("word-1");
        //5.配置一些参数
        Config config = new Config();
        config.setDebug(true);
        //6.提交storm任务并处理
        stormCluster.submitTopology("storm-task", config, builder.createTopology());
    }

    static class WordSpout extends BaseRichSpout {

        private SpoutOutputCollector spoutOutputCollector;

        @Override
        public void open(Map map, TopologyContext topologyContext, SpoutOutputCollector spoutOutputCollector) {
            this.spoutOutputCollector = spoutOutputCollector;
        }

        @Override
        public void nextTuple() {
            try {
                Thread.sleep(10000);
            } catch (InterruptedException e) {
                e.printStackTrace();
            }
            System.out.println("数据初始化中....");
            String initData = "abc";
            spoutOutputCollector.emit(new Values(initData));
        }

        @Override
        public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {
            outputFieldsDeclarer.declare(new Fields("word"));
        }
    }

    static class WordBolt1 extends BaseRichBolt {

        private OutputCollector collector;

        @Override
        public void prepare(Map map, TopologyContext topologyContext, OutputCollector outputCollector) {
            this.collector = outputCollector;
        }

        @Override
        public void execute(Tuple tuple) {
            System.out.println("数据第1次处理中....");
            //给上次获取的单词拼接上def
            collector.emit(tuple, new Values(tuple.getString(0) + "def"));
            collector.ack(tuple);
        }

        @Override
        public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {
            outputFieldsDeclarer.declare(new Fields("word"));
        }
    }


    static class WordBolt2 extends BaseRichBolt {

        private OutputCollector collector;

        @Override
        public void prepare(Map map, TopologyContext topologyContext, OutputCollector outputCollector) {
            this.collector = outputCollector;
        }

        @Override
        public void execute(Tuple tuple) {
            System.out.println("数据第2次处理中....");
            //输出处理结果
            System.out.println("处理结果:" + tuple.getString(0));
            collector.ack(tuple);
        }

        @Override
        public void declareOutputFields(OutputFieldsDeclarer outputFieldsDeclarer) {

        }
    }
}

执行结果:

在这里插入图片描述

利用flink-storm程序实现类似功能

需要更改flink相关依赖的版本到1.7.0,主要依赖了flink-storm的jar包

	<dependency>
      <groupId>org.apache.flink</groupId>
      <artifactId>flink-java</artifactId>
      <version>1.7.0</version>
    </dependency>
    <dependency>
      <groupId>org.apache.flink</groupId>
      <artifactId>flink-streaming-java_2.11</artifactId>
      <version>1.7.0</version>
    </dependency>
    <dependency>
      <groupId>org.apache.flink</groupId>
      <artifactId>flink-clients_2.11</artifactId>
      <version>1.7.0</version>
    </dependency>
    <dependency>
      <groupId>org.apache.flink</groupId>
      <artifactId>flink-storm_2.11</artifactId>
      <version>1.7.0</version>
    </dependency>

只需要修改两处:

LocalCluster替换为FlinkLocalCluster,处理的任务从TopologyBuilder

.createTopology替换为FlinkTopology.createTopology(TopologyBuilder)

在这里插入图片描述

执行结果:
在这里插入图片描述

利用flink程序实现类似功能

利用Kafka发送初始消息“测试数据”。

public class FlinkProducer {

    public static void main(String[] args) throws Exception {
        StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
        Properties properties = new Properties();
        properties.put("bootstrap.servers", "10.225.173.107:9092,10.225.173.108:9092,10.225.173.109:9092");
        FlinkKafkaProducer<String> flinkKafkaProducer = new FlinkKafkaProducer<>("flink", new SimpleStringSchema(), properties);
        DataStreamSource<String> source = env.fromElements("测试数据");
        source.addSink(flinkKafkaProducer);
        env.execute();
    }

}

接收Kafka的初始消息“测试数据”并加工处理。

public class FlinkConsumer {

    public static void main(String[] args) throws Exception {
        //1.创建执行环境
        StreamExecutionEnvironment env = StreamExecutionEnvironment.getExecutionEnvironment();
        //2.配置并创建初始数据源
        Properties properties = new Properties();
        properties.put("bootstrap.servers", "10.225.173.107:9092,10.225.173.108:9092,10.225.173.109:9092");
        FlinkKafkaConsumer<String> flinkKafkaConsumer = new FlinkKafkaConsumer<>("flink", new SimpleStringSchema(), properties);
        DataStreamSource<String> source = env.addSource(flinkKafkaConsumer);
        //3.对数据源进行连续处理
        source.process(new FlinkBolt1()).
               process(new FlinkBolt2()).
               process(new FlinkBolt3());
        //4.执行flink程序
        env.execute();
    }

    static class FlinkBolt1 extends ProcessFunction<String, Object> {

        @Override
        public void open(Configuration parameters) {
            //开始第1次处理....
        }

        @Override
        public void processElement(String s, ProcessFunction<String, Object>.Context context, Collector<Object> collector) throws Exception {
            System.out.println("第1次处理前的值是:" + s);
            s += "abc";
            System.out.println("第1次处理后的值是:" + s);
            collector.collect(s);
        }

        @Override
        public void close() throws Exception {
            //结束第1次处理
        }
    }

    static class FlinkBolt2 extends ProcessFunction<Object, Object> {

        @Override
        public void open(Configuration parameters) {
            //开始第2次处理....
        }

        @Override
        public void processElement(Object s, ProcessFunction<Object, Object>.Context context, Collector<Object> collector) throws Exception {
            s = s.toString();
            System.out.println("第2次处理前的值是:" + s);
            s += "def";
            System.out.println("第2次处理后的值是:" + s);
            collector.collect(s);
        }

        @Override
        public void close() throws Exception {
            //结束第2次处理....
        }
    }

    static class FlinkBolt3 extends ProcessFunction<Object, Object> {

        @Override
        public void open(Configuration parameters) {
            //开始第3次处理....
        }

        @Override
        public void processElement(Object s, ProcessFunction<Object, Object>.Context context, Collector<Object> collector) throws Exception {
            s = s.toString();
            System.out.println("第3次处理前的值是:" + s);
            s += "ghi";
            System.out.println("第3次处理后的值是:" + s);
            collector.collect(s);
        }

        @Override
        public void close() throws Exception {
            //结束第3次处理....
        }
    }

}

处理结果:
在这里插入图片描述

评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值