strom wordcount java 实现案例

1,spout编写,读取文件内容:

package com.storm.test;

import java.io.BufferedReader;
import java.io.FileReader;
import java.util.Map;

import backtype.storm.spout.SpoutOutputCollector;
import backtype.storm.task.TopologyContext;
import backtype.storm.topology.OutputFieldsDeclarer;
import backtype.storm.topology.base.BaseRichSpout;
import backtype.storm.tuple.Fields;
import backtype.storm.tuple.Values;

public class WordReader extends BaseRichSpout{
	
	private SpoutOutputCollector collector;
	private FileReader fileReader;
	private String filePath;
	private boolean completed = false;
	
	
	
	@Override
	public void ack(Object msgId) {
		System.out.println("msgId === "+msgId);
	}
	

	@Override
	public void close() {
	}


	@Override
	public void fail(Object msgId) {
		
		System.out.println("fail === "+msgId);
	}


	@Override
	public void open(Map conf, TopologyContext context,
			SpoutOutputCollector collector) {
		
		try {
			this.fileReader = new FileReader(conf.get("wordFile").toString());
			
		} catch (Exception e) {
			e.printStackTrace();
		}
		this.filePath = conf.get("wordFile").toString();
		this.collector = collector;
		
	}

	@Override
	public void nextTuple() {
		if(completed){
			try {
				Thread.sleep(1000);
			} catch (Exception e) {
				e.printStackTrace();
			}
			return ;
			
		}
		String str;
		BufferedReader reader = new BufferedReader(fileReader);
		try {
			
			while ((str = reader.readLine()) != null){
				System.out.println("read line = "+str);
				this.collector.emit(new Values(str),str);
				System.out.println("WordReader spout = "+str);
			}
		} catch (Exception e) {
			e.printStackTrace();
		}finally{
			completed = true;
		}
	}

	@Override
	public void declareOutputFields(OutputFieldsDeclarer declarer) {
		declarer.declare(new Fields("line"));
	}

	 

}

2,Bolt编写实现句子分割:

package com.storm.test;

import java.util.Map;

import backtype.storm.topology.BasicOutputCollector;
import backtype.storm.topology.OutputFieldsDeclarer;
import backtype.storm.topology.base.BaseBasicBolt;
import backtype.storm.tuple.Fields;
import backtype.storm.tuple.Tuple;
import backtype.storm.tuple.Values;

public class WordNormalizer extends BaseBasicBolt{

	@Override
	public void execute(Tuple input, BasicOutputCollector collector) {
		String sentence = input.getString(0);
		String[] words = sentence.split(",");
		System.out.println("reader line = "+sentence);
		for(String word : words){
			if(!word.trim().isEmpty()){
				collector.emit(new Values(word.trim().toUpperCase()));
			}
		}
	}

	@Override
	public void declareOutputFields(OutputFieldsDeclarer declarer) {
		declarer.declare(new Fields("word"));
	}
 
	 

 
}

3,bolt编写是单词统计:

package com.storm.test;

import java.util.HashMap;
import java.util.Map;

import backtype.storm.task.TopologyContext;
import backtype.storm.topology.BasicOutputCollector;
import backtype.storm.topology.OutputFieldsDeclarer;
import backtype.storm.topology.base.BaseBasicBolt;
import backtype.storm.tuple.Fields;
import backtype.storm.tuple.Tuple;

public class WordCount extends BaseBasicBolt{

	Integer id;
	String name;
	Map<String,Integer> counters ;
	
	
	
	@Override
	public void cleanup() {
		System.out.println("word counter :["+name+"-"+id);
		for(Map.Entry<String, Integer> entry : counters.entrySet()){
			System.out.println(entry.getKey()+":"+entry.getValue());
		}
		System.out.println("counter finish!");
	}

	
	
	@Override
	public void prepare(Map stormConf, TopologyContext context) {
		this.counters = new HashMap<String,Integer>();
		this.name = context.getThisComponentId();
		this.id = context.getThisTaskId();
	}



	@Override
	public void execute(Tuple input, BasicOutputCollector collector) {
		String str = input.getString(0);
		System.out.println("wordCounter recever "+str);
		if(!counters.containsKey(str)){
			counters.put(str,1);
		}else{
			Integer c = counters.get(str) + 1;
			counters.put(str,c);
		}
	}

	@Override
	public void declareOutputFields(OutputFieldsDeclarer declarer) {
		declarer.declare(new Fields("word"));
	}

	 
 
}

4,main方法编写:

package com.storm.test;

import backtype.storm.Config;
import backtype.storm.LocalCluster;
import backtype.storm.topology.TopologyBuilder;
import backtype.storm.tuple.Fields;

public class ToplogyTest {
	public static void main(String[] args) throws Exception {
		
		
		TopologyBuilder builder = new TopologyBuilder();
		builder.setSpout("<span style="color:#FF0000;">word-reader</span>",new WordReader());
		builder.setBolt("<span style="color:#33CC00;">word-normalize</span>r",new WordNormalizer()).shuffleGrouping("<span style="color:#FF0000;">word-reader</span>");
		builder.setBolt("word-count",new WordCount(),1).fieldsGrouping("<span style="color:#009900;"><span style="color:#33FF33;">word-normalize</span>r</span>",new Fields("word"));
		Config conf = new Config();
		conf.setDebug(true);
		conf.setNumWorkers(2);
		//word.txt:hello,world,hello,storm,hello,spark,hadoop,hadoop
		conf.put("wordFile", "/home/lixun/word.txt");
//		conf.setDebug(true);
		conf.put(Config.TOPOLOGY_MAX_SPOUT_PENDING, 1);
		
		LocalCluster cluster = new LocalCluster();
		cluster.submitTopology("wordCounterTopology",conf,builder.createTopology());
		
		Thread.sleep(4000);
		
		cluster.killTopology("wordCounterTopology");
		cluster.shutdown();
		
	
		
		
	}

}


  • 0
    点赞
  • 0
    收藏
    觉得还不错? 一键收藏
  • 0
    评论
评论
添加红包

请填写红包祝福语或标题

红包个数最小为10个

红包金额最低5元

当前余额3.43前往充值 >
需支付:10.00
成就一亿技术人!
领取后你会自动成为博主和红包主的粉丝 规则
hope_wisdom
发出的红包
实付
使用余额支付
点击重新获取
扫码支付
钱包余额 0

抵扣说明:

1.余额是钱包充值的虚拟货币,按照1:1的比例进行支付金额的抵扣。
2.余额无法直接购买下载,可以购买VIP、付费专栏及课程。

余额充值