标签:blog io os ar java for sp div 2014
过程,
Spout 发送msgid 1-10
一级Bolt, msgid1的tuple做为基本组合tuple, 其他8个和一组合, 然后发送给二级Bolt, 同时单个msgid对应的tuple都ack一次,msgid1对象tuple, acker将会跟踪8个二级bolt处理情况.
二级Bolt,发送ack fail(模拟处理失败)
结果:在spout fail下出现msg1-9都失败的情况 .
拓扑代码
package storm.starter; import backtype.storm.Config; import backtype.storm.LocalCluster; import backtype.storm.LocalDRPC; import backtype.storm.StormSubmitter; import backtype.storm.drpc.DRPCSpout; import backtype.storm.task.OutputCollector; import backtype.storm.task.ShellBolt; import backtype.storm.task.TopologyContext; import backtype.storm.topology.BasicOutputCollector; import backtype.storm.topology.IRichBolt; import backtype.storm.topology.OutputFieldsDeclarer; import backtype.storm.topology.TopologyBuilder; import backtype.storm.topology.base.BaseBasicBolt; import backtype.storm.tuple.Fields; import backtype.storm.tuple.Tuple; import backtype.storm.tuple.Values; import storm.starter.spout.RandomSentenceSpout; import java.lang.management.ManagementFactory; import java.util.ArrayList; import java.util.HashMap; import java.util.List; import java.util.Map; import org.apache.log4j.Logger; import org.apache.log4j.PropertyConfigurator; /** * This topology demonstrates Storm‘s stream groupings and multilang * capabilities. */ public class WordCountTopology { public static String GetThreadName() { Thread thread = Thread.currentThread(); return thread.getName(); } public static final Logger logger = Logger .getLogger(WordCountTopology.class); // 切分单词 一级bolt /* * public static class SplitSentence extends ShellBolt implements IRichBolt * { public SplitSentence() { super("python", "splitsentence.py"); * logger.error(GetThreadName() + "SplitSentence create"); } * * // 定义字段发送 * * @Override public void declareOutputFields(OutputFieldsDeclarer declarer) * { declarer.declare(new Fields("word")); logger.error(GetThreadName() + * "declarer.declare(new Fields(word))"); } * * @Override public Map<String, Object> getComponentConfiguration() { * logger.error("getComponentConfiguration"); return null; } } */ public static class SplitSentence implements IRichBolt { private OutputCollector _collector; int num = 0; @Override public void prepare(Map stormConf, TopologyContext context, OutputCollector collector) { _collector = collector; } private Tuple tuple1; @Override public void execute(Tuple tuple) { String sentence = tuple.getString(0); if(sentence.equals("a")) { tuple1 = tuple; } else{ List<Tuple> anchors = new ArrayList<Tuple>(); anchors.add(tuple1); anchors.add(tuple); _collector.emit(anchors, new Values(sentence + "a")); _collector.ack(tuple); _collector.ack(tuple1); } // for (String word : sentence.split(" ")){ // _collector.emit(tuple, new Values(word)); // } // num++; System.out.println("Bolt Thread " + Thread.currentThread().getName() + "recve : " + sentence); System.out.println( num + " bolt recev:" + tuple.getMessageId().getAnchorsToIds()); } @Override public void cleanup() { } @Override public void declareOutputFields(OutputFieldsDeclarer declarer) { declarer.declare(new Fields("word")); } @Override public Map<String, Object> getComponentConfiguration() { // TODO Auto-generated method stub return null; } } public static class CountCount1 implements IRichBolt { Map<String, Integer> counts = new HashMap<String, Integer>(); private OutputCollector _collector; int num = 0; @Override public void execute(Tuple tuple) { String word = tuple.getString(0); //logger.error(this.toString() + "word = " + word); Integer count = counts.get(word); if (count == null) count = 0; count++; counts.put(word, count); num++; _collector.fail(tuple); //_collector.ack(tuple); //_collector.emit(tuple, new Values(word, count)); } @Override public void declareOutputFields(OutputFieldsDeclarer declarer) { // logger.error("declareOutputFields :"); declarer.declare(new Fields("result", "count")); } @Override public void prepare(Map stormConf, TopologyContext context, OutputCollector collector) { // TODO Auto-generated method stub _collector = collector; } @Override public void cleanup() { // TODO Auto-generated method stub } @Override public Map<String, Object> getComponentConfiguration() { // TODO Auto-generated method stub return null; } } public static class WordCount extends BaseBasicBolt { private OutputCollector _collector; Map<String, Integer> counts = new HashMap<String, Integer>(); @Override public void execute(Tuple tuple, BasicOutputCollector collector) { String word = tuple.getString(0); //logger.error(this.toString() + "word = " + word); Integer count = counts.get(word); if (count == null) count = 0; count++; counts.put(word, count); // <key, list<value, count> > //logger.error(this.toString() + "count = " + count); collector.emit(new Values(word, count)); } @Override public void declareOutputFields(OutputFieldsDeclarer declarer) { // logger.error("declareOutputFields :"); declarer.declare(new Fields("result", "count")); } } public static class WordCount1 extends BaseBasicBolt { Map<String, Integer> counts = new HashMap<String, Integer>(); @Override public void execute(Tuple tuple, BasicOutputCollector collector) { // logger.error("WordCount1"); // tuple.getFields()[0]; if (tuple.getFields().contains("result")) { String count = (String) tuple.getValueByField("result"); // tuple.getValueByField(field) long countl = -0;// = Long.valueOf(count); // logger.error(this.toString() + " key = resultkey " + count); } if (tuple.getFields().contains("count")) { Integer count = (Integer) tuple.getValueByField("count"); // tuple.getValueByField(field) long countl = -0;// = Long.valueOf(count); //logger.error(this.toString() + " key = count " + count); } // String word = tuple.getString(0); // logger.error(this.toString() +"word = " + word); // Integer count = counts.get(word); // if (count == null) // count = 0; // count++; // counts.put(word, count); // logger.error(this.toString() + "count = " + count); // collector.emit(new Values(word, count)); } @Override public void declareOutputFields(OutputFieldsDeclarer declarer) { // logger.error("declareOutputFields :"); declarer.declare(new Fields("word1", "count1")); } } public static void main(String[] args) throws Exception { TopologyBuilder builder = new TopologyBuilder(); PropertyConfigurator .configure("/home/hadoop/code1/Kafka/src/Log4j.properties"); // parallelism_hint 代表是executor数量, setNumTasks 代表Tasks数量 builder.setSpout("spout", new RandomSentenceSpout(), 5).setNumTasks(2); builder.setBolt("split", new SplitSentence(), 8).setNumTasks(1).shuffleGrouping("spout"); builder.setBolt("count", new CountCount1(), 12).fieldsGrouping("split", new Fields("word")); // builder.setBolt("WordCount1", new WordCount1(), 1).fieldsGrouping( // "count", new Fields("result", "count")); Config conf = new Config(); conf.setDebug(true);
// 这个设置一个spout task上面最多有多少个没有处理(ack/fail)的tuple,防止tuple队列过大, 只对可靠任务起作用 conf.setMaxSpoutPending(2); conf.setMessageTimeoutSecs(5); // 消息处理延时 conf.setNumAckers(2); // 消息处理acker System.out.println("args.length = " + args.length); if (args != null && args.length > 0) { conf.setNumWorkers(5); // 设置工作进程 StormSubmitter.submitTopology(args[0], conf, builder.createTopology()); } else { // 每个组件的最大executor数 conf.setMaxTaskParallelism(1); conf.setDebug(true); LocalCluster cluster = new LocalCluster(); cluster.submitTopology("word-count", conf, builder.createTopology()); String str = "testdrpc"; // drpc.execute("testdrpc", str); Thread.sleep(1088000); cluster.shutdown(); } } }
spout代码
package storm.starter.spout; import backtype.storm.spout.SpoutOutputCollector; import backtype.storm.task.TopologyContext; import backtype.storm.topology.OutputFieldsDeclarer; import backtype.storm.topology.base.BaseRichSpout; import backtype.storm.tuple.Fields; import backtype.storm.tuple.Values; import backtype.storm.utils.Utils; import java.util.Map; import java.util.Random; import org.apache.log4j.Logger; import storm.starter.WordCountTopology; // IRichSpout public class RandomSentenceSpout extends BaseRichSpout { SpoutOutputCollector _collector; Random _rand; public static final Logger logger = Logger .getLogger(RandomSentenceSpout.class); @Override public void open(Map conf, TopologyContext context, SpoutOutputCollector collector) { _collector = collector; _rand = new Random(); WordCountTopology.logger.error(this.toString() + "RandomSentenceSpout is create"); } private int num = 0; private String gettmstr() { StringBuilder tmp = new StringBuilder(); for (int i = 0; i <= num; i++) tmp.append("a"); num++; return tmp.toString(); } @Override public void nextTuple() { Utils.sleep(200); // String[] sentences = new String[]{ "the cow jumped over the moon", // "an apple a day keeps the doctor away", // "four score and seven years ago", "snow white and the seven dwarfs", // "i am at two with nature" }; String[] sentences = new String[] { "A" }; String sentence = gettmstr(); // sentences[_rand.nextInt(sentences.length)]; if (num < 10) { _collector.emit(new Values(sentence), new Integer(num)); // logger.error(this.toString() + "send sentence = " + sentence); // System.out.println(Thread.currentThread().getName() + " Spout "); } } @Override public void ack(Object id) { logger.error(this.toString() + "spout ack =" + (Integer)id); } @Override public void fail(Object id) { logger.error("spout fail =" + (Integer)id); } @Override public void declareOutputFields(OutputFieldsDeclarer declarer) { declarer.declare(new Fields("word")); } }
运行结果
2014-10-03 21:17:31,149 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =1 2014-10-03 21:17:31,351 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =2 Bolt Thread Thread-22recve : aaa 0 bolt recev:{-3139141336114052337=7131499433188364504} 2014-10-03 21:17:31,552 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =3 Bolt Thread Thread-22recve : aaaa 0 bolt recev:{-4497680640148241887=-615828348570847097} 2014-10-03 21:17:31,754 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =4 Bolt Thread Thread-22recve : aaaaa 0 bolt recev:{-8878772617767839827=-7708082520013359311} 2014-10-03 21:17:31,957 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =5 Bolt Thread Thread-22recve : aaaaaa 0 bolt recev:{-3995020874692495577=-5070846720162762196} 2014-10-03 21:17:32,160 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =6 Bolt Thread Thread-22recve : aaaaaaa 0 bolt recev:{-5994700617723404155=-3738685841476816404} 2014-10-03 21:17:32,362 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =7 Bolt Thread Thread-22recve : aaaaaaaa 0 bolt recev:{-2308734827213127682=-5719708045753233056} 2014-10-03 21:17:32,563 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =8 Bolt Thread Thread-22recve : aaaaaaaaa 0 bolt recev:{-3718844156917119468=-6359724009048981605} 2014-10-03 21:17:32,766 ERROR (storm.starter.spout.RandomSentenceSpout:71) - spout fail =9
标签:blog io os ar java for sp div 2014
原文地址:http://www.cnblogs.com/chengxin1982/p/4005359.html