Github user jerrypeng commented on a diff in the pull request:
https://github.com/apache/storm/pull/934#discussion_r47741484
--- Diff:
examples/storm-starter/src/jvm/storm/starter/BlobStoreAPIWordCountTopology.java
---
@@ -0,0 +1,288 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package storm.starter;
+
+import backtype.storm.Config;
+import backtype.storm.StormSubmitter;
+import backtype.storm.blobstore.AtomicOutputStream;
+import backtype.storm.blobstore.ClientBlobStore;
+import backtype.storm.blobstore.InputStreamWithMeta;
+import backtype.storm.blobstore.NimbusBlobStore;
+
+import backtype.storm.generated.AccessControl;
+import backtype.storm.generated.AccessControlType;
+import backtype.storm.generated.AlreadyAliveException;
+import backtype.storm.generated.AuthorizationException;
+import backtype.storm.generated.InvalidTopologyException;
+import backtype.storm.generated.KeyAlreadyExistsException;
+import backtype.storm.generated.KeyNotFoundException;
+import backtype.storm.generated.SettableBlobMeta;
+import backtype.storm.spout.SpoutOutputCollector;
+import backtype.storm.task.ShellBolt;
+import backtype.storm.task.TopologyContext;
+import backtype.storm.topology.BasicOutputCollector;
+import backtype.storm.topology.IRichBolt;
+import backtype.storm.topology.OutputFieldsDeclarer;
+import backtype.storm.topology.TopologyBuilder;
+import backtype.storm.topology.base.BaseBasicBolt;
+import backtype.storm.topology.base.BaseRichSpout;
+import backtype.storm.blobstore.BlobStoreAclHandler;
+import backtype.storm.tuple.Fields;
+import backtype.storm.tuple.Tuple;
+import backtype.storm.tuple.Values;
+import backtype.storm.utils.Utils;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.io.BufferedReader;
+import java.io.BufferedWriter;
+import java.io.File;
+import java.io.FileReader;
+import java.io.FileWriter;
+import java.io.IOException;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.Random;
+import java.util.Set;
+import java.util.StringTokenizer;
+
+public class BlobStoreAPIWordCountTopology {
+ private static ClientBlobStore store; // Client API to invoke blob
store API functionality
+ private static String key = "key";
+ private static String fileName = "blacklist.txt";
+ private static final Logger LOG =
LoggerFactory.getLogger(BlobStoreAPIWordCountTopology.class);
+
+ public static void prepare() {
+ Config conf = new Config();
+ conf.putAll(Utils.readStormConfig());
+ store = Utils.getClientBlobStore(conf);
+ }
+
+ // Spout implementation
+ public static class RandomSentenceSpout extends BaseRichSpout {
+ SpoutOutputCollector _collector;
+
+ @Override
+ public void open(Map conf, TopologyContext context,
SpoutOutputCollector collector) {
+ _collector = collector;
+ }
+
+ @Override
+ public void nextTuple() {
+ Utils.sleep(100);
+ _collector.emit(new Values(getRandomSentence()));
+ }
+
+ @Override
+ public void ack(Object id) {
+ }
+
+ @Override
+ public void fail(Object id) {
+ }
+
+ @Override
+ public void declareOutputFields(OutputFieldsDeclarer declarer) {
+ declarer.declare(new Fields("sentence"));
+ }
+
+ }
+
+ // Bolt implementation
+ public static class SplitSentence extends ShellBolt implements
IRichBolt {
+
+ public SplitSentence() {
+ super("python", "splitsentence.py");
+ }
+
+ @Override
+ public void declareOutputFields(OutputFieldsDeclarer declarer) {
+ declarer.declare(new Fields("word"));
+ }
+
+ @Override
+ public Map<String, Object> getComponentConfiguration() {
+ return null;
+ }
+ }
+
+ public static class FilterWords extends BaseBasicBolt {
+ String fileName = "blacklist.txt";
+ @Override
+ public void execute(Tuple tuple, BasicOutputCollector collector) {
+ try {
+ String word = tuple.getString(0);
+ Set<String> wordSet = parseFile(fileName);
+ if (!wordSet.contains(word)) {
+ collector.emit(new Values(word));
+ }
+ } catch (IOException exp) {
+ throw new RuntimeException(exp);
+ }
+ }
+
+ @Override
+ public void declareOutputFields(OutputFieldsDeclarer declarer) {
+ declarer.declare(new Fields("word"));
+ }
+ }
+
+ public void buildAndLaunchWordCountTopology(String[] args) {
+ TopologyBuilder builder = new TopologyBuilder();
+ builder.setSpout("spout", new RandomSentenceSpout(), 5);
+ builder.setBolt("split", new SplitSentence(),
8).shuffleGrouping("spout");
+ builder.setBolt("filter", new FilterWords(),
6).shuffleGrouping("split");
+
+ Config conf = new Config();
+ conf.setDebug(true);
+ try {
+ conf.setNumWorkers(3);
+ StormSubmitter.submitTopologyWithProgressBar(args[0], conf,
builder.createTopology());
+ } catch (InvalidTopologyException | AuthorizationException |
AlreadyAliveException exp) {
+ throw new RuntimeException(exp);
+ }
+ }
+
+ // Equivalent create command on command line
+ // storm blobstore create --file blacklist.txt --acl o::rwa key
+ private static void createBlobWithContent(String blobKey,
ClientBlobStore clientBlobStore, File file)
+ throws AuthorizationException, KeyAlreadyExistsException,
IOException,KeyNotFoundException {
+ String stringBlobACL = "o::rwa";
+ AccessControl blobACL =
BlobStoreAclHandler.parseAccessControl(stringBlobACL);
+ List<AccessControl> acls = new LinkedList<AccessControl>();
+ acls.add(blobACL); // more ACLs can be added here
+ SettableBlobMeta settableBlobMeta = new SettableBlobMeta(acls);
+ AtomicOutputStream blobStream =
clientBlobStore.createBlob(blobKey,settableBlobMeta);
+ blobStream.write(readFile(file).toString().getBytes());
+ blobStream.close();
+ }
+
+ // Equivalent update command on command line
+ // storm blobstore update --file blacklist.txt key
+ private static void updateBlobWithContent(String blobKey,
ClientBlobStore clientBlobStore, File file)
+ throws KeyNotFoundException, AuthorizationException,
IOException {
+ AtomicOutputStream blobOutputStream =
clientBlobStore.updateBlob(blobKey);
+ blobOutputStream.write(readFile(file).toString().getBytes());
+ blobOutputStream.close();
+ }
+
+ private static String getRandomSentence() {
+ String[] sentences = new String[]{ "the cow jumped over the moon",
"an apple a day keeps the doctor away",
+ "four score and seven years ago", "snow white and the
seven dwarfs", "i am at two with nature" };
+ String sentence = sentences[new
Random().nextInt(sentences.length)];
+ return sentence;
+ }
+
+ private static Set<String> getRandomWordSet() {
+ Set<String> randomWordSet = new HashSet<>();
+ Random random = new Random();
+ String[] words = new String[]{ "cow", "jumped", "over", "the",
"moon", "apple", "day", "doctor", "away",
+ "four", "seven", "ago", "snow", "white", "seven",
"dwarfs", "nature", "two" };
+ // Choosing atmost 5 words to update the blacklist file for
filtering
+ for (int i=0; i<5; i++) {
+ randomWordSet.add(words[random.nextInt(words.length)]);
+ }
+ return randomWordSet;
+ }
+
+ private static Set<String> parseFile(String fileName) throws
IOException {
+ File file = new File(fileName);
+ Set<String> wordSet = new HashSet<>();
+ if (!file.exists()) {
+ return wordSet;
+ }
+ StringTokenizer tokens = new
StringTokenizer(readFile(file).toString(), "\r\n");
+ while (tokens.hasMoreElements()) {
+ wordSet.add(tokens.nextToken());
+ }
+ LOG.info("parseFile {}", wordSet);
--- End diff --
perhaps debug?
---
If your project is set up for it, you can reply to this email and have your
reply appear on GitHub as well. If your project does not have this feature
enabled and wishes so, or if the feature is enabled but not working, please
contact infrastructure at [email protected] or file a JIRA ticket
with INFRA.
---