[
https://issues.apache.org/jira/browse/STORM-1373?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15059592#comment-15059592
]
ASF GitHub Bot commented on STORM-1373:
---------------------------------------
Github user jerrypeng commented on a diff in the pull request:
https://github.com/apache/storm/pull/934#discussion_r47741484
--- Diff:
examples/storm-starter/src/jvm/storm/starter/BlobStoreAPIWordCountTopology.java
---
@@ -0,0 +1,288 @@
+/**
+ * Licensed to the Apache Software Foundation (ASF) under one
+ * or more contributor license agreements. See the NOTICE file
+ * distributed with this work for additional information
+ * regarding copyright ownership. The ASF licenses this file
+ * to you under the Apache License, Version 2.0 (the
+ * "License"); you may not use this file except in compliance
+ * with the License. You may obtain a copy of the License at
+ *
+ * http://www.apache.org/licenses/LICENSE-2.0
+ *
+ * Unless required by applicable law or agreed to in writing, software
+ * distributed under the License is distributed on an "AS IS" BASIS,
+ * WITHOUT WARRANTIES OR CONDITIONS OF ANY KIND, either express or implied.
+ * See the License for the specific language governing permissions and
+ * limitations under the License.
+ */
+package storm.starter;
+
+import backtype.storm.Config;
+import backtype.storm.StormSubmitter;
+import backtype.storm.blobstore.AtomicOutputStream;
+import backtype.storm.blobstore.ClientBlobStore;
+import backtype.storm.blobstore.InputStreamWithMeta;
+import backtype.storm.blobstore.NimbusBlobStore;
+
+import backtype.storm.generated.AccessControl;
+import backtype.storm.generated.AccessControlType;
+import backtype.storm.generated.AlreadyAliveException;
+import backtype.storm.generated.AuthorizationException;
+import backtype.storm.generated.InvalidTopologyException;
+import backtype.storm.generated.KeyAlreadyExistsException;
+import backtype.storm.generated.KeyNotFoundException;
+import backtype.storm.generated.SettableBlobMeta;
+import backtype.storm.spout.SpoutOutputCollector;
+import backtype.storm.task.ShellBolt;
+import backtype.storm.task.TopologyContext;
+import backtype.storm.topology.BasicOutputCollector;
+import backtype.storm.topology.IRichBolt;
+import backtype.storm.topology.OutputFieldsDeclarer;
+import backtype.storm.topology.TopologyBuilder;
+import backtype.storm.topology.base.BaseBasicBolt;
+import backtype.storm.topology.base.BaseRichSpout;
+import backtype.storm.blobstore.BlobStoreAclHandler;
+import backtype.storm.tuple.Fields;
+import backtype.storm.tuple.Tuple;
+import backtype.storm.tuple.Values;
+import backtype.storm.utils.Utils;
+import org.slf4j.Logger;
+import org.slf4j.LoggerFactory;
+
+import java.io.BufferedReader;
+import java.io.BufferedWriter;
+import java.io.File;
+import java.io.FileReader;
+import java.io.FileWriter;
+import java.io.IOException;
+import java.util.HashSet;
+import java.util.Iterator;
+import java.util.LinkedList;
+import java.util.List;
+import java.util.Map;
+import java.util.Random;
+import java.util.Set;
+import java.util.StringTokenizer;
+
+public class BlobStoreAPIWordCountTopology {
+ private static ClientBlobStore store; // Client API to invoke blob
store API functionality
+ private static String key = "key";
+ private static String fileName = "blacklist.txt";
+ private static final Logger LOG =
LoggerFactory.getLogger(BlobStoreAPIWordCountTopology.class);
+
+ public static void prepare() {
+ Config conf = new Config();
+ conf.putAll(Utils.readStormConfig());
+ store = Utils.getClientBlobStore(conf);
+ }
+
+ // Spout implementation
+ public static class RandomSentenceSpout extends BaseRichSpout {
+ SpoutOutputCollector _collector;
+
+ @Override
+ public void open(Map conf, TopologyContext context,
SpoutOutputCollector collector) {
+ _collector = collector;
+ }
+
+ @Override
+ public void nextTuple() {
+ Utils.sleep(100);
+ _collector.emit(new Values(getRandomSentence()));
+ }
+
+ @Override
+ public void ack(Object id) {
+ }
+
+ @Override
+ public void fail(Object id) {
+ }
+
+ @Override
+ public void declareOutputFields(OutputFieldsDeclarer declarer) {
+ declarer.declare(new Fields("sentence"));
+ }
+
+ }
+
+ // Bolt implementation
+ public static class SplitSentence extends ShellBolt implements
IRichBolt {
+
+ public SplitSentence() {
+ super("python", "splitsentence.py");
+ }
+
+ @Override
+ public void declareOutputFields(OutputFieldsDeclarer declarer) {
+ declarer.declare(new Fields("word"));
+ }
+
+ @Override
+ public Map<String, Object> getComponentConfiguration() {
+ return null;
+ }
+ }
+
+ public static class FilterWords extends BaseBasicBolt {
+ String fileName = "blacklist.txt";
+ @Override
+ public void execute(Tuple tuple, BasicOutputCollector collector) {
+ try {
+ String word = tuple.getString(0);
+ Set<String> wordSet = parseFile(fileName);
+ if (!wordSet.contains(word)) {
+ collector.emit(new Values(word));
+ }
+ } catch (IOException exp) {
+ throw new RuntimeException(exp);
+ }
+ }
+
+ @Override
+ public void declareOutputFields(OutputFieldsDeclarer declarer) {
+ declarer.declare(new Fields("word"));
+ }
+ }
+
+ public void buildAndLaunchWordCountTopology(String[] args) {
+ TopologyBuilder builder = new TopologyBuilder();
+ builder.setSpout("spout", new RandomSentenceSpout(), 5);
+ builder.setBolt("split", new SplitSentence(),
8).shuffleGrouping("spout");
+ builder.setBolt("filter", new FilterWords(),
6).shuffleGrouping("split");
+
+ Config conf = new Config();
+ conf.setDebug(true);
+ try {
+ conf.setNumWorkers(3);
+ StormSubmitter.submitTopologyWithProgressBar(args[0], conf,
builder.createTopology());
+ } catch (InvalidTopologyException | AuthorizationException |
AlreadyAliveException exp) {
+ throw new RuntimeException(exp);
+ }
+ }
+
+ // Equivalent create command on command line
+ // storm blobstore create --file blacklist.txt --acl o::rwa key
+ private static void createBlobWithContent(String blobKey,
ClientBlobStore clientBlobStore, File file)
+ throws AuthorizationException, KeyAlreadyExistsException,
IOException,KeyNotFoundException {
+ String stringBlobACL = "o::rwa";
+ AccessControl blobACL =
BlobStoreAclHandler.parseAccessControl(stringBlobACL);
+ List<AccessControl> acls = new LinkedList<AccessControl>();
+ acls.add(blobACL); // more ACLs can be added here
+ SettableBlobMeta settableBlobMeta = new SettableBlobMeta(acls);
+ AtomicOutputStream blobStream =
clientBlobStore.createBlob(blobKey,settableBlobMeta);
+ blobStream.write(readFile(file).toString().getBytes());
+ blobStream.close();
+ }
+
+ // Equivalent update command on command line
+ // storm blobstore update --file blacklist.txt key
+ private static void updateBlobWithContent(String blobKey,
ClientBlobStore clientBlobStore, File file)
+ throws KeyNotFoundException, AuthorizationException,
IOException {
+ AtomicOutputStream blobOutputStream =
clientBlobStore.updateBlob(blobKey);
+ blobOutputStream.write(readFile(file).toString().getBytes());
+ blobOutputStream.close();
+ }
+
+ private static String getRandomSentence() {
+ String[] sentences = new String[]{ "the cow jumped over the moon",
"an apple a day keeps the doctor away",
+ "four score and seven years ago", "snow white and the
seven dwarfs", "i am at two with nature" };
+ String sentence = sentences[new
Random().nextInt(sentences.length)];
+ return sentence;
+ }
+
+ private static Set<String> getRandomWordSet() {
+ Set<String> randomWordSet = new HashSet<>();
+ Random random = new Random();
+ String[] words = new String[]{ "cow", "jumped", "over", "the",
"moon", "apple", "day", "doctor", "away",
+ "four", "seven", "ago", "snow", "white", "seven",
"dwarfs", "nature", "two" };
+ // Choosing atmost 5 words to update the blacklist file for
filtering
+ for (int i=0; i<5; i++) {
+ randomWordSet.add(words[random.nextInt(words.length)]);
+ }
+ return randomWordSet;
+ }
+
+ private static Set<String> parseFile(String fileName) throws
IOException {
+ File file = new File(fileName);
+ Set<String> wordSet = new HashSet<>();
+ if (!file.exists()) {
+ return wordSet;
+ }
+ StringTokenizer tokens = new
StringTokenizer(readFile(file).toString(), "\r\n");
+ while (tokens.hasMoreElements()) {
+ wordSet.add(tokens.nextToken());
+ }
+ LOG.info("parseFile {}", wordSet);
--- End diff --
perhaps debug?
> Sample Topology making use of BlobStore API
> -------------------------------------------
>
> Key: STORM-1373
> URL: https://issues.apache.org/jira/browse/STORM-1373
> Project: Apache Storm
> Issue Type: Story
> Reporter: Sanket Reddy
> Priority: Minor
>
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)