http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/b467b0b4/pirk-2.log
----------------------------------------------------------------------
diff --git a/pirk-2.log b/pirk-2.log
deleted file mode 100644
index 4c22d5b..0000000
--- a/pirk-2.log
+++ /dev/null
@@ -1,6349 +0,0 @@
-2016-08-30 09:55:32,677 INFO  localizer.Localizer  [] - Reconstruct localized 
resource: 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/960b6925-fc51-4264-984e-d7937568606f/supervisor/usercache
-2016-08-30 09:55:32,682 WARN  localizer.Localizer  [] - No left over resources 
found for any user during reconstructing of local resources at: 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/960b6925-fc51-4264-984e-d7937568606f/supervisor/usercache
-2016-08-30 09:55:32,684 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:32,684 INFO  zookeeper.ZooKeeper  [] - Initiating client 
connection, connectString=localhost:2000 sessionTimeout=20000 
watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@635900ef
-2016-08-30 09:55:32,688 INFO  zookeeper.ClientCnxn  [] - Opening socket 
connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate 
using SASL (unknown error)
-2016-08-30 09:55:32,689 INFO  zookeeper.ClientCnxn  [] - Socket connection 
established to localhost/127.0.0.1:2000, initiating session
-2016-08-30 09:55:32,689 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56311
-2016-08-30 09:55:32,689 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56311
-2016-08-30 09:55:32,690 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d0009 with negotiated timeout 20000 for client /127.0.0.1:56311
-2016-08-30 09:55:32,690 INFO  zookeeper.ClientCnxn  [] - Session establishment 
complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d0009, 
negotiated timeout = 20000
-2016-08-30 09:55:32,691 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:32,691 INFO  storm.zookeeper    [] - Zookeeper state update: 
:connected:none
-2016-08-30 09:55:32,693 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:32,693 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0009
-2016-08-30 09:55:32,694 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d0009 closed
-2016-08-30 09:55:32,694 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56311 which had sessionid 0x156dbbbca4d0009
-2016-08-30 09:55:32,694 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:32,695 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:32,697 INFO  zookeeper.ZooKeeper  [] - Initiating client 
connection, connectString=localhost:2000/storm sessionTimeout=20000 
watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@4fd9deed
-2016-08-30 09:55:32,702 INFO  zookeeper.ClientCnxn  [] - Opening socket 
connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate 
using SASL (unknown error)
-2016-08-30 09:55:32,703 INFO  zookeeper.ClientCnxn  [] - Socket connection 
established to localhost/127.0.0.1:2000, initiating session
-2016-08-30 09:55:32,703 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56312
-2016-08-30 09:55:32,704 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56312
-2016-08-30 09:55:32,705 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d000a with negotiated timeout 20000 for client /127.0.0.1:56312
-2016-08-30 09:55:32,705 INFO  zookeeper.ClientCnxn  [] - Session establishment 
complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000a, 
negotiated timeout = 20000
-2016-08-30 09:55:32,706 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:32,718 INFO  daemon.supervisor  [] - Starting supervisor with 
id 59a01bf6-2a07-4e98-a935-3031a25e93b7 at host traveler.ltsnet.net
-2016-08-30 09:55:32,719 INFO  logging.ThriftAccessLogger  [] - Request ID: 1 
access from:  principal:  operation: submitTopology
-2016-08-30 09:55:32,757 INFO  daemon.nimbus      [] - Received topology 
submission for pirk_integration_test with conf {"topology.max.task.parallelism" 
nil, "qSchema" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml",
 "output" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt",
 "topology.executor.send.buffer.size" 1024, "topology.submitter.principal" "", 
"topology.acker.executors" 1, "topology.eventlogger.executors" 0, "queryInfo" 
{"dataPartitionsBitSize" 8, "numPartitionsPerDataElement" 36, "hashBitSize" 12, 
"queryType" "dn
 s-hostname-query", "embedSelector" true, "uuid" 
"74cbee54-9a78-48c3-905d-c83a0782b051", "hashKey" "someKey", 
"useExpLookupTable" false, "useHDFSExpLookupTable" false, 
"numBitsPerDataElement" 256, "numSelectors" 6}, "splitPartitions" true, 
"topology.workers" 1, "topology.debug" false, "topology.max.spout.pending" 300, 
"maxHitsPerSelector" 100, "storm.zookeeper.superACL" nil, "useHdfs" false, 
"limitHitsPerSelector" true, "hdfsUri" "localhost", "saltColumns" true, 
"topology.transfer.buffer.size" 32, "topology.executor.receive.buffer.size" 
1024, "allowAdHocQuerySchemas" false, "queryFile" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/pir_integrationTest-query7554668182117678094.txt",
 "topology.users" (), "topology.submitter.user" "voyager", 
"topology.component.resources.onheap.memory.mb" 128.0, "rowDivisions" 2, 
"topology.kryo.register" nil, "worker.heap.memory.mb" 750, "nSquared" 
"893714063091493252140805543927462484189909163470256324669366233674468493934531351808411750089403379605
 
919371579520956313614309759596321404680678803731365842878590760838574242511806531716813184820975781862008105030373882425133453241",
 "enccolmultboltPar" 2, "encrowcalcboltPar" 2, "dschema" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml",
 "topology.kryo.decorators" (), "storm.id" 
"pirk_integration_test-1-1472565332", "topology.name" "pirk_integration_test"}
-2016-08-30 09:55:32,773 INFO  daemon.nimbus      [] - uploadedJar 
-2016-08-30 09:55:32,794 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:32,795 INFO  zookeeper.ZooKeeper  [] - Initiating client 
connection, connectString=localhost:2000/storm sessionTimeout=20000 
watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@799bb4ed
-2016-08-30 09:55:32,799 INFO  zookeeper.ClientCnxn  [] - Opening socket 
connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate 
using SASL (unknown error)
-2016-08-30 09:55:32,800 INFO  zookeeper.ClientCnxn  [] - Socket connection 
established to localhost/127.0.0.1:2000, initiating session
-2016-08-30 09:55:32,801 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56313
-2016-08-30 09:55:32,802 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56313
-2016-08-30 09:55:32,803 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d000b with negotiated timeout 20000 for client /127.0.0.1:56313
-2016-08-30 09:55:32,803 INFO  zookeeper.ClientCnxn  [] - Session establishment 
complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000b, 
negotiated timeout = 20000
-2016-08-30 09:55:32,804 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:32,807 INFO  server.PrepRequestProcessor  [] - Got user-level 
KeeperException when processing sessionid:0x156dbbbca4d000b type:create 
cxid:0x2 zxid:0x22 txntype:-1 reqpath:n/a Error 
Path:/storm/blobstoremaxkeysequencenumber Error:KeeperErrorCode = NoNode for 
/storm/blobstoremaxkeysequencenumber
-2016-08-30 09:55:32,815 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:32,817 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d000b
-2016-08-30 09:55:32,818 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56313 which had sessionid 0x156dbbbca4d000b
-2016-08-30 09:55:32,819 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d000b closed
-2016-08-30 09:55:32,819 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:32,821 INFO  storm.cluster      [] - 
setup-path/blobstore/pirk_integration_test-1-1472565332-stormconf.ser/traveler.ltsnet.net:6627-1
-2016-08-30 09:55:32,829 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:32,830 INFO  zookeeper.ZooKeeper  [] - Initiating client 
connection, connectString=localhost:2000/storm sessionTimeout=20000 
watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@4e64e44c
-2016-08-30 09:55:32,832 INFO  zookeeper.ClientCnxn  [] - Opening socket 
connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate 
using SASL (unknown error)
-2016-08-30 09:55:32,833 INFO  zookeeper.ClientCnxn  [] - Socket connection 
established to localhost/127.0.0.1:2000, initiating session
-2016-08-30 09:55:32,833 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56314
-2016-08-30 09:55:32,833 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56314
-2016-08-30 09:55:32,834 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d000c with negotiated timeout 20000 for client /127.0.0.1:56314
-2016-08-30 09:55:32,834 INFO  zookeeper.ClientCnxn  [] - Session establishment 
complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000c, 
negotiated timeout = 20000
-2016-08-30 09:55:32,834 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:32,838 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:32,839 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d000c
-2016-08-30 09:55:32,840 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56314 which had sessionid 0x156dbbbca4d000c
-2016-08-30 09:55:32,840 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:32,841 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d000c closed
-2016-08-30 09:55:32,841 INFO  storm.cluster      [] - 
setup-path/blobstore/pirk_integration_test-1-1472565332-stormcode.ser/traveler.ltsnet.net:6627-1
-2016-08-30 09:55:32,852 INFO  daemon.nimbus      [] - desired replication 
count of 1 not achieved but we have hit the max wait time 60 so moving on with 
replication count for conf key = 1 for code key = 1for jar key = 1
-2016-08-30 09:55:32,919 INFO  daemon.nimbus      [] - Activating 
pirk_integration_test: pirk_integration_test-1-1472565332
-2016-08-30 09:55:32,924 INFO  storm.KafkaStormIntegrationTest  [] - Pausing 
for setup.
-2016-08-30 09:55:33,671 INFO  scheduler.EvenScheduler  [] - Available slots: 
(["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1025] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1026])
-2016-08-30 09:55:33,674 INFO  daemon.nimbus      [] - Setting new assignment 
for topology id pirk_integration_test-1-1472565332: 
#org.apache.storm.daemon.common.Assignment{:master-code-dir 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60",
 :node->host {"59a01bf6-2a07-4e98-a935-3031a25e93b7" "traveler.ltsnet.net"}, 
:executor->node+port {[8 8] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [7 
7] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [6 6] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [5 5] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [4 4] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [3 3] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [2 2] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [1 1] 
["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024]}, :executor->start-time-secs {[1 
1] 1472565333, [2 2] 1472565333, [3 3] 1472565333, [4 4] 1472565333, [5 5] 
1472565333, [6 6] 1472565333, [7 7] 1472565333, [8 8] 1472565333}, 
:worker->resources {["
 59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024] [0.0 0.0 0.0]}}
-2016-08-30 09:55:33,689 INFO  daemon.supervisor  [] - Downloading code for 
storm id pirk_integration_test-1-1472565332
-2016-08-30 09:55:33,690 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:33,691 INFO  zookeeper.ZooKeeper  [] - Initiating client 
connection, connectString=localhost:2000/storm sessionTimeout=20000 
watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@1537621d
-2016-08-30 09:55:33,693 INFO  zookeeper.ClientCnxn  [] - Opening socket 
connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate 
using SASL (unknown error)
-2016-08-30 09:55:33,693 INFO  zookeeper.ClientCnxn  [] - Socket connection 
established to localhost/127.0.0.1:2000, initiating session
-2016-08-30 09:55:33,693 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56315
-2016-08-30 09:55:33,694 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56315
-2016-08-30 09:55:33,695 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d000d with negotiated timeout 20000 for client /127.0.0.1:56315
-2016-08-30 09:55:33,695 INFO  zookeeper.ClientCnxn  [] - Session establishment 
complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000d, 
negotiated timeout = 20000
-2016-08-30 09:55:33,695 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:33,699 INFO  blobstore.FileBlobStoreImpl  [] - Creating new 
blob store based in 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60/blobs
-2016-08-30 09:55:33,727 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:33,730 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d000d
-2016-08-30 09:55:33,731 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:33,731 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d000d closed
-2016-08-30 09:55:33,731 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56315 which had sessionid 0x156dbbbca4d000d
-2016-08-30 09:55:34,000 INFO  server.SessionTrackerImpl  [] - 
SessionTrackerImpl exited loop!
-2016-08-30 09:55:34,763 INFO  daemon.supervisor  [] - Finished downloading 
code for storm id pirk_integration_test-1-1472565332
-2016-08-30 09:55:34,770 INFO  daemon.supervisor  [] - Launching worker with 
assignment {:storm-id "pirk_integration_test-1-1472565332", :executors [[8 8] 
[7 7] [6 6] [5 5] [4 4] [3 3] [2 2] [1 1]], :resources 
#object[org.apache.storm.generated.WorkerResources 0x6ff2384a 
"WorkerResources(mem_on_heap:0.0, mem_off_heap:0.0, cpu:0.0)"]} for this 
supervisor 59a01bf6-2a07-4e98-a935-3031a25e93b7 on port 1024 with id 
cacfc67b-c838-449e-8e35-dca6d3f1da94
-2016-08-30 09:55:34,773 INFO  daemon.worker      [] - Launching worker for 
pirk_integration_test-1-1472565332 on 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024 
with id cacfc67b-c838-449e-8e35-dca6d3f1da94 and conf 
{"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" 
"-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "local", 
"storm.messaging.netty.client_worker_threads" 1, 
"logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" 
false, "topology.max.task.parallelism" nil, "topology.priority" 29, 
"zmq.threads" 1, "storm.group.mapping.service" 
"org.apache.storm.security.auth.ShellBasedGroupsMapping", 
"transactional.zookeeper.root" "/transactional", 
"topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" 
false, "topology.max.replication.wait.time.sec" 60, "qSchema" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file66
 
01659287734868610.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml",
 "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 
10240, "topology.multilang.serializer" 
"org.apache.storm.multilang.JsonSerializer", 
"storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" 
"org.apache.storm.blobstore.LocalFsBlobStore", 
"resource.aware.scheduler.eviction.strategy" 
"org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy",
 "topology.max.error.report.per.interval" 5, "storm.thrift.transport" 
"org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "output" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt",
 "storm.group.mapping.service.params" n
 il, "worker.profiler.enabled" false, "storm.principal.tolocal" 
"org.apache.storm.security.auth.DefaultPrincipalToLocal", 
"supervisor.worker.shutdown.sleep.secs" 1, "pacemaker.host" "localhost", 
"storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 
0, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, 
"storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, 
"storm.zookeeper.port" 2000, "supervisor.heartbeat.frequency.secs" 5, 
"topology.enable.message.timeouts" false, "supervisor.cpu.capacity" 400.0, 
"drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, 
"drpc.queue.size" 128, "topology.backpressure.enable" true, 
"supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", 
"storm.blobstore.inputstream.buffer.size.bytes" 65536, 
"topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", 
"nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, 
"topology.scheduler.strategy" "org.apach
 
e.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", 
"topology.executor.send.buffer.size" 1024, 
"resource.aware.scheduler.priority.strategy" 
"org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy",
 "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" 
["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], 
"topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" 
"org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", 
"storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 
3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", 
"topology.acker.executors" 1, "topology.fall.back.on.java.serialization" true, 
"topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 
600000, "storm.zookeeper.servers" ["localhost"], "nimbus.thrift.threads" 64, 
"logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topolog
 y.classpath" nil, "supervisor.monitor.frequency.secs" 3, 
"nimbus.credential.renewers.freq.secs" 600, 
"topology.skip.missing.kryo.registrations" true, "drpc.authorizer.acl.filename" 
"drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], 
"storm.group.mapping.service.cache.duration.secs" 120, 
"topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, 
"queryInfo" {"useHDFSExpLookupTable" false, "hashKey" "someKey", "numSelectors" 
6, "dataPartitionsBitSize" 8, "hashBitSize" 12, "numPartitionsPerDataElement" 
36, "numBitsPerDataElement" 256, "uuid" "74cbee54-9a78-48c3-905d-c83a0782b051", 
"useExpLookupTable" false, "embedSelector" true, "queryType" 
"dns-hostname-query"}, "storm.health.check.timeout.ms" 5000, 
"supervisor.supervisors" [], "topology.tasks" nil, 
"topology.bolts.outgoing.overflow.buffer.enable" false, "splitPartitions" true, 
"storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, 
"pacemaker.base.threads" 10, "storm.local.dir" "/var/folders/83/07wq_8t55
 pn3mgj99512hb8c0000gn/T//960b6925-fc51-4264-984e-d7937568606f", 
"topology.disable.loadaware" false, "worker.childopts" "-Xmx%HEAP-MEM%m 
-XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps 
-XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 
-XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError 
-XX:HeapDumpPath=artifacts/heapdump", "storm.auth.simple-white-list.users" [], 
"topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 
30, "topology.state.synchronization.timeout.secs" 60, 
"topology.tuple.serializer" 
"org.apache.storm.serialization.types.ListDelegateSerializer", 
"supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, 
"logviewer.childopts" "-Xmx128m", "topology.environment" nil, "topology.debug" 
false, "topology.disruptor.batch.size" 100, "storm.messaging.netty.max_retries" 
300, "ui.childopts" "-Xmx768m", "storm.network.topography.plugin" 
"org.apache.storm.networktopography.DefaultRackDNSToSwitchMap
 ping", "storm.zookeeper.session.timeout" 20000, "drpc.childopts" "-Xmx768m", 
"drpc.http.creds.plugin" 
"org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", 
"storm.zookeeper.connection.timeout" 15000, "storm.zookeeper.auth.user" nil, 
"storm.meta.serialization.delegate" 
"org.apache.storm.serialization.GzipThriftSerializationDelegate", 
"topology.max.spout.pending" 300, "storm.codedistributor.class" 
"org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", 
"nimbus.supervisor.timeout.secs" 60, "maxHitsPerSelector" 100, 
"nimbus.task.timeout.secs" 30, "drpc.port" 3772, "pacemaker.max.threads" 50, 
"storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 
6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 
10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, 
"supervisor.worker.start.timeout.secs" 120, "useHdfs" false, 
"storm.zookeeper.retry.interval" 1000, "limitHitsPerSelector" true, 
"logs.users" nil, "wor
 ker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, 
"drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, 
"task.credentials.poll.secs" 30, "blobstore.superuser" "voyager", 
"drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, 
"topology.state.checkpoint.interval.ms" 1000, "hdfsUri" "localhost", 
"supervisor.slots.ports" (1024 1025 1026), "saltColumns" true, 
"topology.transfer.buffer.size" 32, "storm.health.check.dir" "healthchecks", 
"topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" 
false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" 
"-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", 
"topology.executor.receive.buffer.size" 1024, 
"backpressure.disruptor.low.watermark" 0.4, "allowAdHocQuerySchemas" false, 
"queryFile" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/pir_integrationTest-query7554668182117678094.txt",
 "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.mess
 aging.netty.buffer_size" 5242880, "storm.cluster.state.store" 
"org.apache.storm.cluster_state.zookeeper_state_factory", 
"worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", 
"ui.http.creds.plugin" 
"org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", 
"storm.zookeeper.root" "/storm", "topology.tick.tuple.freq.secs" nil, 
"drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", 
"supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, 
"storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, 
"pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, 
"nimbus.impersonation.authorizer" 
"org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer", 
"topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, 
"topology.error.throttle.interval.secs" 10, "storm.messaging.transport" 
"org.apache.storm.messaging.netty.Context", 
"storm.messaging.netty.authentication" false, "topology.component.resources.
 onheap.memory.mb" 128.0, "rowDivisions" 2, "topology.kryo.factory" 
"org.apache.storm.serialization.DefaultKryoFactory", "worker.gc.childopts" "", 
"nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", 
"nimbus.seeds" ["localhost"], "nimbus.queue.size" 100000, 
"nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, 
"worker.heap.memory.mb" 750, "logviewer.max.sum.worker.logs.size.mb" 4096, 
"pacemaker.childopts" "-Xmx1024m", "nSquared" 
"893714063091493252140805543927462484189909163470256324669366233674468493934531351808411750089403379605919371579520956313614309759596321404680678803731365842878590760838574242511806531716813184820975781862008105030373882425133453241",
 "ui.users" nil, "transactional.zookeeper.servers" nil, 
"supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, 
"client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", 
"supervisor.childopts" "-Xmx256m", "topology.worker.max.heap.size.mb" 768.0, "b
 ackpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, 
"ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, 
"topology.disruptor.wait.timeout.millis" 1000, "enccolmultboltPar" 2, 
"storm.nimbus.retry.intervalceiling.millis" 60000, 
"topology.trident.batch.emit.interval.millis" 50, "storm.auth.simple-acl.users" 
[], "drpc.invocations.threads" 64, "java.library.path" 
"/usr/local/lib:/opt/local/lib:/usr/lib", "encrowcalcboltPar" 2, "ui.port" 
8080, "dschema" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml",
 "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", 
"storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" 
"A1", "nimbus.thrift.max_buffer_size" 1048576, 
"storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600}
-2016-08-30 09:55:34,777 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:34,777 INFO  zookeeper.ZooKeeper  [] - Initiating client 
connection, connectString=localhost:2000 sessionTimeout=20000 
watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@5ec68ddc
-2016-08-30 09:55:34,779 INFO  zookeeper.ClientCnxn  [] - Opening socket 
connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate 
using SASL (unknown error)
-2016-08-30 09:55:34,780 INFO  zookeeper.ClientCnxn  [] - Socket connection 
established to localhost/127.0.0.1:2000, initiating session
-2016-08-30 09:55:34,780 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56316
-2016-08-30 09:55:34,780 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56316
-2016-08-30 09:55:34,781 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d000e with negotiated timeout 20000 for client /127.0.0.1:56316
-2016-08-30 09:55:34,781 INFO  zookeeper.ClientCnxn  [] - Session establishment 
complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000e, 
negotiated timeout = 20000
-2016-08-30 09:55:34,781 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:34,781 INFO  storm.zookeeper    [] - Zookeeper state update: 
:connected:none
-2016-08-30 09:55:34,782 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:34,783 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d000e
-2016-08-30 09:55:34,783 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d000e closed
-2016-08-30 09:55:34,783 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:34,784 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56316 which had sessionid 0x156dbbbca4d000e
-2016-08-30 09:55:34,784 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:34,784 INFO  zookeeper.ZooKeeper  [] - Initiating client 
connection, connectString=localhost:2000/storm sessionTimeout=20000 
watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@2b377b2c
-2016-08-30 09:55:34,786 INFO  zookeeper.ClientCnxn  [] - Opening socket 
connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate 
using SASL (unknown error)
-2016-08-30 09:55:34,786 INFO  zookeeper.ClientCnxn  [] - Socket connection 
established to localhost/127.0.0.1:2000, initiating session
-2016-08-30 09:55:34,786 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56317
-2016-08-30 09:55:34,787 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56317
-2016-08-30 09:55:34,788 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d000f with negotiated timeout 20000 for client /127.0.0.1:56317
-2016-08-30 09:55:34,788 INFO  zookeeper.ClientCnxn  [] - Session establishment 
complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000f, 
negotiated timeout = 20000
-2016-08-30 09:55:34,788 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:34,793 INFO  auth.AuthUtils     [] - Got AutoCreds []
-2016-08-30 09:55:34,793 INFO  daemon.worker      [] - Reading Assignments.
-2016-08-30 09:55:34,806 INFO  daemon.worker      [] - Registering 
IConnectionCallbacks for 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024
-2016-08-30 09:55:34,818 INFO  daemon.executor    [] - Loading executor 
partitiondataBolt:[8 8]
-2016-08-30 09:55:34,819 INFO  daemon.executor    [] - Loaded executor tasks 
partitiondataBolt:[8 8]
-2016-08-30 09:55:34,824 INFO  daemon.executor    [] - Finished loading 
executor partitiondataBolt:[8 8]
-2016-08-30 09:55:34,829 INFO  daemon.executor    [] - Loading executor 
enccolmultbolt:[2 2]
-2016-08-30 09:55:34,830 INFO  daemon.executor    [] - Loaded executor tasks 
enccolmultbolt:[2 2]
-2016-08-30 09:55:34,832 INFO  daemon.executor    [] - Finished loading 
executor enccolmultbolt:[2 2]
-2016-08-30 09:55:34,836 INFO  daemon.executor    [] - Loading executor 
outputbolt:[7 7]
-2016-08-30 09:55:34,837 INFO  daemon.executor    [] - Loaded executor tasks 
outputbolt:[7 7]
-2016-08-30 09:55:34,839 INFO  daemon.executor    [] - Finished loading 
executor outputbolt:[7 7]
-2016-08-30 09:55:34,844 INFO  daemon.executor    [] - Loading executor 
enccolmultbolt:[3 3]
-2016-08-30 09:55:34,845 INFO  daemon.executor    [] - Loaded executor tasks 
enccolmultbolt:[3 3]
-2016-08-30 09:55:34,848 INFO  daemon.executor    [] - Finished loading 
executor enccolmultbolt:[3 3]
-2016-08-30 09:55:34,852 INFO  daemon.executor    [] - Loading executor 
__acker:[1 1]
-2016-08-30 09:55:34,853 INFO  daemon.executor    [] - Loaded executor tasks 
__acker:[1 1]
-2016-08-30 09:55:34,854 INFO  daemon.executor    [] - Timeouts disabled for 
executor __acker:[1 1]
-2016-08-30 09:55:34,855 INFO  daemon.executor    [] - Finished loading 
executor __acker:[1 1]
-2016-08-30 09:55:34,859 INFO  daemon.executor    [] - Loading executor 
kafkaspout:[6 6]
-2016-08-30 09:55:34,861 INFO  daemon.executor    [] - Loaded executor tasks 
kafkaspout:[6 6]
-2016-08-30 09:55:34,864 INFO  daemon.executor    [] - Timeouts disabled for 
executor kafkaspout:[6 6]
-2016-08-30 09:55:34,865 INFO  daemon.executor    [] - Finished loading 
executor kafkaspout:[6 6]
-2016-08-30 09:55:34,869 INFO  daemon.executor    [] - Loading executor 
__system:[-1 -1]
-2016-08-30 09:55:34,870 INFO  daemon.executor    [] - Loaded executor tasks 
__system:[-1 -1]
-2016-08-30 09:55:34,873 INFO  daemon.executor    [] - Finished loading 
executor __system:[-1 -1]
-2016-08-30 09:55:34,877 INFO  daemon.executor    [] - Loading executor 
encrowcalcbolt:[5 5]
-2016-08-30 09:55:34,878 INFO  daemon.executor    [] - Loaded executor tasks 
encrowcalcbolt:[5 5]
-2016-08-30 09:55:34,880 INFO  daemon.executor    [] - Finished loading 
executor encrowcalcbolt:[5 5]
-2016-08-30 09:55:34,885 INFO  daemon.executor    [] - Loading executor 
encrowcalcbolt:[4 4]
-2016-08-30 09:55:34,886 INFO  daemon.executor    [] - Loaded executor tasks 
encrowcalcbolt:[4 4]
-2016-08-30 09:55:34,888 INFO  daemon.executor    [] - Finished loading 
executor encrowcalcbolt:[4 4]
-2016-08-30 09:55:34,890 INFO  daemon.worker      [] - Started with log levels: 
{"org.apache.hadoop.hbase" #object[org.apache.logging.log4j.Level 0x7af08580 
"INFO"], "" #object[org.apache.logging.log4j.Level 0x7af08580 "INFO"], 
"org.apache.hadoop.hbase.zookeeper" #object[org.apache.logging.log4j.Level 
0x6d660c48 "WARN"], "org.apache.zookeeper" 
#object[org.apache.logging.log4j.Level 0xf759057 "ERROR"]}
-2016-08-30 09:55:34,893 INFO  daemon.worker      [] - Worker has topology 
config {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" 
"-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "local", 
"storm.messaging.netty.client_worker_threads" 1, 
"logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" 
false, "topology.max.task.parallelism" nil, "topology.priority" 29, 
"zmq.threads" 1, "storm.group.mapping.service" 
"org.apache.storm.security.auth.ShellBasedGroupsMapping", 
"transactional.zookeeper.root" "/transactional", 
"topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" 
false, "topology.max.replication.wait.time.sec" 60, "qSchema" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml,/var/folde
 
rs/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml",
 "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 
10240, "topology.multilang.serializer" 
"org.apache.storm.multilang.JsonSerializer", 
"storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" 
"org.apache.storm.blobstore.LocalFsBlobStore", 
"resource.aware.scheduler.eviction.strategy" 
"org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy",
 "topology.max.error.report.per.interval" 5, "storm.thrift.transport" 
"org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "output" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt",
 "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, 
"storm.principal.tolocal" 
"org.apache.storm.security.auth.DefaultPrincipalToLocal", "superv
 isor.worker.shutdown.sleep.secs" 1, "pacemaker.host" "localhost", 
"storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 
0, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, 
"storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, 
"storm.zookeeper.port" 2000, "supervisor.heartbeat.frequency.secs" 5, 
"topology.enable.message.timeouts" false, "supervisor.cpu.capacity" 400.0, 
"drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, 
"drpc.queue.size" 128, "topology.backpressure.enable" true, 
"supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", 
"storm.blobstore.inputstream.buffer.size.bytes" 65536, 
"topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", 
"nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, 
"topology.scheduler.strategy" 
"org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy",
 "topology.executor.send.buffer.size" 1024, "reso
 urce.aware.scheduler.priority.strategy" 
"org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy",
 "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" 
["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], 
"topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" 
"org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", 
"topology.submitter.principal" "", "storm.nimbus.retry.interval.millis" 2000, 
"nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" 
"/tmp/dev-storm-zookeeper", "topology.acker.executors" 1, 
"topology.fall.back.on.java.serialization" true, 
"topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 
600000, "storm.zookeeper.servers" ["localhost"], "nimbus.thrift.threads" 64, 
"logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, 
"topology.classpath" nil, "supervisor.monitor.frequency.secs" 3, 
"nimbus.credential.renewers.freq.sec
 s" 600, "topology.skip.missing.kryo.registrations" true, 
"drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" 
[], "storm.group.mapping.service.cache.duration.secs" 120, 
"topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, 
"queryInfo" {"dataPartitionsBitSize" 8, "numPartitionsPerDataElement" 36, 
"hashBitSize" 12, "queryType" "dns-hostname-query", "embedSelector" true, 
"uuid" "74cbee54-9a78-48c3-905d-c83a0782b051", "hashKey" "someKey", 
"useExpLookupTable" false, "useHDFSExpLookupTable" false, 
"numBitsPerDataElement" 256, "numSelectors" 6}, "storm.health.check.timeout.ms" 
5000, "supervisor.supervisors" [], "topology.tasks" nil, 
"topology.bolts.outgoing.overflow.buffer.enable" false, "splitPartitions" true, 
"storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, 
"pacemaker.base.threads" 10, "storm.local.dir" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60",
 "topology.disable.loadaware" 
 false, "worker.childopts" "-Xmx%HEAP-MEM%m -XX:+PrintGCDetails 
-Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps 
-XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M 
-XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump", 
"storm.auth.simple-white-list.users" [], 
"topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 
30, "topology.state.synchronization.timeout.secs" 60, 
"topology.tuple.serializer" 
"org.apache.storm.serialization.types.ListDelegateSerializer", 
"supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, 
"logviewer.childopts" "-Xmx128m", "topology.environment" nil, "topology.debug" 
false, "topology.disruptor.batch.size" 100, "storm.messaging.netty.max_retries" 
300, "ui.childopts" "-Xmx768m", "storm.network.topography.plugin" 
"org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", 
"storm.zookeeper.session.timeout" 20000, "drpc.childopts" "-Xmx768m", 
"drpc.http.creds
 .plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", 
"storm.zookeeper.connection.timeout" 15000, "storm.zookeeper.auth.user" nil, 
"storm.meta.serialization.delegate" 
"org.apache.storm.serialization.GzipThriftSerializationDelegate", 
"topology.max.spout.pending" 300, "storm.codedistributor.class" 
"org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", 
"nimbus.supervisor.timeout.secs" 60, "maxHitsPerSelector" 100, 
"nimbus.task.timeout.secs" 30, "storm.zookeeper.superACL" nil, "drpc.port" 
3772, "pacemaker.max.threads" 50, 
"storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 
6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 
10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, 
"supervisor.worker.start.timeout.secs" 120, "useHdfs" false, 
"storm.zookeeper.retry.interval" 1000, "limitHitsPerSelector" true, 
"logs.users" nil, "worker.profiler.command" "flight.bash", 
"transactional.zookeeper
 .port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, 
"task.credentials.poll.secs" 30, "blobstore.superuser" "voyager", 
"drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, 
"topology.state.checkpoint.interval.ms" 1000, "hdfsUri" "localhost", 
"supervisor.slots.ports" [6700 6701 6702 6703], "saltColumns" true, 
"topology.transfer.buffer.size" 32, "storm.health.check.dir" "healthchecks", 
"topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" 
false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" 
"-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", 
"topology.executor.receive.buffer.size" 1024, 
"backpressure.disruptor.low.watermark" 0.4, "allowAdHocQuerySchemas" false, 
"queryFile" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/pir_integrationTest-query7554668182117678094.txt",
 "topology.users" [], "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" 
false, "storm.messaging.netty.buffer_size" 5242880, "
 storm.cluster.state.store" 
"org.apache.storm.cluster_state.zookeeper_state_factory", 
"worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", 
"ui.http.creds.plugin" 
"org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", 
"storm.zookeeper.root" "/storm", "topology.submitter.user" "voyager", 
"topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, 
"storm.workers.artifacts.dir" "workers-artifacts", 
"supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, 
"storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, 
"pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, 
"nimbus.impersonation.authorizer" 
"org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer", 
"topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, 
"topology.error.throttle.interval.secs" 10, "storm.messaging.transport" 
"org.apache.storm.messaging.netty.Context", 
"storm.messaging.netty.authentication" false, "topology.component.resource
 s.onheap.memory.mb" 128.0, "rowDivisions" 2, "topology.kryo.factory" 
"org.apache.storm.serialization.DefaultKryoFactory", "topology.kryo.register" 
nil, "worker.gc.childopts" "", "nimbus.topology.validator" 
"org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" 
["localhost"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 
600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 750, 
"logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" 
"-Xmx1024m", "nSquared" 
"893714063091493252140805543927462484189909163470256324669366233674468493934531351808411750089403379605919371579520956313614309759596321404680678803731365842878590760838574242511806531716813184820975781862008105030373882425133453241",
 "ui.users" nil, "transactional.zookeeper.servers" nil, 
"supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, 
"client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", 
"supervisor.childopts" "-Xmx256m", "topology.wo
 rker.max.heap.size.mb" 768.0, "backpressure.disruptor.high.watermark" 0.9, 
"ui.filter" nil, "ui.header.buffer.bytes" 4096, 
"topology.min.replication.count" 1, "topology.disruptor.wait.timeout.millis" 
1000, "enccolmultboltPar" 2, "storm.nimbus.retry.intervalceiling.millis" 60000, 
"topology.trident.batch.emit.interval.millis" 50, "storm.auth.simple-acl.users" 
[], "drpc.invocations.threads" 64, "java.library.path" 
"/usr/local/lib:/opt/local/lib:/usr/lib", "encrowcalcboltPar" 2, "ui.port" 
8080, "dschema" 
"/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml",
 "topology.kryo.decorators" [], "storm.id" 
"pirk_integration_test-1-1472565332", "topology.name" "pirk_integration_test", 
"storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", 
"storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" 
"A1", "nimbus.thrift.max_buffer_size" 1048576, 
"storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600}
-2016-08-30 09:55:34,894 INFO  daemon.worker      [] - Worker 
cacfc67b-c838-449e-8e35-dca6d3f1da94 for storm 
pirk_integration_test-1-1472565332 on 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024 
has finished loading
-2016-08-30 09:55:34,894 INFO  storm.config       [] - SET worker-user 
cacfc67b-c838-449e-8e35-dca6d3f1da94 
-2016-08-30 09:55:35,801 INFO  daemon.worker      [] - All connections are 
ready for worker 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024 with id 
cacfc67b-c838-449e-8e35-dca6d3f1da94
-2016-08-30 09:55:35,844 INFO  daemon.executor    [] - Preparing bolt 
partitiondataBolt:(8)
-2016-08-30 09:55:35,845 INFO  storm.PartitionDataBolt  [] - partition databolt 
hdfs = false
-2016-08-30 09:55:35,845 INFO  data.DataSchemaLoader  [] - Loading schemaFile = 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml
 hdfs = false
-2016-08-30 09:55:35,845 INFO  data.DataSchemaLoader  [] - localFS: inputFile = 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml
-2016-08-30 09:55:35,846 INFO  data.DataSchemaLoader  [] - Root element: schema
-2016-08-30 09:55:35,847 INFO  data.DataSchemaLoader  [] - schemaName = 
testDataSchema
-2016-08-30 09:55:35,847 INFO  data.DataSchemaLoader  [] - name = date javaType 
= string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner
-2016-08-30 09:55:35,847 INFO  data.DataSchemaLoader  [] - name = qname 
javaType = string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,847 INFO  data.DataSchemaLoader  [] - name = src_ip 
javaType = string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,847 INFO  data.DataSchemaLoader  [] - name = dest_ip 
javaType = string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,848 INFO  data.DataSchemaLoader  [] - name = qtype 
javaType = short isArray = true partitioner 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,848 INFO  data.DataSchemaLoader  [] - name = rcode 
javaType = int isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,848 INFO  data.DataSchemaLoader  [] - name = ip javaType = 
string isArray = true partitioner 
org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,848 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml
-2016-08-30 09:55:35,848 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml
-2016-08-30 09:55:35,849 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,849 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-hostname-query
-2016-08-30 09:55:35,849 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - selectorName = 
qname
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = date 
partitionerName = org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = date bits = 
64 dataElementSize = 64
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = src_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = src_ip bits 
= 32 dataElementSize = 96
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 128
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = qtype 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,850 INFO  query.QuerySchemaLoader  [] - name = qtype bits 
= 32 dataElementSize = 160
-2016-08-30 09:55:35,851 INFO  query.QuerySchemaLoader  [] - name = rcode 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,851 INFO  query.QuerySchemaLoader  [] - name = rcode bits 
= 32 dataElementSize = 192
-2016-08-30 09:55:35,851 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,851 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 256
-2016-08-30 09:55:35,851 INFO  query.QuerySchemaLoader  [] - filterName = qname
-2016-08-30 09:55:35,858 INFO  daemon.executor    [] - Preparing bolt 
outputbolt:(7)
-2016-08-30 09:55:35,859 INFO  storm.OutputBolt   [] - Intitialized OutputBolt.
-2016-08-30 09:55:35,859 INFO  daemon.executor    [] - Prepared bolt 
outputbolt:(7)
-2016-08-30 09:55:35,859 INFO  daemon.executor    [] - Preparing bolt 
enccolmultbolt:(2)
-2016-08-30 09:55:35,859 INFO  storm.EncColMultBolt  [] - Initialized 
EncColMultBolt. 
-2016-08-30 09:55:35,860 INFO  daemon.executor    [] - Prepared bolt 
enccolmultbolt:(2)
-2016-08-30 09:55:35,865 INFO  daemon.executor    [] - Preparing bolt 
enccolmultbolt:(3)
-2016-08-30 09:55:35,866 INFO  storm.EncColMultBolt  [] - Initialized 
EncColMultBolt. 
-2016-08-30 09:55:35,866 INFO  daemon.executor    [] - Prepared bolt 
enccolmultbolt:(3)
-2016-08-30 09:55:35,872 INFO  daemon.executor    [] - Preparing bolt 
__acker:(1)
-2016-08-30 09:55:35,873 INFO  daemon.executor    [] - Prepared bolt __acker:(1)
-2016-08-30 09:55:35,874 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml
-2016-08-30 09:55:35,874 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-ip-query
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - selectorName = ip
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - name = src_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - name = src_ip bits 
= 32 dataElementSize = 32
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,876 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 64
-2016-08-30 09:55:35,877 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,877 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 128
-2016-08-30 09:55:35,877 INFO  query.QuerySchemaLoader  [] - filterName = qname
-2016-08-30 09:55:35,887 INFO  daemon.executor    [] - Opening spout 
kafkaspout:(6)
-2016-08-30 09:55:35,888 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:35,892 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:35,894 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml
-2016-08-30 09:55:35,894 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml
-2016-08-30 09:55:35,896 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,897 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-nxdomain-query
-2016-08-30 09:55:35,897 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,897 INFO  query.QuerySchemaLoader  [] - selectorName = 
qname
-2016-08-30 09:55:35,897 INFO  query.QuerySchemaLoader  [] - name = qname 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,898 INFO  query.QuerySchemaLoader  [] - name = qname bits 
= 64 dataElementSize = 64
-2016-08-30 09:55:35,898 INFO  query.QuerySchemaLoader  [] - name = src_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,898 INFO  query.QuerySchemaLoader  [] - name = src_ip bits 
= 32 dataElementSize = 96
-2016-08-30 09:55:35,898 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,898 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 128
-2016-08-30 09:55:35,898 INFO  query.QuerySchemaLoader  [] - filterName = qname
-2016-08-30 09:55:35,899 INFO  server.NIOServerCnxnFactory  [] - Accepted 
socket connection from /127.0.0.1:56318
-2016-08-30 09:55:35,899 INFO  server.ZooKeeperServer  [] - Client attempting 
to establish new session at /127.0.0.1:56318
-2016-08-30 09:55:35,900 INFO  daemon.executor    [] - Preparing bolt 
__system:(-1)
-2016-08-30 09:55:35,900 INFO  daemon.executor    [] - Preparing bolt 
encrowcalcbolt:(5)
-2016-08-30 09:55:35,901 INFO  storm.EncRowCalcBolt  [] - partition databolt 
hdfs = false
-2016-08-30 09:55:35,901 INFO  server.ZooKeeperServer  [] - Established session 
0x156dbbbca4d0010 with negotiated timeout 20000 for client /127.0.0.1:56318
-2016-08-30 09:55:35,901 INFO  storm.EncRowCalcBolt  [] - Initialized 
EncRowCalcBolt.
-2016-08-30 09:55:35,902 INFO  daemon.executor    [] - Prepared bolt 
encrowcalcbolt:(5)
-2016-08-30 09:55:35,902 INFO  daemon.executor    [] - Prepared bolt 
__system:(-1)
-2016-08-30 09:55:35,902 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:35,902 INFO  daemon.executor    [] - Preparing bolt 
encrowcalcbolt:(4)
-2016-08-30 09:55:35,902 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:35,902 INFO  storm.EncRowCalcBolt  [] - partition databolt 
hdfs = false
-2016-08-30 09:55:35,902 INFO  storm.EncRowCalcBolt  [] - Initialized 
EncRowCalcBolt.
-2016-08-30 09:55:35,902 INFO  daemon.executor    [] - Prepared bolt 
encrowcalcbolt:(4)
-2016-08-30 09:55:35,906 INFO  kafka.DynamicBrokersReader  [] - Read partition 
info from zookeeper: GlobalPartitionInformation{topic=pirk_test_topic, 
partitionMap={0=localhost:11111}}
-2016-08-30 09:55:35,906 INFO  imps.CuratorFrameworkImpl  [] - Starting
-2016-08-30 09:55:35,908 INFO  daemon.executor    [] - Opened spout 
kafkaspout:(6)
-2016-08-30 09:55:35,909 INFO  daemon.executor    [] - Activating spout 
kafkaspout:(6)
-2016-08-30 09:55:35,909 INFO  kafka.ZkCoordinator  [] - Task [1/1] Refreshing 
partition manager connections
-2016-08-30 09:55:35,918 INFO  state.ConnectionStateManager  [] - State change: 
CONNECTED
-2016-08-30 09:55:35,920 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml
-2016-08-30 09:55:35,920 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-srcip-query
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - selectorName = 
src_ip
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - name = qname 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - name = qname bits 
= 64 dataElementSize = 64
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,922 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 96
-2016-08-30 09:55:35,923 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,923 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 160
-2016-08-30 09:55:35,923 INFO  query.QuerySchemaLoader  [] - filterName = src_ip
-2016-08-30 09:55:35,923 INFO  query.QuerySchemaLoader  [] - filterName = ip
-2016-08-30 09:55:35,923 INFO  kafka.DynamicBrokersReader  [] - Read partition 
info from zookeeper: GlobalPartitionInformation{topic=pirk_test_topic, 
partitionMap={0=localhost:11111}}
-2016-08-30 09:55:35,923 INFO  kafka.KafkaUtils   [] - Task [1/1] assigned 
[Partition{host=localhost:11111, topic=pirk_test_topic, partition=0}]
-2016-08-30 09:55:35,923 INFO  kafka.ZkCoordinator  [] - Task [1/1] Deleted 
partition managers: []
-2016-08-30 09:55:35,923 INFO  kafka.ZkCoordinator  [] - Task [1/1] New 
partition managers: [Partition{host=localhost:11111, topic=pirk_test_topic, 
partition=0}]
-2016-08-30 09:55:35,925 INFO  kafka.PartitionManager  [] - Read partition 
information from: /pirk_test_root/pirk_integr_test_spout/partition_0  --> null
-2016-08-30 09:55:35,933 INFO  kafka.PartitionManager  [] - No partition 
information found, using configuration to determine offset
-2016-08-30 09:55:35,934 INFO  kafka.PartitionManager  [] - Last commit offset 
from zookeeper: 0
-2016-08-30 09:55:35,934 INFO  kafka.PartitionManager  [] - Commit offset 0 is 
more than 9223372036854775807 behind latest offset 0, resetting to 
startOffsetTime=-2
-2016-08-30 09:55:35,934 INFO  kafka.PartitionManager  [] - Starting Kafka 
localhost:0 from offset 0
-2016-08-30 09:55:35,934 INFO  kafka.ZkCoordinator  [] - Task [1/1] Finished 
refreshing
-2016-08-30 09:55:35,939 INFO  data.DataSchemaLoader  [] - Loading schemaFile = 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml
 hdfs = false
-2016-08-30 09:55:35,939 INFO  data.DataSchemaLoader  [] - localFS: inputFile = 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml
-2016-08-30 09:55:35,941 INFO  data.DataSchemaLoader  [] - Root element: schema
-2016-08-30 09:55:35,941 INFO  data.DataSchemaLoader  [] - schemaName = 
testDataSchema
-2016-08-30 09:55:35,941 INFO  data.DataSchemaLoader  [] - name = date javaType 
= string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner
-2016-08-30 09:55:35,941 INFO  data.DataSchemaLoader  [] - name = qname 
javaType = string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,942 INFO  data.DataSchemaLoader  [] - name = src_ip 
javaType = string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,942 INFO  data.DataSchemaLoader  [] - name = dest_ip 
javaType = string isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,942 INFO  data.DataSchemaLoader  [] - name = qtype 
javaType = short isArray = true partitioner 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,942 INFO  data.DataSchemaLoader  [] - name = rcode 
javaType = int isArray = false partitioner 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,942 INFO  data.DataSchemaLoader  [] - name = ip javaType = 
string isArray = true partitioner 
org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,942 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml
-2016-08-30 09:55:35,942 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml
-2016-08-30 09:55:35,943 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml
-2016-08-30 09:55:35,943 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml
-2016-08-30 09:55:35,943 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,943 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-hostname-query
-2016-08-30 09:55:35,943 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - selectorName = 
qname
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = date 
partitionerName = org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = date bits = 
64 dataElementSize = 64
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = src_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = src_ip bits 
= 32 dataElementSize = 96
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-srcip-query-no-filter
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 128
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - selectorName = 
src_ip
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = qname 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,944 INFO  query.QuerySchemaLoader  [] - name = qname bits 
= 64 dataElementSize = 64
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = qtype 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 96
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = qtype bits 
= 32 dataElementSize = 160
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = rcode 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 160
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = rcode bits 
= 32 dataElementSize = 192
-2016-08-30 09:55:35,945 INFO  storm.PartitionDataBolt  [] - Initialized 
ExtractAndPartitionDataBolt.
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 256
-2016-08-30 09:55:35,945 INFO  query.QuerySchemaLoader  [] - filterName = qname
-2016-08-30 09:55:35,946 INFO  daemon.executor    [] - Prepared bolt 
partitiondataBolt:(8)
-2016-08-30 09:55:35,963 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml
-2016-08-30 09:55:35,963 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml
-2016-08-30 09:55:35,964 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,964 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-ip-query
-2016-08-30 09:55:35,964 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - selectorName = ip
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - name = src_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - name = src_ip bits 
= 32 dataElementSize = 32
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 64
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 128
-2016-08-30 09:55:35,965 INFO  query.QuerySchemaLoader  [] - filterName = qname
-2016-08-30 09:55:35,981 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml
-2016-08-30 09:55:35,981 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml
-2016-08-30 09:55:35,983 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:35,983 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-nxdomain-query
-2016-08-30 09:55:35,983 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:35,983 INFO  query.QuerySchemaLoader  [] - selectorName = 
qname
-2016-08-30 09:55:35,983 INFO  query.QuerySchemaLoader  [] - name = qname 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:35,983 INFO  query.QuerySchemaLoader  [] - name = qname bits 
= 64 dataElementSize = 64
-2016-08-30 09:55:35,984 INFO  query.QuerySchemaLoader  [] - name = src_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,984 INFO  query.QuerySchemaLoader  [] - name = src_ip bits 
= 32 dataElementSize = 96
-2016-08-30 09:55:35,984 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:35,984 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 128
-2016-08-30 09:55:35,984 INFO  query.QuerySchemaLoader  [] - filterName = qname
-2016-08-30 09:55:35,999 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml
-2016-08-30 09:55:35,999 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml
-2016-08-30 09:55:36,000 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:36,000 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-srcip-query
-2016-08-30 09:55:36,001 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:36,001 INFO  query.QuerySchemaLoader  [] - selectorName = 
src_ip
-2016-08-30 09:55:36,001 INFO  query.QuerySchemaLoader  [] - name = qname 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:36,001 INFO  query.QuerySchemaLoader  [] - name = qname bits 
= 64 dataElementSize = 64
-2016-08-30 09:55:36,001 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:36,001 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 96
-2016-08-30 09:55:36,001 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:36,002 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 160
-2016-08-30 09:55:36,002 INFO  query.QuerySchemaLoader  [] - filterName = src_ip
-2016-08-30 09:55:36,002 INFO  query.QuerySchemaLoader  [] - filterName = ip
-2016-08-30 09:55:36,016 INFO  query.QuerySchemaLoader  [] - Loading schemaFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml
-2016-08-30 09:55:36,017 INFO  query.QuerySchemaLoader  [] - localFS: inputFile 
= 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml
-2016-08-30 09:55:36,018 INFO  query.QuerySchemaLoader  [] - Root element: 
schema
-2016-08-30 09:55:36,018 INFO  query.QuerySchemaLoader  [] - schemaName = 
dns-srcip-query-no-filter
-2016-08-30 09:55:36,018 INFO  query.QuerySchemaLoader  [] - dataSchemaName = 
testDataSchema
-2016-08-30 09:55:36,018 INFO  query.QuerySchemaLoader  [] - selectorName = 
src_ip
-2016-08-30 09:55:36,019 INFO  query.QuerySchemaLoader  [] - name = qname 
partitionerName = 
org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner
-2016-08-30 09:55:36,019 INFO  query.QuerySchemaLoader  [] - name = qname bits 
= 64 dataElementSize = 64
-2016-08-30 09:55:36,019 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:36,019 INFO  query.QuerySchemaLoader  [] - name = dest_ip 
bits = 32 dataElementSize = 96
-2016-08-30 09:55:36,019 INFO  query.QuerySchemaLoader  [] - name = ip 
partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner
-2016-08-30 09:55:36,019 INFO  query.QuerySchemaLoader  [] - name = ip bits = 
64 dataElementSize = 160
-2016-08-30 09:55:38,022 INFO  server.PrepRequestProcessor  [] - Got user-level 
KeeperException when processing sessionid:0x156dbbbca4d0010 type:create 
cxid:0x3 zxid:0x3b txntype:-1 reqpath:n/a Error 
Path:/pirk_test_root/pirk_integr_test_spout Error:KeeperErrorCode = NoNode for 
/pirk_test_root/pirk_integr_test_spout
-2016-08-30 09:55:42,895 INFO  storm.OutputBolt   [] - TimeToFlush reached - 
outputting response to 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt
 with columns.size = 180
-2016-08-30 09:55:42,951 INFO  storm.KafkaStormIntegrationTest  [] - Finished...
-2016-08-30 09:55:42,951 INFO  daemon.nimbus      [] - Shutting down master
-2016-08-30 09:55:42,952 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,952 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0003
-2016-08-30 09:55:42,953 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d0003 closed
-2016-08-30 09:55:42,953 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,953 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56305 which had sessionid 0x156dbbbca4d0003
-2016-08-30 09:55:42,953 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,954 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0004
-2016-08-30 09:55:42,954 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d0004 closed
-2016-08-30 09:55:42,954 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,954 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56306 which had sessionid 0x156dbbbca4d0004
-2016-08-30 09:55:42,954 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,955 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0001
-2016-08-30 09:55:42,955 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d0001 closed
-2016-08-30 09:55:42,955 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,955 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56303 which had sessionid 0x156dbbbca4d0001
-2016-08-30 09:55:42,956 INFO  storm.zookeeper    [] - closing zookeeper 
connection of leader elector.
-2016-08-30 09:55:42,956 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,957 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0000
-2016-08-30 09:55:42,958 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d0000 closed
-2016-08-30 09:55:42,958 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56302 which had sessionid 0x156dbbbca4d0000
-2016-08-30 09:55:42,958 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,958 INFO  daemon.nimbus      [] - Shut down master
-2016-08-30 09:55:42,959 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,960 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0006
-2016-08-30 09:55:42,960 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d0006 closed
-2016-08-30 09:55:42,960 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56308 which had sessionid 0x156dbbbca4d0006
-2016-08-30 09:55:42,960 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,961 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,961 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0008
-2016-08-30 09:55:42,962 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d0008 closed
-2016-08-30 09:55:42,962 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,962 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56310 which had sessionid 0x156dbbbca4d0008
-2016-08-30 09:55:42,962 INFO  daemon.supervisor  [] - Shutting down 
59a01bf6-2a07-4e98-a935-3031a25e93b7:cacfc67b-c838-449e-8e35-dca6d3f1da94
-2016-08-30 09:55:42,963 INFO  storm.config       [] - GET worker-user 
cacfc67b-c838-449e-8e35-dca6d3f1da94
-2016-08-30 09:55:42,963 INFO  storm.process-simulator  [] - Killing process 
3e97964a-19b4-43ae-9718-6834ddb4a775
-2016-08-30 09:55:42,964 INFO  daemon.worker      [] - Shutting down worker 
pirk_integration_test-1-1472565332 59a01bf6-2a07-4e98-a935-3031a25e93b7 1024
-2016-08-30 09:55:42,964 INFO  daemon.worker      [] - Terminating messaging 
context
-2016-08-30 09:55:42,964 INFO  daemon.worker      [] - Shutting down executors
-2016-08-30 09:55:42,964 INFO  daemon.executor    [] - Shutting down executor 
partitiondataBolt:[8 8]
-2016-08-30 09:55:42,965 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,965 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,965 INFO  daemon.executor    [] - Shut down executor 
partitiondataBolt:[8 8]
-2016-08-30 09:55:42,966 INFO  daemon.executor    [] - Shutting down executor 
enccolmultbolt:[2 2]
-2016-08-30 09:55:42,966 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,966 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,966 INFO  daemon.executor    [] - Shut down executor 
enccolmultbolt:[2 2]
-2016-08-30 09:55:42,966 INFO  daemon.executor    [] - Shutting down executor 
outputbolt:[7 7]
-2016-08-30 09:55:42,967 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,967 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,967 INFO  daemon.executor    [] - Shut down executor 
outputbolt:[7 7]
-2016-08-30 09:55:42,967 INFO  daemon.executor    [] - Shutting down executor 
enccolmultbolt:[3 3]
-2016-08-30 09:55:42,967 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,967 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,968 INFO  daemon.executor    [] - Shut down executor 
enccolmultbolt:[3 3]
-2016-08-30 09:55:42,968 INFO  daemon.executor    [] - Shutting down executor 
__acker:[1 1]
-2016-08-30 09:55:42,968 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,968 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,969 INFO  daemon.executor    [] - Shut down executor 
__acker:[1 1]
-2016-08-30 09:55:42,969 INFO  daemon.executor    [] - Shutting down executor 
kafkaspout:[6 6]
-2016-08-30 09:55:42,969 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,970 WARN  kafka.KafkaUtils   [] - Network error when 
fetching messages:
-java.nio.channels.ClosedByInterruptException
-       at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
 ~[?:1.8.0_102]
-       at sun.nio.ch.SocketChannelImpl.poll(SocketChannelImpl.java:957) 
~[?:1.8.0_102]
-       at 
sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:204) 
~[?:1.8.0_102]
-       at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103) 
~[?:1.8.0_102]
-       at 
java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:385) 
~[?:1.8.0_102]
-       at 
org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:81)
 ~[kafka-clients-0.9.0.1.jar:?]
-       at 
kafka.network.BlockingChannel.readCompletely(BlockingChannel.scala:129) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.network.BlockingChannel.receive(BlockingChannel.scala:120) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer.scala:86) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$$sendRequest(SimpleConsumer.scala:83)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(SimpleConsumer.scala:132)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply$mcV$sp(SimpleConsumer.scala:131)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:130) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.javaapi.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:47) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at org.apache.storm.kafka.KafkaUtils.fetchMessages(KafkaUtils.java:191) 
[storm-kafka-1.0.1.jar:1.0.1]
-       at 
org.apache.storm.kafka.PartitionManager.fill(PartitionManager.java:189) 
[storm-kafka-1.0.1.jar:1.0.1]
-       at 
org.apache.storm.kafka.PartitionManager.next(PartitionManager.java:138) 
[storm-kafka-1.0.1.jar:1.0.1]
-       at org.apache.storm.kafka.KafkaSpout.nextTuple(KafkaSpout.java:135) 
[storm-kafka-1.0.1.jar:1.0.1]
-       at 
org.apache.storm.daemon.executor$fn__7885$fn__7900$fn__7931.invoke(executor.clj:645)
 [storm-core-1.0.1.jar:1.0.1]
-       at org.apache.storm.util$async_loop$fn__625.invoke(util.clj:484) 
[storm-core-1.0.1.jar:1.0.1]
-       at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]
-       at java.lang.Thread.run(Thread.java:745) [?:1.8.0_102]
-2016-08-30 09:55:42,977 WARN  kafka.KafkaSpout   [] - Fetch failed
-org.apache.storm.kafka.FailedFetchException: 
java.nio.channels.ClosedByInterruptException
-       at org.apache.storm.kafka.KafkaUtils.fetchMessages(KafkaUtils.java:199) 
~[storm-kafka-1.0.1.jar:1.0.1]
-       at 
org.apache.storm.kafka.PartitionManager.fill(PartitionManager.java:189) 
~[storm-kafka-1.0.1.jar:1.0.1]
-       at 
org.apache.storm.kafka.PartitionManager.next(PartitionManager.java:138) 
~[storm-kafka-1.0.1.jar:1.0.1]
-       at org.apache.storm.kafka.KafkaSpout.nextTuple(KafkaSpout.java:135) 
[storm-kafka-1.0.1.jar:1.0.1]
-       at 
org.apache.storm.daemon.executor$fn__7885$fn__7900$fn__7931.invoke(executor.clj:645)
 [storm-core-1.0.1.jar:1.0.1]
-       at org.apache.storm.util$async_loop$fn__625.invoke(util.clj:484) 
[storm-core-1.0.1.jar:1.0.1]
-       at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?]
-       at java.lang.Thread.run(Thread.java:745) [?:1.8.0_102]
-Caused by: java.nio.channels.ClosedByInterruptException
-       at 
java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202)
 ~[?:1.8.0_102]
-       at sun.nio.ch.SocketChannelImpl.poll(SocketChannelImpl.java:957) 
~[?:1.8.0_102]
-       at 
sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:204) 
~[?:1.8.0_102]
-       at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103) 
~[?:1.8.0_102]
-       at 
java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:385) 
~[?:1.8.0_102]
-       at 
org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:81)
 ~[kafka-clients-0.9.0.1.jar:?]
-       at 
kafka.network.BlockingChannel.readCompletely(BlockingChannel.scala:129) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.network.BlockingChannel.receive(BlockingChannel.scala:120) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer.scala:86) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$$sendRequest(SimpleConsumer.scala:83)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(SimpleConsumer.scala:132)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply$mcV$sp(SimpleConsumer.scala:131)
 ~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at 
kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:130) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at kafka.javaapi.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:47) 
~[kafka_2.10-0.9.0.1.jar:?]
-       at org.apache.storm.kafka.KafkaUtils.fetchMessages(KafkaUtils.java:191) 
~[storm-kafka-1.0.1.jar:1.0.1]
-       ... 7 more
-2016-08-30 09:55:42,977 INFO  kafka.ZkCoordinator  [] - Task [1/1] Refreshing 
partition manager connections
-2016-08-30 09:55:42,978 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,978 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,979 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d0010
-2016-08-30 09:55:42,980 INFO  daemon.executor    [] - Shut down executor 
kafkaspout:[6 6]
-2016-08-30 09:55:42,980 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56318 which had sessionid 0x156dbbbca4d0010
-2016-08-30 09:55:42,980 INFO  daemon.executor    [] - Shutting down executor 
__system:[-1 -1]
-2016-08-30 09:55:42,981 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,981 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,981 INFO  daemon.executor    [] - Shut down executor 
__system:[-1 -1]
-2016-08-30 09:55:42,981 INFO  daemon.executor    [] - Shutting down executor 
encrowcalcbolt:[5 5]
-2016-08-30 09:55:42,982 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,982 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,982 INFO  daemon.executor    [] - Shut down executor 
encrowcalcbolt:[5 5]
-2016-08-30 09:55:42,982 INFO  daemon.executor    [] - Shutting down executor 
encrowcalcbolt:[4 4]
-2016-08-30 09:55:42,982 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,982 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,983 INFO  daemon.executor    [] - Shut down executor 
encrowcalcbolt:[4 4]
-2016-08-30 09:55:42,983 INFO  daemon.worker      [] - Shut down executors
-2016-08-30 09:55:42,983 INFO  daemon.worker      [] - Shutting down transfer 
thread
-2016-08-30 09:55:42,984 INFO  storm.util         [] - Async loop interrupted!
-2016-08-30 09:55:42,984 INFO  daemon.worker      [] - Shut down transfer thread
-2016-08-30 09:55:42,984 INFO  daemon.worker      [] - Shut down backpressure 
thread
-2016-08-30 09:55:42,986 INFO  daemon.worker      [] - Shutting down default 
resources
-2016-08-30 09:55:42,986 INFO  daemon.worker      [] - Shut down default 
resources
-2016-08-30 09:55:42,987 INFO  daemon.worker      [] - Trigger any worker 
shutdown hooks
-2016-08-30 09:55:42,988 INFO  daemon.worker      [] - Disconnecting from storm 
cluster state context
-2016-08-30 09:55:42,988 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,989 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d000f
-2016-08-30 09:55:42,989 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d000f closed
-2016-08-30 09:55:42,989 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,989 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56317 which had sessionid 0x156dbbbca4d000f
-2016-08-30 09:55:42,990 INFO  daemon.worker      [] - Shut down worker 
pirk_integration_test-1-1472565332 59a01bf6-2a07-4e98-a935-3031a25e93b7 1024
-2016-08-30 09:55:42,993 INFO  storm.config       [] - REMOVE worker-user 
cacfc67b-c838-449e-8e35-dca6d3f1da94
-2016-08-30 09:55:42,993 INFO  daemon.supervisor  [] - Shut down 
59a01bf6-2a07-4e98-a935-3031a25e93b7:cacfc67b-c838-449e-8e35-dca6d3f1da94
-2016-08-30 09:55:42,993 INFO  daemon.supervisor  [] - Shutting down supervisor 
59a01bf6-2a07-4e98-a935-3031a25e93b7
-2016-08-30 09:55:42,994 INFO  storm.event        [] - Event manager interrupted
-2016-08-30 09:55:42,994 INFO  storm.event        [] - Event manager interrupted
-2016-08-30 09:55:42,994 INFO  imps.CuratorFrameworkImpl  [] - 
backgroundOperationsLoop exiting
-2016-08-30 09:55:42,995 INFO  server.PrepRequestProcessor  [] - Processed 
session termination for sessionid: 0x156dbbbca4d000a
-2016-08-30 09:55:42,995 INFO  zookeeper.ZooKeeper  [] - Session: 
0x156dbbbca4d000a closed
-2016-08-30 09:55:42,996 INFO  zookeeper.ClientCnxn  [] - EventThread shut down
-2016-08-30 09:55:42,996 INFO  server.NIOServerCnxn  [] - Closed socket 
connection for client /127.0.0.1:56312 which had sessionid 0x156dbbbca4d000a
-2016-08-30 09:55:42,996 INFO  storm.testing      [] - Shutting down in process 
zookeeper
-2016-08-30 09:55:42,996 INFO  server.NIOServerCnxnFactory  [] - NIOServerCnxn 
factory exited run method
-2016-08-30 09:55:42,997 INFO  server.ZooKeeperServer  [] - shutting down
-2016-08-30 09:55:42,997 INFO  server.SessionTrackerImpl  [] - Shutting down
-2016-08-30 09:55:42,997 INFO  server.PrepRequestProcessor  [] - Shutting down
-2016-08-30 09:55:42,997 INFO  server.SyncRequestProcessor  [] - Shutting down
-2016-08-30 09:55:42,997 INFO  server.PrepRequestProcessor  [] - 
PrepRequestProcessor exited loop!
-2016-08-30 09:55:42,997 INFO  server.SyncRequestProcessor  [] - 
SyncRequestProcessor exited!
-2016-08-30 09:55:42,997 INFO  server.FinalRequestProcessor  [] - shutdown of 
request processor complete
-2016-08-30 09:55:42,998 INFO  storm.testing      [] - Done shutting down in 
process zookeeper
-2016-08-30 09:55:42,998 INFO  storm.testing      [] - Deleting temporary path 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60
-2016-08-30 09:55:43,001 INFO  storm.testing      [] - Deleting temporary path 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//e665f001-f902-4bd0-89ed-30254a84f1ef
-2016-08-30 09:55:43,002 INFO  storm.testing      [] - Deleting temporary path 
/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//960b6925-fc51-4264-984e-d7937568606f
-2016-08-30 09:55:43,051 INFO  storm.KafkaStormIntegrationTest  [] - Decrypting 
results. 14716
-2016-08-30 09:55:43,218 INFO  utils.BaseTests    [] - results:
-2016-08-30 09:55:43,218 INFO  utils.BaseTests    [] - 
{"date":"2016-02-20T23:29:05.000Z","src_ip":"55.55.55.55","event_type":"dns-hostname-query","query_id":"74cbee54-9a78-48c3-905d-c83a0782b051","qtype":[1,0],"dest_ip":"1.2.3.6","ip":["10.20.30.40","10.20.30.60"],"qname":"a.b.c.com","match":"a.b.c.com","rcode":0}
-2016-08-30 09:55:43,218 INFO  utils.BaseTests    [] - 
{"date":"2016-02-20T23:29:11.000Z","src_ip":"1.1.1.1","event_type":"dns-hostname-query","query_id":"74cbee54-9a78-48c3-905d-c83a0782b051","qtype":[1,0],"dest_ip":"2.2.2.2","ip":["3.3.3.3","0.0.0.0"],"qname":"something.else","match":"something.else","rcode":0}
-2016-08-30 09:55:43,218 INFO  utils.BaseTests    [] - 
{"date":"2016-02-20T23:29:06.000Z","src_ip":"127.128.129.130","event_

<TRUNCATED>

Reply via email to