http://git-wip-us.apache.org/repos/asf/incubator-pirk/blob/b467b0b4/pirk-2.log ---------------------------------------------------------------------- diff --git a/pirk-2.log b/pirk-2.log deleted file mode 100644 index 4c22d5b..0000000 --- a/pirk-2.log +++ /dev/null @@ -1,6349 +0,0 @@ -2016-08-30 09:55:32,677 INFO localizer.Localizer [] - Reconstruct localized resource: /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/960b6925-fc51-4264-984e-d7937568606f/supervisor/usercache -2016-08-30 09:55:32,682 WARN localizer.Localizer [] - No left over resources found for any user during reconstructing of local resources at: /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/960b6925-fc51-4264-984e-d7937568606f/supervisor/usercache -2016-08-30 09:55:32,684 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:32,684 INFO zookeeper.ZooKeeper [] - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@635900ef -2016-08-30 09:55:32,688 INFO zookeeper.ClientCnxn [] - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error) -2016-08-30 09:55:32,689 INFO zookeeper.ClientCnxn [] - Socket connection established to localhost/127.0.0.1:2000, initiating session -2016-08-30 09:55:32,689 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56311 -2016-08-30 09:55:32,689 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56311 -2016-08-30 09:55:32,690 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d0009 with negotiated timeout 20000 for client /127.0.0.1:56311 -2016-08-30 09:55:32,690 INFO zookeeper.ClientCnxn [] - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d0009, negotiated timeout = 20000 -2016-08-30 09:55:32,691 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:32,691 INFO storm.zookeeper [] - Zookeeper state update: :connected:none -2016-08-30 09:55:32,693 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:32,693 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0009 -2016-08-30 09:55:32,694 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d0009 closed -2016-08-30 09:55:32,694 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56311 which had sessionid 0x156dbbbca4d0009 -2016-08-30 09:55:32,694 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:32,695 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:32,697 INFO zookeeper.ZooKeeper [] - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@4fd9deed -2016-08-30 09:55:32,702 INFO zookeeper.ClientCnxn [] - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error) -2016-08-30 09:55:32,703 INFO zookeeper.ClientCnxn [] - Socket connection established to localhost/127.0.0.1:2000, initiating session -2016-08-30 09:55:32,703 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56312 -2016-08-30 09:55:32,704 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56312 -2016-08-30 09:55:32,705 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d000a with negotiated timeout 20000 for client /127.0.0.1:56312 -2016-08-30 09:55:32,705 INFO zookeeper.ClientCnxn [] - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000a, negotiated timeout = 20000 -2016-08-30 09:55:32,706 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:32,718 INFO daemon.supervisor [] - Starting supervisor with id 59a01bf6-2a07-4e98-a935-3031a25e93b7 at host traveler.ltsnet.net -2016-08-30 09:55:32,719 INFO logging.ThriftAccessLogger [] - Request ID: 1 access from: principal: operation: submitTopology -2016-08-30 09:55:32,757 INFO daemon.nimbus [] - Received topology submission for pirk_integration_test with conf {"topology.max.task.parallelism" nil, "qSchema" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml", "output" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt", "topology.executor.send.buffer.size" 1024, "topology.submitter.principal" "", "topology.acker.executors" 1, "topology.eventlogger.executors" 0, "queryInfo" {"dataPartitionsBitSize" 8, "numPartitionsPerDataElement" 36, "hashBitSize" 12, "queryType" "dn s-hostname-query", "embedSelector" true, "uuid" "74cbee54-9a78-48c3-905d-c83a0782b051", "hashKey" "someKey", "useExpLookupTable" false, "useHDFSExpLookupTable" false, "numBitsPerDataElement" 256, "numSelectors" 6}, "splitPartitions" true, "topology.workers" 1, "topology.debug" false, "topology.max.spout.pending" 300, "maxHitsPerSelector" 100, "storm.zookeeper.superACL" nil, "useHdfs" false, "limitHitsPerSelector" true, "hdfsUri" "localhost", "saltColumns" true, "topology.transfer.buffer.size" 32, "topology.executor.receive.buffer.size" 1024, "allowAdHocQuerySchemas" false, "queryFile" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/pir_integrationTest-query7554668182117678094.txt", "topology.users" (), "topology.submitter.user" "voyager", "topology.component.resources.onheap.memory.mb" 128.0, "rowDivisions" 2, "topology.kryo.register" nil, "worker.heap.memory.mb" 750, "nSquared" "893714063091493252140805543927462484189909163470256324669366233674468493934531351808411750089403379605 919371579520956313614309759596321404680678803731365842878590760838574242511806531716813184820975781862008105030373882425133453241", "enccolmultboltPar" 2, "encrowcalcboltPar" 2, "dschema" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml", "topology.kryo.decorators" (), "storm.id" "pirk_integration_test-1-1472565332", "topology.name" "pirk_integration_test"} -2016-08-30 09:55:32,773 INFO daemon.nimbus [] - uploadedJar -2016-08-30 09:55:32,794 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:32,795 INFO zookeeper.ZooKeeper [] - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@799bb4ed -2016-08-30 09:55:32,799 INFO zookeeper.ClientCnxn [] - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error) -2016-08-30 09:55:32,800 INFO zookeeper.ClientCnxn [] - Socket connection established to localhost/127.0.0.1:2000, initiating session -2016-08-30 09:55:32,801 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56313 -2016-08-30 09:55:32,802 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56313 -2016-08-30 09:55:32,803 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d000b with negotiated timeout 20000 for client /127.0.0.1:56313 -2016-08-30 09:55:32,803 INFO zookeeper.ClientCnxn [] - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000b, negotiated timeout = 20000 -2016-08-30 09:55:32,804 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:32,807 INFO server.PrepRequestProcessor [] - Got user-level KeeperException when processing sessionid:0x156dbbbca4d000b type:create cxid:0x2 zxid:0x22 txntype:-1 reqpath:n/a Error Path:/storm/blobstoremaxkeysequencenumber Error:KeeperErrorCode = NoNode for /storm/blobstoremaxkeysequencenumber -2016-08-30 09:55:32,815 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:32,817 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d000b -2016-08-30 09:55:32,818 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56313 which had sessionid 0x156dbbbca4d000b -2016-08-30 09:55:32,819 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d000b closed -2016-08-30 09:55:32,819 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:32,821 INFO storm.cluster [] - setup-path/blobstore/pirk_integration_test-1-1472565332-stormconf.ser/traveler.ltsnet.net:6627-1 -2016-08-30 09:55:32,829 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:32,830 INFO zookeeper.ZooKeeper [] - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@4e64e44c -2016-08-30 09:55:32,832 INFO zookeeper.ClientCnxn [] - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error) -2016-08-30 09:55:32,833 INFO zookeeper.ClientCnxn [] - Socket connection established to localhost/127.0.0.1:2000, initiating session -2016-08-30 09:55:32,833 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56314 -2016-08-30 09:55:32,833 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56314 -2016-08-30 09:55:32,834 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d000c with negotiated timeout 20000 for client /127.0.0.1:56314 -2016-08-30 09:55:32,834 INFO zookeeper.ClientCnxn [] - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000c, negotiated timeout = 20000 -2016-08-30 09:55:32,834 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:32,838 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:32,839 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d000c -2016-08-30 09:55:32,840 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56314 which had sessionid 0x156dbbbca4d000c -2016-08-30 09:55:32,840 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:32,841 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d000c closed -2016-08-30 09:55:32,841 INFO storm.cluster [] - setup-path/blobstore/pirk_integration_test-1-1472565332-stormcode.ser/traveler.ltsnet.net:6627-1 -2016-08-30 09:55:32,852 INFO daemon.nimbus [] - desired replication count of 1 not achieved but we have hit the max wait time 60 so moving on with replication count for conf key = 1 for code key = 1for jar key = 1 -2016-08-30 09:55:32,919 INFO daemon.nimbus [] - Activating pirk_integration_test: pirk_integration_test-1-1472565332 -2016-08-30 09:55:32,924 INFO storm.KafkaStormIntegrationTest [] - Pausing for setup. -2016-08-30 09:55:33,671 INFO scheduler.EvenScheduler [] - Available slots: (["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1025] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1026]) -2016-08-30 09:55:33,674 INFO daemon.nimbus [] - Setting new assignment for topology id pirk_integration_test-1-1472565332: #org.apache.storm.daemon.common.Assignment{:master-code-dir "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60", :node->host {"59a01bf6-2a07-4e98-a935-3031a25e93b7" "traveler.ltsnet.net"}, :executor->node+port {[8 8] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [7 7] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [6 6] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [5 5] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [4 4] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [3 3] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [2 2] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024], [1 1] ["59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024]}, :executor->start-time-secs {[1 1] 1472565333, [2 2] 1472565333, [3 3] 1472565333, [4 4] 1472565333, [5 5] 1472565333, [6 6] 1472565333, [7 7] 1472565333, [8 8] 1472565333}, :worker->resources {[" 59a01bf6-2a07-4e98-a935-3031a25e93b7" 1024] [0.0 0.0 0.0]}} -2016-08-30 09:55:33,689 INFO daemon.supervisor [] - Downloading code for storm id pirk_integration_test-1-1472565332 -2016-08-30 09:55:33,690 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:33,691 INFO zookeeper.ZooKeeper [] - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@1537621d -2016-08-30 09:55:33,693 INFO zookeeper.ClientCnxn [] - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error) -2016-08-30 09:55:33,693 INFO zookeeper.ClientCnxn [] - Socket connection established to localhost/127.0.0.1:2000, initiating session -2016-08-30 09:55:33,693 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56315 -2016-08-30 09:55:33,694 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56315 -2016-08-30 09:55:33,695 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d000d with negotiated timeout 20000 for client /127.0.0.1:56315 -2016-08-30 09:55:33,695 INFO zookeeper.ClientCnxn [] - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000d, negotiated timeout = 20000 -2016-08-30 09:55:33,695 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:33,699 INFO blobstore.FileBlobStoreImpl [] - Creating new blob store based in /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60/blobs -2016-08-30 09:55:33,727 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:33,730 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d000d -2016-08-30 09:55:33,731 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:33,731 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d000d closed -2016-08-30 09:55:33,731 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56315 which had sessionid 0x156dbbbca4d000d -2016-08-30 09:55:34,000 INFO server.SessionTrackerImpl [] - SessionTrackerImpl exited loop! -2016-08-30 09:55:34,763 INFO daemon.supervisor [] - Finished downloading code for storm id pirk_integration_test-1-1472565332 -2016-08-30 09:55:34,770 INFO daemon.supervisor [] - Launching worker with assignment {:storm-id "pirk_integration_test-1-1472565332", :executors [[8 8] [7 7] [6 6] [5 5] [4 4] [3 3] [2 2] [1 1]], :resources #object[org.apache.storm.generated.WorkerResources 0x6ff2384a "WorkerResources(mem_on_heap:0.0, mem_off_heap:0.0, cpu:0.0)"]} for this supervisor 59a01bf6-2a07-4e98-a935-3031a25e93b7 on port 1024 with id cacfc67b-c838-449e-8e35-dca6d3f1da94 -2016-08-30 09:55:34,773 INFO daemon.worker [] - Launching worker for pirk_integration_test-1-1472565332 on 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024 with id cacfc67b-c838-449e-8e35-dca6d3f1da94 and conf {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "local", "storm.messaging.netty.client_worker_threads" 1, "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "qSchema" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file66 01659287734868610.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml", "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "output" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt", "storm.group.mapping.service.params" n il, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "supervisor.worker.shutdown.sleep.secs" 1, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 0, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2000, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" false, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "drpc.queue.size" 128, "topology.backpressure.enable" true, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "topology.scheduler.strategy" "org.apach e.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "resource.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" 1, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["localhost"], "nimbus.thrift.threads" 64, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topolog y.classpath" nil, "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.secs" 600, "topology.skip.missing.kryo.registrations" true, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "queryInfo" {"useHDFSExpLookupTable" false, "hashKey" "someKey", "numSelectors" 6, "dataPartitionsBitSize" 8, "hashBitSize" 12, "numPartitionsPerDataElement" 36, "numBitsPerDataElement" 256, "uuid" "74cbee54-9a78-48c3-905d-c83a0782b051", "useExpLookupTable" false, "embedSelector" true, "queryType" "dns-hostname-query"}, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "splitPartitions" true, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/var/folders/83/07wq_8t55 pn3mgj99512hb8c0000gn/T//960b6925-fc51-4264-984e-d7937568606f", "topology.disable.loadaware" false, "worker.childopts" "-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.messaging.netty.max_retries" 300, "ui.childopts" "-Xmx768m", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMap ping", "storm.zookeeper.session.timeout" 20000, "drpc.childopts" "-Xmx768m", "drpc.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 15000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" 300, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "maxHitsPerSelector" 100, "nimbus.task.timeout.secs" 30, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "useHdfs" false, "storm.zookeeper.retry.interval" 1000, "limitHitsPerSelector" true, "logs.users" nil, "wor ker.profiler.command" "flight.bash", "transactional.zookeeper.port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "task.credentials.poll.secs" 30, "blobstore.superuser" "voyager", "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "hdfsUri" "localhost", "supervisor.slots.ports" (1024 1025 1026), "saltColumns" true, "topology.transfer.buffer.size" 32, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "allowAdHocQuerySchemas" false, "queryFile" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/pir_integrationTest-query7554668182117678094.txt", "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.mess aging.netty.buffer_size" 5242880, "storm.cluster.state.store" "org.apache.storm.cluster_state.zookeeper_state_factory", "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "nimbus.impersonation.authorizer" "org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer", "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "storm.messaging.netty.authentication" false, "topology.component.resources. onheap.memory.mb" 128.0, "rowDivisions" 2, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["localhost"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 750, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "nSquared" "893714063091493252140805543927462484189909163470256324669366233674468493934531351808411750089403379605919371579520956313614309759596321404680678803731365842878590760838574242511806531716813184820975781862008105030373882425133453241", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "supervisor.childopts" "-Xmx256m", "topology.worker.max.heap.size.mb" 768.0, "b ackpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "topology.disruptor.wait.timeout.millis" 1000, "enccolmultboltPar" 2, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 50, "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib", "encrowcalcboltPar" 2, "ui.port" 8080, "dschema" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml", "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600} -2016-08-30 09:55:34,777 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:34,777 INFO zookeeper.ZooKeeper [] - Initiating client connection, connectString=localhost:2000 sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@5ec68ddc -2016-08-30 09:55:34,779 INFO zookeeper.ClientCnxn [] - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error) -2016-08-30 09:55:34,780 INFO zookeeper.ClientCnxn [] - Socket connection established to localhost/127.0.0.1:2000, initiating session -2016-08-30 09:55:34,780 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56316 -2016-08-30 09:55:34,780 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56316 -2016-08-30 09:55:34,781 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d000e with negotiated timeout 20000 for client /127.0.0.1:56316 -2016-08-30 09:55:34,781 INFO zookeeper.ClientCnxn [] - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000e, negotiated timeout = 20000 -2016-08-30 09:55:34,781 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:34,781 INFO storm.zookeeper [] - Zookeeper state update: :connected:none -2016-08-30 09:55:34,782 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:34,783 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d000e -2016-08-30 09:55:34,783 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d000e closed -2016-08-30 09:55:34,783 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:34,784 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56316 which had sessionid 0x156dbbbca4d000e -2016-08-30 09:55:34,784 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:34,784 INFO zookeeper.ZooKeeper [] - Initiating client connection, connectString=localhost:2000/storm sessionTimeout=20000 watcher=org.apache.storm.shade.org.apache.curator.ConnectionState@2b377b2c -2016-08-30 09:55:34,786 INFO zookeeper.ClientCnxn [] - Opening socket connection to server localhost/127.0.0.1:2000. Will not attempt to authenticate using SASL (unknown error) -2016-08-30 09:55:34,786 INFO zookeeper.ClientCnxn [] - Socket connection established to localhost/127.0.0.1:2000, initiating session -2016-08-30 09:55:34,786 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56317 -2016-08-30 09:55:34,787 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56317 -2016-08-30 09:55:34,788 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d000f with negotiated timeout 20000 for client /127.0.0.1:56317 -2016-08-30 09:55:34,788 INFO zookeeper.ClientCnxn [] - Session establishment complete on server localhost/127.0.0.1:2000, sessionid = 0x156dbbbca4d000f, negotiated timeout = 20000 -2016-08-30 09:55:34,788 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:34,793 INFO auth.AuthUtils [] - Got AutoCreds [] -2016-08-30 09:55:34,793 INFO daemon.worker [] - Reading Assignments. -2016-08-30 09:55:34,806 INFO daemon.worker [] - Registering IConnectionCallbacks for 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024 -2016-08-30 09:55:34,818 INFO daemon.executor [] - Loading executor partitiondataBolt:[8 8] -2016-08-30 09:55:34,819 INFO daemon.executor [] - Loaded executor tasks partitiondataBolt:[8 8] -2016-08-30 09:55:34,824 INFO daemon.executor [] - Finished loading executor partitiondataBolt:[8 8] -2016-08-30 09:55:34,829 INFO daemon.executor [] - Loading executor enccolmultbolt:[2 2] -2016-08-30 09:55:34,830 INFO daemon.executor [] - Loaded executor tasks enccolmultbolt:[2 2] -2016-08-30 09:55:34,832 INFO daemon.executor [] - Finished loading executor enccolmultbolt:[2 2] -2016-08-30 09:55:34,836 INFO daemon.executor [] - Loading executor outputbolt:[7 7] -2016-08-30 09:55:34,837 INFO daemon.executor [] - Loaded executor tasks outputbolt:[7 7] -2016-08-30 09:55:34,839 INFO daemon.executor [] - Finished loading executor outputbolt:[7 7] -2016-08-30 09:55:34,844 INFO daemon.executor [] - Loading executor enccolmultbolt:[3 3] -2016-08-30 09:55:34,845 INFO daemon.executor [] - Loaded executor tasks enccolmultbolt:[3 3] -2016-08-30 09:55:34,848 INFO daemon.executor [] - Finished loading executor enccolmultbolt:[3 3] -2016-08-30 09:55:34,852 INFO daemon.executor [] - Loading executor __acker:[1 1] -2016-08-30 09:55:34,853 INFO daemon.executor [] - Loaded executor tasks __acker:[1 1] -2016-08-30 09:55:34,854 INFO daemon.executor [] - Timeouts disabled for executor __acker:[1 1] -2016-08-30 09:55:34,855 INFO daemon.executor [] - Finished loading executor __acker:[1 1] -2016-08-30 09:55:34,859 INFO daemon.executor [] - Loading executor kafkaspout:[6 6] -2016-08-30 09:55:34,861 INFO daemon.executor [] - Loaded executor tasks kafkaspout:[6 6] -2016-08-30 09:55:34,864 INFO daemon.executor [] - Timeouts disabled for executor kafkaspout:[6 6] -2016-08-30 09:55:34,865 INFO daemon.executor [] - Finished loading executor kafkaspout:[6 6] -2016-08-30 09:55:34,869 INFO daemon.executor [] - Loading executor __system:[-1 -1] -2016-08-30 09:55:34,870 INFO daemon.executor [] - Loaded executor tasks __system:[-1 -1] -2016-08-30 09:55:34,873 INFO daemon.executor [] - Finished loading executor __system:[-1 -1] -2016-08-30 09:55:34,877 INFO daemon.executor [] - Loading executor encrowcalcbolt:[5 5] -2016-08-30 09:55:34,878 INFO daemon.executor [] - Loaded executor tasks encrowcalcbolt:[5 5] -2016-08-30 09:55:34,880 INFO daemon.executor [] - Finished loading executor encrowcalcbolt:[5 5] -2016-08-30 09:55:34,885 INFO daemon.executor [] - Loading executor encrowcalcbolt:[4 4] -2016-08-30 09:55:34,886 INFO daemon.executor [] - Loaded executor tasks encrowcalcbolt:[4 4] -2016-08-30 09:55:34,888 INFO daemon.executor [] - Finished loading executor encrowcalcbolt:[4 4] -2016-08-30 09:55:34,890 INFO daemon.worker [] - Started with log levels: {"org.apache.hadoop.hbase" #object[org.apache.logging.log4j.Level 0x7af08580 "INFO"], "" #object[org.apache.logging.log4j.Level 0x7af08580 "INFO"], "org.apache.hadoop.hbase.zookeeper" #object[org.apache.logging.log4j.Level 0x6d660c48 "WARN"], "org.apache.zookeeper" #object[org.apache.logging.log4j.Level 0xf759057 "ERROR"]} -2016-08-30 09:55:34,893 INFO daemon.worker [] - Worker has topology config {"topology.builtin.metrics.bucket.size.secs" 60, "nimbus.childopts" "-Xmx1024m", "ui.filter.params" nil, "storm.cluster.mode" "local", "storm.messaging.netty.client_worker_threads" 1, "logviewer.max.per.worker.logs.size.mb" 2048, "supervisor.run.worker.as.user" false, "topology.max.task.parallelism" nil, "topology.priority" 29, "zmq.threads" 1, "storm.group.mapping.service" "org.apache.storm.security.auth.ShellBasedGroupsMapping", "transactional.zookeeper.root" "/transactional", "topology.sleep.spout.wait.strategy.time.ms" 1, "scheduler.display.resource" false, "topology.max.replication.wait.time.sec" 60, "qSchema" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml,/var/folde rs/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml,/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml", "drpc.invocations.port" 3773, "supervisor.localizer.cache.target.size.mb" 10240, "topology.multilang.serializer" "org.apache.storm.multilang.JsonSerializer", "storm.messaging.netty.server_worker_threads" 1, "nimbus.blobstore.class" "org.apache.storm.blobstore.LocalFsBlobStore", "resource.aware.scheduler.eviction.strategy" "org.apache.storm.scheduler.resource.strategies.eviction.DefaultEvictionStrategy", "topology.max.error.report.per.interval" 5, "storm.thrift.transport" "org.apache.storm.security.auth.SimpleTransportPlugin", "zmq.hwm" 0, "output" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt", "storm.group.mapping.service.params" nil, "worker.profiler.enabled" false, "storm.principal.tolocal" "org.apache.storm.security.auth.DefaultPrincipalToLocal", "superv isor.worker.shutdown.sleep.secs" 1, "pacemaker.host" "localhost", "storm.zookeeper.retry.times" 5, "ui.actions.enabled" true, "zmq.linger.millis" 0, "supervisor.enable" true, "topology.stats.sample.rate" 0.05, "storm.messaging.netty.min_wait_ms" 100, "worker.log.level.reset.poll.secs" 30, "storm.zookeeper.port" 2000, "supervisor.heartbeat.frequency.secs" 5, "topology.enable.message.timeouts" false, "supervisor.cpu.capacity" 400.0, "drpc.worker.threads" 64, "supervisor.blobstore.download.thread.count" 5, "drpc.queue.size" 128, "topology.backpressure.enable" true, "supervisor.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "storm.blobstore.inputstream.buffer.size.bytes" 65536, "topology.shellbolt.max.pending" 100, "drpc.https.keystore.password" "", "nimbus.code.sync.freq.secs" 120, "logviewer.port" 8000, "topology.scheduler.strategy" "org.apache.storm.scheduler.resource.strategies.scheduling.DefaultResourceAwareStrategy", "topology.executor.send.buffer.size" 1024, "reso urce.aware.scheduler.priority.strategy" "org.apache.storm.scheduler.resource.strategies.priority.DefaultSchedulingPriorityStrategy", "pacemaker.auth.method" "NONE", "storm.daemon.metrics.reporter.plugins" ["org.apache.storm.daemon.metrics.reporters.JmxPreparableReporter"], "topology.worker.logwriter.childopts" "-Xmx64m", "topology.spout.wait.strategy" "org.apache.storm.spout.SleepSpoutWaitStrategy", "ui.host" "0.0.0.0", "topology.submitter.principal" "", "storm.nimbus.retry.interval.millis" 2000, "nimbus.inbox.jar.expiration.secs" 3600, "dev.zookeeper.path" "/tmp/dev-storm-zookeeper", "topology.acker.executors" 1, "topology.fall.back.on.java.serialization" true, "topology.eventlogger.executors" 0, "supervisor.localizer.cleanup.interval.ms" 600000, "storm.zookeeper.servers" ["localhost"], "nimbus.thrift.threads" 64, "logviewer.cleanup.age.mins" 10080, "topology.worker.childopts" nil, "topology.classpath" nil, "supervisor.monitor.frequency.secs" 3, "nimbus.credential.renewers.freq.sec s" 600, "topology.skip.missing.kryo.registrations" true, "drpc.authorizer.acl.filename" "drpc-auth-acl.yaml", "pacemaker.kerberos.users" [], "storm.group.mapping.service.cache.duration.secs" 120, "topology.testing.always.try.serialize" false, "nimbus.monitor.freq.secs" 10, "queryInfo" {"dataPartitionsBitSize" 8, "numPartitionsPerDataElement" 36, "hashBitSize" 12, "queryType" "dns-hostname-query", "embedSelector" true, "uuid" "74cbee54-9a78-48c3-905d-c83a0782b051", "hashKey" "someKey", "useExpLookupTable" false, "useHDFSExpLookupTable" false, "numBitsPerDataElement" 256, "numSelectors" 6}, "storm.health.check.timeout.ms" 5000, "supervisor.supervisors" [], "topology.tasks" nil, "topology.bolts.outgoing.overflow.buffer.enable" false, "splitPartitions" true, "storm.messaging.netty.socket.backlog" 500, "topology.workers" 1, "pacemaker.base.threads" 10, "storm.local.dir" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60", "topology.disable.loadaware" false, "worker.childopts" "-Xmx%HEAP-MEM%m -XX:+PrintGCDetails -Xloggc:artifacts/gc.log -XX:+PrintGCDateStamps -XX:+PrintGCTimeStamps -XX:+UseGCLogFileRotation -XX:NumberOfGCLogFiles=10 -XX:GCLogFileSize=1M -XX:+HeapDumpOnOutOfMemoryError -XX:HeapDumpPath=artifacts/heapdump", "storm.auth.simple-white-list.users" [], "topology.disruptor.batch.timeout.millis" 1, "topology.message.timeout.secs" 30, "topology.state.synchronization.timeout.secs" 60, "topology.tuple.serializer" "org.apache.storm.serialization.types.ListDelegateSerializer", "supervisor.supervisors.commands" [], "nimbus.blobstore.expiration.secs" 600, "logviewer.childopts" "-Xmx128m", "topology.environment" nil, "topology.debug" false, "topology.disruptor.batch.size" 100, "storm.messaging.netty.max_retries" 300, "ui.childopts" "-Xmx768m", "storm.network.topography.plugin" "org.apache.storm.networktopography.DefaultRackDNSToSwitchMapping", "storm.zookeeper.session.timeout" 20000, "drpc.childopts" "-Xmx768m", "drpc.http.creds .plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.connection.timeout" 15000, "storm.zookeeper.auth.user" nil, "storm.meta.serialization.delegate" "org.apache.storm.serialization.GzipThriftSerializationDelegate", "topology.max.spout.pending" 300, "storm.codedistributor.class" "org.apache.storm.codedistributor.LocalFileSystemCodeDistributor", "nimbus.supervisor.timeout.secs" 60, "maxHitsPerSelector" 100, "nimbus.task.timeout.secs" 30, "storm.zookeeper.superACL" nil, "drpc.port" 3772, "pacemaker.max.threads" 50, "storm.zookeeper.retry.intervalceiling.millis" 30000, "nimbus.thrift.port" 6627, "storm.auth.simple-acl.admins" [], "topology.component.cpu.pcore.percent" 10.0, "supervisor.memory.capacity.mb" 3072.0, "storm.nimbus.retry.times" 5, "supervisor.worker.start.timeout.secs" 120, "useHdfs" false, "storm.zookeeper.retry.interval" 1000, "limitHitsPerSelector" true, "logs.users" nil, "worker.profiler.command" "flight.bash", "transactional.zookeeper .port" nil, "drpc.max_buffer_size" 1048576, "pacemaker.thread.timeout" 10, "task.credentials.poll.secs" 30, "blobstore.superuser" "voyager", "drpc.https.keystore.type" "JKS", "topology.worker.receiver.thread.count" 1, "topology.state.checkpoint.interval.ms" 1000, "hdfsUri" "localhost", "supervisor.slots.ports" [6700 6701 6702 6703], "saltColumns" true, "topology.transfer.buffer.size" 32, "storm.health.check.dir" "healthchecks", "topology.worker.shared.thread.pool.size" 4, "drpc.authorizer.acl.strict" false, "nimbus.file.copy.expiration.secs" 600, "worker.profiler.childopts" "-XX:+UnlockCommercialFeatures -XX:+FlightRecorder", "topology.executor.receive.buffer.size" 1024, "backpressure.disruptor.low.watermark" 0.4, "allowAdHocQuerySchemas" false, "queryFile" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/pir_integrationTest-query7554668182117678094.txt", "topology.users" [], "nimbus.task.launch.secs" 120, "storm.local.mode.zmq" false, "storm.messaging.netty.buffer_size" 5242880, " storm.cluster.state.store" "org.apache.storm.cluster_state.zookeeper_state_factory", "worker.heartbeat.frequency.secs" 1, "storm.log4j2.conf.dir" "log4j2", "ui.http.creds.plugin" "org.apache.storm.security.auth.DefaultHttpCredentialsPlugin", "storm.zookeeper.root" "/storm", "topology.submitter.user" "voyager", "topology.tick.tuple.freq.secs" nil, "drpc.https.port" -1, "storm.workers.artifacts.dir" "workers-artifacts", "supervisor.blobstore.download.max_retries" 3, "task.refresh.poll.secs" 10, "storm.exhibitor.port" 8080, "task.heartbeat.frequency.secs" 3, "pacemaker.port" 6699, "storm.messaging.netty.max_wait_ms" 1000, "nimbus.impersonation.authorizer" "org.apache.storm.security.auth.authorizer.ImpersonationAuthorizer", "topology.component.resources.offheap.memory.mb" 0.0, "drpc.http.port" 3774, "topology.error.throttle.interval.secs" 10, "storm.messaging.transport" "org.apache.storm.messaging.netty.Context", "storm.messaging.netty.authentication" false, "topology.component.resource s.onheap.memory.mb" 128.0, "rowDivisions" 2, "topology.kryo.factory" "org.apache.storm.serialization.DefaultKryoFactory", "topology.kryo.register" nil, "worker.gc.childopts" "", "nimbus.topology.validator" "org.apache.storm.nimbus.DefaultTopologyValidator", "nimbus.seeds" ["localhost"], "nimbus.queue.size" 100000, "nimbus.cleanup.inbox.freq.secs" 600, "storm.blobstore.replication.factor" 3, "worker.heap.memory.mb" 750, "logviewer.max.sum.worker.logs.size.mb" 4096, "pacemaker.childopts" "-Xmx1024m", "nSquared" "893714063091493252140805543927462484189909163470256324669366233674468493934531351808411750089403379605919371579520956313614309759596321404680678803731365842878590760838574242511806531716813184820975781862008105030373882425133453241", "ui.users" nil, "transactional.zookeeper.servers" nil, "supervisor.worker.timeout.secs" 30, "storm.zookeeper.auth.password" nil, "client.blobstore.class" "org.apache.storm.blobstore.NimbusBlobStore", "supervisor.childopts" "-Xmx256m", "topology.wo rker.max.heap.size.mb" 768.0, "backpressure.disruptor.high.watermark" 0.9, "ui.filter" nil, "ui.header.buffer.bytes" 4096, "topology.min.replication.count" 1, "topology.disruptor.wait.timeout.millis" 1000, "enccolmultboltPar" 2, "storm.nimbus.retry.intervalceiling.millis" 60000, "topology.trident.batch.emit.interval.millis" 50, "storm.auth.simple-acl.users" [], "drpc.invocations.threads" 64, "java.library.path" "/usr/local/lib:/opt/local/lib:/usr/lib", "encrowcalcboltPar" 2, "ui.port" 8080, "dschema" "/var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml", "topology.kryo.decorators" [], "storm.id" "pirk_integration_test-1-1472565332", "topology.name" "pirk_integration_test", "storm.exhibitor.poll.uripath" "/exhibitor/v1/cluster/list", "storm.messaging.netty.transfer.batch.size" 262144, "logviewer.appender.name" "A1", "nimbus.thrift.max_buffer_size" 1048576, "storm.auth.simple-acl.users.commands" [], "drpc.request.timeout.secs" 600} -2016-08-30 09:55:34,894 INFO daemon.worker [] - Worker cacfc67b-c838-449e-8e35-dca6d3f1da94 for storm pirk_integration_test-1-1472565332 on 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024 has finished loading -2016-08-30 09:55:34,894 INFO storm.config [] - SET worker-user cacfc67b-c838-449e-8e35-dca6d3f1da94 -2016-08-30 09:55:35,801 INFO daemon.worker [] - All connections are ready for worker 59a01bf6-2a07-4e98-a935-3031a25e93b7:1024 with id cacfc67b-c838-449e-8e35-dca6d3f1da94 -2016-08-30 09:55:35,844 INFO daemon.executor [] - Preparing bolt partitiondataBolt:(8) -2016-08-30 09:55:35,845 INFO storm.PartitionDataBolt [] - partition databolt hdfs = false -2016-08-30 09:55:35,845 INFO data.DataSchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml hdfs = false -2016-08-30 09:55:35,845 INFO data.DataSchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml -2016-08-30 09:55:35,846 INFO data.DataSchemaLoader [] - Root element: schema -2016-08-30 09:55:35,847 INFO data.DataSchemaLoader [] - schemaName = testDataSchema -2016-08-30 09:55:35,847 INFO data.DataSchemaLoader [] - name = date javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner -2016-08-30 09:55:35,847 INFO data.DataSchemaLoader [] - name = qname javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,847 INFO data.DataSchemaLoader [] - name = src_ip javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,847 INFO data.DataSchemaLoader [] - name = dest_ip javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,848 INFO data.DataSchemaLoader [] - name = qtype javaType = short isArray = true partitioner org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,848 INFO data.DataSchemaLoader [] - name = rcode javaType = int isArray = false partitioner org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,848 INFO data.DataSchemaLoader [] - name = ip javaType = string isArray = true partitioner org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,848 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml -2016-08-30 09:55:35,848 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml -2016-08-30 09:55:35,849 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,849 INFO query.QuerySchemaLoader [] - schemaName = dns-hostname-query -2016-08-30 09:55:35,849 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - selectorName = qname -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = date partitionerName = org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = date bits = 64 dataElementSize = 64 -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = src_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = src_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 128 -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = qtype partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,850 INFO query.QuerySchemaLoader [] - name = qtype bits = 32 dataElementSize = 160 -2016-08-30 09:55:35,851 INFO query.QuerySchemaLoader [] - name = rcode partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,851 INFO query.QuerySchemaLoader [] - name = rcode bits = 32 dataElementSize = 192 -2016-08-30 09:55:35,851 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,851 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 256 -2016-08-30 09:55:35,851 INFO query.QuerySchemaLoader [] - filterName = qname -2016-08-30 09:55:35,858 INFO daemon.executor [] - Preparing bolt outputbolt:(7) -2016-08-30 09:55:35,859 INFO storm.OutputBolt [] - Intitialized OutputBolt. -2016-08-30 09:55:35,859 INFO daemon.executor [] - Prepared bolt outputbolt:(7) -2016-08-30 09:55:35,859 INFO daemon.executor [] - Preparing bolt enccolmultbolt:(2) -2016-08-30 09:55:35,859 INFO storm.EncColMultBolt [] - Initialized EncColMultBolt. -2016-08-30 09:55:35,860 INFO daemon.executor [] - Prepared bolt enccolmultbolt:(2) -2016-08-30 09:55:35,865 INFO daemon.executor [] - Preparing bolt enccolmultbolt:(3) -2016-08-30 09:55:35,866 INFO storm.EncColMultBolt [] - Initialized EncColMultBolt. -2016-08-30 09:55:35,866 INFO daemon.executor [] - Prepared bolt enccolmultbolt:(3) -2016-08-30 09:55:35,872 INFO daemon.executor [] - Preparing bolt __acker:(1) -2016-08-30 09:55:35,873 INFO daemon.executor [] - Prepared bolt __acker:(1) -2016-08-30 09:55:35,874 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml -2016-08-30 09:55:35,874 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - schemaName = dns-ip-query -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - selectorName = ip -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - name = src_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - name = src_ip bits = 32 dataElementSize = 32 -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,876 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 64 -2016-08-30 09:55:35,877 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,877 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 128 -2016-08-30 09:55:35,877 INFO query.QuerySchemaLoader [] - filterName = qname -2016-08-30 09:55:35,887 INFO daemon.executor [] - Opening spout kafkaspout:(6) -2016-08-30 09:55:35,888 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:35,892 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:35,894 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml -2016-08-30 09:55:35,894 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml -2016-08-30 09:55:35,896 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,897 INFO query.QuerySchemaLoader [] - schemaName = dns-nxdomain-query -2016-08-30 09:55:35,897 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,897 INFO query.QuerySchemaLoader [] - selectorName = qname -2016-08-30 09:55:35,897 INFO query.QuerySchemaLoader [] - name = qname partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,898 INFO query.QuerySchemaLoader [] - name = qname bits = 64 dataElementSize = 64 -2016-08-30 09:55:35,898 INFO query.QuerySchemaLoader [] - name = src_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,898 INFO query.QuerySchemaLoader [] - name = src_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:35,898 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,898 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 128 -2016-08-30 09:55:35,898 INFO query.QuerySchemaLoader [] - filterName = qname -2016-08-30 09:55:35,899 INFO server.NIOServerCnxnFactory [] - Accepted socket connection from /127.0.0.1:56318 -2016-08-30 09:55:35,899 INFO server.ZooKeeperServer [] - Client attempting to establish new session at /127.0.0.1:56318 -2016-08-30 09:55:35,900 INFO daemon.executor [] - Preparing bolt __system:(-1) -2016-08-30 09:55:35,900 INFO daemon.executor [] - Preparing bolt encrowcalcbolt:(5) -2016-08-30 09:55:35,901 INFO storm.EncRowCalcBolt [] - partition databolt hdfs = false -2016-08-30 09:55:35,901 INFO server.ZooKeeperServer [] - Established session 0x156dbbbca4d0010 with negotiated timeout 20000 for client /127.0.0.1:56318 -2016-08-30 09:55:35,901 INFO storm.EncRowCalcBolt [] - Initialized EncRowCalcBolt. -2016-08-30 09:55:35,902 INFO daemon.executor [] - Prepared bolt encrowcalcbolt:(5) -2016-08-30 09:55:35,902 INFO daemon.executor [] - Prepared bolt __system:(-1) -2016-08-30 09:55:35,902 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:35,902 INFO daemon.executor [] - Preparing bolt encrowcalcbolt:(4) -2016-08-30 09:55:35,902 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:35,902 INFO storm.EncRowCalcBolt [] - partition databolt hdfs = false -2016-08-30 09:55:35,902 INFO storm.EncRowCalcBolt [] - Initialized EncRowCalcBolt. -2016-08-30 09:55:35,902 INFO daemon.executor [] - Prepared bolt encrowcalcbolt:(4) -2016-08-30 09:55:35,906 INFO kafka.DynamicBrokersReader [] - Read partition info from zookeeper: GlobalPartitionInformation{topic=pirk_test_topic, partitionMap={0=localhost:11111}} -2016-08-30 09:55:35,906 INFO imps.CuratorFrameworkImpl [] - Starting -2016-08-30 09:55:35,908 INFO daemon.executor [] - Opened spout kafkaspout:(6) -2016-08-30 09:55:35,909 INFO daemon.executor [] - Activating spout kafkaspout:(6) -2016-08-30 09:55:35,909 INFO kafka.ZkCoordinator [] - Task [1/1] Refreshing partition manager connections -2016-08-30 09:55:35,918 INFO state.ConnectionStateManager [] - State change: CONNECTED -2016-08-30 09:55:35,920 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml -2016-08-30 09:55:35,920 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - schemaName = dns-srcip-query -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - selectorName = src_ip -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - name = qname partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - name = qname bits = 64 dataElementSize = 64 -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,922 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:35,923 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,923 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 160 -2016-08-30 09:55:35,923 INFO query.QuerySchemaLoader [] - filterName = src_ip -2016-08-30 09:55:35,923 INFO query.QuerySchemaLoader [] - filterName = ip -2016-08-30 09:55:35,923 INFO kafka.DynamicBrokersReader [] - Read partition info from zookeeper: GlobalPartitionInformation{topic=pirk_test_topic, partitionMap={0=localhost:11111}} -2016-08-30 09:55:35,923 INFO kafka.KafkaUtils [] - Task [1/1] assigned [Partition{host=localhost:11111, topic=pirk_test_topic, partition=0}] -2016-08-30 09:55:35,923 INFO kafka.ZkCoordinator [] - Task [1/1] Deleted partition managers: [] -2016-08-30 09:55:35,923 INFO kafka.ZkCoordinator [] - Task [1/1] New partition managers: [Partition{host=localhost:11111, topic=pirk_test_topic, partition=0}] -2016-08-30 09:55:35,925 INFO kafka.PartitionManager [] - Read partition information from: /pirk_test_root/pirk_integr_test_spout/partition_0 --> null -2016-08-30 09:55:35,933 INFO kafka.PartitionManager [] - No partition information found, using configuration to determine offset -2016-08-30 09:55:35,934 INFO kafka.PartitionManager [] - Last commit offset from zookeeper: 0 -2016-08-30 09:55:35,934 INFO kafka.PartitionManager [] - Commit offset 0 is more than 9223372036854775807 behind latest offset 0, resetting to startOffsetTime=-2 -2016-08-30 09:55:35,934 INFO kafka.PartitionManager [] - Starting Kafka localhost:0 from offset 0 -2016-08-30 09:55:35,934 INFO kafka.ZkCoordinator [] - Task [1/1] Finished refreshing -2016-08-30 09:55:35,939 INFO data.DataSchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml hdfs = false -2016-08-30 09:55:35,939 INFO data.DataSchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/testDataSchemaFile5716412757641366065.xml -2016-08-30 09:55:35,941 INFO data.DataSchemaLoader [] - Root element: schema -2016-08-30 09:55:35,941 INFO data.DataSchemaLoader [] - schemaName = testDataSchema -2016-08-30 09:55:35,941 INFO data.DataSchemaLoader [] - name = date javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner -2016-08-30 09:55:35,941 INFO data.DataSchemaLoader [] - name = qname javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,942 INFO data.DataSchemaLoader [] - name = src_ip javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,942 INFO data.DataSchemaLoader [] - name = dest_ip javaType = string isArray = false partitioner org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,942 INFO data.DataSchemaLoader [] - name = qtype javaType = short isArray = true partitioner org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,942 INFO data.DataSchemaLoader [] - name = rcode javaType = int isArray = false partitioner org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,942 INFO data.DataSchemaLoader [] - name = ip javaType = string isArray = true partitioner org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,942 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml -2016-08-30 09:55:35,942 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-hostname-query_file4532447393685025940.xml -2016-08-30 09:55:35,943 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml -2016-08-30 09:55:35,943 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml -2016-08-30 09:55:35,943 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,943 INFO query.QuerySchemaLoader [] - schemaName = dns-hostname-query -2016-08-30 09:55:35,943 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - selectorName = qname -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = date partitionerName = org.apache.pirk.schema.data.partitioner.ISO8601DatePartitioner -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = date bits = 64 dataElementSize = 64 -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = src_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = src_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - schemaName = dns-srcip-query-no-filter -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 128 -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - selectorName = src_ip -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = qname partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,944 INFO query.QuerySchemaLoader [] - name = qname bits = 64 dataElementSize = 64 -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = qtype partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = qtype bits = 32 dataElementSize = 160 -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = rcode partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 160 -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = rcode bits = 32 dataElementSize = 192 -2016-08-30 09:55:35,945 INFO storm.PartitionDataBolt [] - Initialized ExtractAndPartitionDataBolt. -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 256 -2016-08-30 09:55:35,945 INFO query.QuerySchemaLoader [] - filterName = qname -2016-08-30 09:55:35,946 INFO daemon.executor [] - Prepared bolt partitiondataBolt:(8) -2016-08-30 09:55:35,963 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml -2016-08-30 09:55:35,963 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-ip-query_file6601659287734868610.xml -2016-08-30 09:55:35,964 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,964 INFO query.QuerySchemaLoader [] - schemaName = dns-ip-query -2016-08-30 09:55:35,964 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - selectorName = ip -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - name = src_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - name = src_ip bits = 32 dataElementSize = 32 -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 64 -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 128 -2016-08-30 09:55:35,965 INFO query.QuerySchemaLoader [] - filterName = qname -2016-08-30 09:55:35,981 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml -2016-08-30 09:55:35,981 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-nxdomain-query_file2033186626462123619.xml -2016-08-30 09:55:35,983 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:35,983 INFO query.QuerySchemaLoader [] - schemaName = dns-nxdomain-query -2016-08-30 09:55:35,983 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:35,983 INFO query.QuerySchemaLoader [] - selectorName = qname -2016-08-30 09:55:35,983 INFO query.QuerySchemaLoader [] - name = qname partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:35,983 INFO query.QuerySchemaLoader [] - name = qname bits = 64 dataElementSize = 64 -2016-08-30 09:55:35,984 INFO query.QuerySchemaLoader [] - name = src_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,984 INFO query.QuerySchemaLoader [] - name = src_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:35,984 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:35,984 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 128 -2016-08-30 09:55:35,984 INFO query.QuerySchemaLoader [] - filterName = qname -2016-08-30 09:55:35,999 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml -2016-08-30 09:55:35,999 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query_file6934837797661139339.xml -2016-08-30 09:55:36,000 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:36,000 INFO query.QuerySchemaLoader [] - schemaName = dns-srcip-query -2016-08-30 09:55:36,001 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:36,001 INFO query.QuerySchemaLoader [] - selectorName = src_ip -2016-08-30 09:55:36,001 INFO query.QuerySchemaLoader [] - name = qname partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:36,001 INFO query.QuerySchemaLoader [] - name = qname bits = 64 dataElementSize = 64 -2016-08-30 09:55:36,001 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:36,001 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:36,001 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:36,002 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 160 -2016-08-30 09:55:36,002 INFO query.QuerySchemaLoader [] - filterName = src_ip -2016-08-30 09:55:36,002 INFO query.QuerySchemaLoader [] - filterName = ip -2016-08-30 09:55:36,016 INFO query.QuerySchemaLoader [] - Loading schemaFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml -2016-08-30 09:55:36,017 INFO query.QuerySchemaLoader [] - localFS: inputFile = /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/dns-srcip-query-no-filter_file4280566018573806206.xml -2016-08-30 09:55:36,018 INFO query.QuerySchemaLoader [] - Root element: schema -2016-08-30 09:55:36,018 INFO query.QuerySchemaLoader [] - schemaName = dns-srcip-query-no-filter -2016-08-30 09:55:36,018 INFO query.QuerySchemaLoader [] - dataSchemaName = testDataSchema -2016-08-30 09:55:36,018 INFO query.QuerySchemaLoader [] - selectorName = src_ip -2016-08-30 09:55:36,019 INFO query.QuerySchemaLoader [] - name = qname partitionerName = org.apache.pirk.schema.data.partitioner.PrimitiveTypePartitioner -2016-08-30 09:55:36,019 INFO query.QuerySchemaLoader [] - name = qname bits = 64 dataElementSize = 64 -2016-08-30 09:55:36,019 INFO query.QuerySchemaLoader [] - name = dest_ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:36,019 INFO query.QuerySchemaLoader [] - name = dest_ip bits = 32 dataElementSize = 96 -2016-08-30 09:55:36,019 INFO query.QuerySchemaLoader [] - name = ip partitionerName = org.apache.pirk.schema.data.partitioner.IPDataPartitioner -2016-08-30 09:55:36,019 INFO query.QuerySchemaLoader [] - name = ip bits = 64 dataElementSize = 160 -2016-08-30 09:55:38,022 INFO server.PrepRequestProcessor [] - Got user-level KeeperException when processing sessionid:0x156dbbbca4d0010 type:create cxid:0x3 zxid:0x3b txntype:-1 reqpath:n/a Error Path:/pirk_test_root/pirk_integr_test_spout Error:KeeperErrorCode = NoNode for /pirk_test_root/pirk_integr_test_spout -2016-08-30 09:55:42,895 INFO storm.OutputBolt [] - TimeToFlush reached - outputting response to /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T/responderFile5665813788069927912.txt with columns.size = 180 -2016-08-30 09:55:42,951 INFO storm.KafkaStormIntegrationTest [] - Finished... -2016-08-30 09:55:42,951 INFO daemon.nimbus [] - Shutting down master -2016-08-30 09:55:42,952 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,952 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0003 -2016-08-30 09:55:42,953 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d0003 closed -2016-08-30 09:55:42,953 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,953 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56305 which had sessionid 0x156dbbbca4d0003 -2016-08-30 09:55:42,953 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,954 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0004 -2016-08-30 09:55:42,954 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d0004 closed -2016-08-30 09:55:42,954 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,954 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56306 which had sessionid 0x156dbbbca4d0004 -2016-08-30 09:55:42,954 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,955 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0001 -2016-08-30 09:55:42,955 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d0001 closed -2016-08-30 09:55:42,955 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,955 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56303 which had sessionid 0x156dbbbca4d0001 -2016-08-30 09:55:42,956 INFO storm.zookeeper [] - closing zookeeper connection of leader elector. -2016-08-30 09:55:42,956 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,957 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0000 -2016-08-30 09:55:42,958 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d0000 closed -2016-08-30 09:55:42,958 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56302 which had sessionid 0x156dbbbca4d0000 -2016-08-30 09:55:42,958 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,958 INFO daemon.nimbus [] - Shut down master -2016-08-30 09:55:42,959 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,960 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0006 -2016-08-30 09:55:42,960 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d0006 closed -2016-08-30 09:55:42,960 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56308 which had sessionid 0x156dbbbca4d0006 -2016-08-30 09:55:42,960 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,961 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,961 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0008 -2016-08-30 09:55:42,962 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d0008 closed -2016-08-30 09:55:42,962 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,962 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56310 which had sessionid 0x156dbbbca4d0008 -2016-08-30 09:55:42,962 INFO daemon.supervisor [] - Shutting down 59a01bf6-2a07-4e98-a935-3031a25e93b7:cacfc67b-c838-449e-8e35-dca6d3f1da94 -2016-08-30 09:55:42,963 INFO storm.config [] - GET worker-user cacfc67b-c838-449e-8e35-dca6d3f1da94 -2016-08-30 09:55:42,963 INFO storm.process-simulator [] - Killing process 3e97964a-19b4-43ae-9718-6834ddb4a775 -2016-08-30 09:55:42,964 INFO daemon.worker [] - Shutting down worker pirk_integration_test-1-1472565332 59a01bf6-2a07-4e98-a935-3031a25e93b7 1024 -2016-08-30 09:55:42,964 INFO daemon.worker [] - Terminating messaging context -2016-08-30 09:55:42,964 INFO daemon.worker [] - Shutting down executors -2016-08-30 09:55:42,964 INFO daemon.executor [] - Shutting down executor partitiondataBolt:[8 8] -2016-08-30 09:55:42,965 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,965 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,965 INFO daemon.executor [] - Shut down executor partitiondataBolt:[8 8] -2016-08-30 09:55:42,966 INFO daemon.executor [] - Shutting down executor enccolmultbolt:[2 2] -2016-08-30 09:55:42,966 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,966 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,966 INFO daemon.executor [] - Shut down executor enccolmultbolt:[2 2] -2016-08-30 09:55:42,966 INFO daemon.executor [] - Shutting down executor outputbolt:[7 7] -2016-08-30 09:55:42,967 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,967 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,967 INFO daemon.executor [] - Shut down executor outputbolt:[7 7] -2016-08-30 09:55:42,967 INFO daemon.executor [] - Shutting down executor enccolmultbolt:[3 3] -2016-08-30 09:55:42,967 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,967 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,968 INFO daemon.executor [] - Shut down executor enccolmultbolt:[3 3] -2016-08-30 09:55:42,968 INFO daemon.executor [] - Shutting down executor __acker:[1 1] -2016-08-30 09:55:42,968 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,968 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,969 INFO daemon.executor [] - Shut down executor __acker:[1 1] -2016-08-30 09:55:42,969 INFO daemon.executor [] - Shutting down executor kafkaspout:[6 6] -2016-08-30 09:55:42,969 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,970 WARN kafka.KafkaUtils [] - Network error when fetching messages: -java.nio.channels.ClosedByInterruptException - at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202) ~[?:1.8.0_102] - at sun.nio.ch.SocketChannelImpl.poll(SocketChannelImpl.java:957) ~[?:1.8.0_102] - at sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:204) ~[?:1.8.0_102] - at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103) ~[?:1.8.0_102] - at java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:385) ~[?:1.8.0_102] - at org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:81) ~[kafka-clients-0.9.0.1.jar:?] - at kafka.network.BlockingChannel.readCompletely(BlockingChannel.scala:129) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.network.BlockingChannel.receive(BlockingChannel.scala:120) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer.scala:86) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$$sendRequest(SimpleConsumer.scala:83) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(SimpleConsumer.scala:132) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply$mcV$sp(SimpleConsumer.scala:131) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:130) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.javaapi.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:47) ~[kafka_2.10-0.9.0.1.jar:?] - at org.apache.storm.kafka.KafkaUtils.fetchMessages(KafkaUtils.java:191) [storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.kafka.PartitionManager.fill(PartitionManager.java:189) [storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.kafka.PartitionManager.next(PartitionManager.java:138) [storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.kafka.KafkaSpout.nextTuple(KafkaSpout.java:135) [storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.daemon.executor$fn__7885$fn__7900$fn__7931.invoke(executor.clj:645) [storm-core-1.0.1.jar:1.0.1] - at org.apache.storm.util$async_loop$fn__625.invoke(util.clj:484) [storm-core-1.0.1.jar:1.0.1] - at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?] - at java.lang.Thread.run(Thread.java:745) [?:1.8.0_102] -2016-08-30 09:55:42,977 WARN kafka.KafkaSpout [] - Fetch failed -org.apache.storm.kafka.FailedFetchException: java.nio.channels.ClosedByInterruptException - at org.apache.storm.kafka.KafkaUtils.fetchMessages(KafkaUtils.java:199) ~[storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.kafka.PartitionManager.fill(PartitionManager.java:189) ~[storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.kafka.PartitionManager.next(PartitionManager.java:138) ~[storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.kafka.KafkaSpout.nextTuple(KafkaSpout.java:135) [storm-kafka-1.0.1.jar:1.0.1] - at org.apache.storm.daemon.executor$fn__7885$fn__7900$fn__7931.invoke(executor.clj:645) [storm-core-1.0.1.jar:1.0.1] - at org.apache.storm.util$async_loop$fn__625.invoke(util.clj:484) [storm-core-1.0.1.jar:1.0.1] - at clojure.lang.AFn.run(AFn.java:22) [clojure-1.7.0.jar:?] - at java.lang.Thread.run(Thread.java:745) [?:1.8.0_102] -Caused by: java.nio.channels.ClosedByInterruptException - at java.nio.channels.spi.AbstractInterruptibleChannel.end(AbstractInterruptibleChannel.java:202) ~[?:1.8.0_102] - at sun.nio.ch.SocketChannelImpl.poll(SocketChannelImpl.java:957) ~[?:1.8.0_102] - at sun.nio.ch.SocketAdaptor$SocketInputStream.read(SocketAdaptor.java:204) ~[?:1.8.0_102] - at sun.nio.ch.ChannelInputStream.read(ChannelInputStream.java:103) ~[?:1.8.0_102] - at java.nio.channels.Channels$ReadableByteChannelImpl.read(Channels.java:385) ~[?:1.8.0_102] - at org.apache.kafka.common.network.NetworkReceive.readFromReadableChannel(NetworkReceive.java:81) ~[kafka-clients-0.9.0.1.jar:?] - at kafka.network.BlockingChannel.readCompletely(BlockingChannel.scala:129) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.network.BlockingChannel.receive(BlockingChannel.scala:120) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer.liftedTree1$1(SimpleConsumer.scala:86) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer.kafka$consumer$SimpleConsumer$$sendRequest(SimpleConsumer.scala:83) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply$mcV$sp(SimpleConsumer.scala:132) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1$$anonfun$apply$mcV$sp$1.apply(SimpleConsumer.scala:132) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply$mcV$sp(SimpleConsumer.scala:131) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer$$anonfun$fetch$1.apply(SimpleConsumer.scala:131) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.metrics.KafkaTimer.time(KafkaTimer.scala:33) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:130) ~[kafka_2.10-0.9.0.1.jar:?] - at kafka.javaapi.consumer.SimpleConsumer.fetch(SimpleConsumer.scala:47) ~[kafka_2.10-0.9.0.1.jar:?] - at org.apache.storm.kafka.KafkaUtils.fetchMessages(KafkaUtils.java:191) ~[storm-kafka-1.0.1.jar:1.0.1] - ... 7 more -2016-08-30 09:55:42,977 INFO kafka.ZkCoordinator [] - Task [1/1] Refreshing partition manager connections -2016-08-30 09:55:42,978 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,978 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,979 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d0010 -2016-08-30 09:55:42,980 INFO daemon.executor [] - Shut down executor kafkaspout:[6 6] -2016-08-30 09:55:42,980 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56318 which had sessionid 0x156dbbbca4d0010 -2016-08-30 09:55:42,980 INFO daemon.executor [] - Shutting down executor __system:[-1 -1] -2016-08-30 09:55:42,981 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,981 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,981 INFO daemon.executor [] - Shut down executor __system:[-1 -1] -2016-08-30 09:55:42,981 INFO daemon.executor [] - Shutting down executor encrowcalcbolt:[5 5] -2016-08-30 09:55:42,982 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,982 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,982 INFO daemon.executor [] - Shut down executor encrowcalcbolt:[5 5] -2016-08-30 09:55:42,982 INFO daemon.executor [] - Shutting down executor encrowcalcbolt:[4 4] -2016-08-30 09:55:42,982 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,982 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,983 INFO daemon.executor [] - Shut down executor encrowcalcbolt:[4 4] -2016-08-30 09:55:42,983 INFO daemon.worker [] - Shut down executors -2016-08-30 09:55:42,983 INFO daemon.worker [] - Shutting down transfer thread -2016-08-30 09:55:42,984 INFO storm.util [] - Async loop interrupted! -2016-08-30 09:55:42,984 INFO daemon.worker [] - Shut down transfer thread -2016-08-30 09:55:42,984 INFO daemon.worker [] - Shut down backpressure thread -2016-08-30 09:55:42,986 INFO daemon.worker [] - Shutting down default resources -2016-08-30 09:55:42,986 INFO daemon.worker [] - Shut down default resources -2016-08-30 09:55:42,987 INFO daemon.worker [] - Trigger any worker shutdown hooks -2016-08-30 09:55:42,988 INFO daemon.worker [] - Disconnecting from storm cluster state context -2016-08-30 09:55:42,988 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,989 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d000f -2016-08-30 09:55:42,989 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d000f closed -2016-08-30 09:55:42,989 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,989 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56317 which had sessionid 0x156dbbbca4d000f -2016-08-30 09:55:42,990 INFO daemon.worker [] - Shut down worker pirk_integration_test-1-1472565332 59a01bf6-2a07-4e98-a935-3031a25e93b7 1024 -2016-08-30 09:55:42,993 INFO storm.config [] - REMOVE worker-user cacfc67b-c838-449e-8e35-dca6d3f1da94 -2016-08-30 09:55:42,993 INFO daemon.supervisor [] - Shut down 59a01bf6-2a07-4e98-a935-3031a25e93b7:cacfc67b-c838-449e-8e35-dca6d3f1da94 -2016-08-30 09:55:42,993 INFO daemon.supervisor [] - Shutting down supervisor 59a01bf6-2a07-4e98-a935-3031a25e93b7 -2016-08-30 09:55:42,994 INFO storm.event [] - Event manager interrupted -2016-08-30 09:55:42,994 INFO storm.event [] - Event manager interrupted -2016-08-30 09:55:42,994 INFO imps.CuratorFrameworkImpl [] - backgroundOperationsLoop exiting -2016-08-30 09:55:42,995 INFO server.PrepRequestProcessor [] - Processed session termination for sessionid: 0x156dbbbca4d000a -2016-08-30 09:55:42,995 INFO zookeeper.ZooKeeper [] - Session: 0x156dbbbca4d000a closed -2016-08-30 09:55:42,996 INFO zookeeper.ClientCnxn [] - EventThread shut down -2016-08-30 09:55:42,996 INFO server.NIOServerCnxn [] - Closed socket connection for client /127.0.0.1:56312 which had sessionid 0x156dbbbca4d000a -2016-08-30 09:55:42,996 INFO storm.testing [] - Shutting down in process zookeeper -2016-08-30 09:55:42,996 INFO server.NIOServerCnxnFactory [] - NIOServerCnxn factory exited run method -2016-08-30 09:55:42,997 INFO server.ZooKeeperServer [] - shutting down -2016-08-30 09:55:42,997 INFO server.SessionTrackerImpl [] - Shutting down -2016-08-30 09:55:42,997 INFO server.PrepRequestProcessor [] - Shutting down -2016-08-30 09:55:42,997 INFO server.SyncRequestProcessor [] - Shutting down -2016-08-30 09:55:42,997 INFO server.PrepRequestProcessor [] - PrepRequestProcessor exited loop! -2016-08-30 09:55:42,997 INFO server.SyncRequestProcessor [] - SyncRequestProcessor exited! -2016-08-30 09:55:42,997 INFO server.FinalRequestProcessor [] - shutdown of request processor complete -2016-08-30 09:55:42,998 INFO storm.testing [] - Done shutting down in process zookeeper -2016-08-30 09:55:42,998 INFO storm.testing [] - Deleting temporary path /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//f9c8fb23-e6d6-415d-b29b-2ecc9cfd9b60 -2016-08-30 09:55:43,001 INFO storm.testing [] - Deleting temporary path /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//e665f001-f902-4bd0-89ed-30254a84f1ef -2016-08-30 09:55:43,002 INFO storm.testing [] - Deleting temporary path /var/folders/83/07wq_8t55pn3mgj99512hb8c0000gn/T//960b6925-fc51-4264-984e-d7937568606f -2016-08-30 09:55:43,051 INFO storm.KafkaStormIntegrationTest [] - Decrypting results. 14716 -2016-08-30 09:55:43,218 INFO utils.BaseTests [] - results: -2016-08-30 09:55:43,218 INFO utils.BaseTests [] - {"date":"2016-02-20T23:29:05.000Z","src_ip":"55.55.55.55","event_type":"dns-hostname-query","query_id":"74cbee54-9a78-48c3-905d-c83a0782b051","qtype":[1,0],"dest_ip":"1.2.3.6","ip":["10.20.30.40","10.20.30.60"],"qname":"a.b.c.com","match":"a.b.c.com","rcode":0} -2016-08-30 09:55:43,218 INFO utils.BaseTests [] - {"date":"2016-02-20T23:29:11.000Z","src_ip":"1.1.1.1","event_type":"dns-hostname-query","query_id":"74cbee54-9a78-48c3-905d-c83a0782b051","qtype":[1,0],"dest_ip":"2.2.2.2","ip":["3.3.3.3","0.0.0.0"],"qname":"something.else","match":"something.else","rcode":0} -2016-08-30 09:55:43,218 INFO utils.BaseTests [] - {"date":"2016-02-20T23:29:06.000Z","src_ip":"127.128.129.130","event_
<TRUNCATED>
