Thank you very much, Kosala I have get it run on my production, it works good at first try, it get data from kafkaSpout, and write into postgresq DB, I count the number it is the number we need, but when I make second run, it comes with such error:
java.lang.RuntimeException: java.lang.RuntimeException: org.apache.storm.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /partition_1/126188 at backtype.storm.utils.DisruptorQueue.consumeBatchToCursor(DisruptorQueue.java:128) at backtype.storm.utils.DisruptorQueue.consumeBatchWhenAvailable(DisruptorQueue.java:99) at backtype.storm.disruptor$consume_batch_when_available.invoke(disruptor.clj:80) at backtype.storm.daemon.executor$fn__3441$fn__3453$fn__3500.invoke(executor.clj:748) at backtype.storm.util$async_loop$fn__464.invoke(util.clj:463) at clojure.lang.AFn.run(AFn.java:24) at java.lang.Thread.run(Thread.java:745) Caused by: java.lang.RuntimeException: org.apache.storm.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /partition_1/126188 at storm.trident.topology.state.TransactionalState.delete(TransactionalState.java:92) at storm.trident.topology.state.RotatingTransactionalState.removeState(RotatingTransactionalState.java:59) at storm.trident.spout.OpaquePartitionedTridentSpoutExecutor$Emitter.emitBatch(OpaquePartitionedTridentSpoutExecutor.java:124) at storm.trident.spout.TridentSpoutExecutor.execute(TridentSpoutExecutor.java:82) at storm.trident.topology.TridentBoltExecutor.execute(TridentBoltExecutor.java:369) at backtype.storm.daemon.executor$fn__3441$tuple_action_fn__3443.invoke(executor.clj:633) at backtype.storm.daemon.executor$mk_task_receiver$fn__3364.invoke(executor.clj:401) at backtype.storm.disruptor$clojure_handler$reify__1447.onEvent(disruptor.clj:58) at backtype.storm.utils.DisruptorQueue.consumeBatchToCursor(DisruptorQueue.java:125) ... 6 more Caused by: org.apache.storm.zookeeper.KeeperException$NoNodeException: KeeperErrorCode = NoNode for /partition_1/126188 at org.apache.storm.zookeeper.KeeperException.create(KeeperException.java:111) at org.apache.storm.zookeeper.KeeperException.create(KeeperException.java:51) at org.apache.storm.zookeeper.ZooKeeper.delete(ZooKeeper.java:873) at org.apache.storm.curator.framework.imps.DeleteBuilderImpl$5.call(DeleteBuilderImpl.java:239) at org.apache.storm.curator.framework.imps.DeleteBuilderImpl$5.call(DeleteBuilderImpl.java:234) at org.apache.storm.curator.RetryLoop.callWithRetry(RetryLoop.java:107) at org.apache.storm.curator.framework.imps.DeleteBuilderImpl.pathInForeground(DeleteBuilderImpl.java:230) at org.apache.storm.curator.framework.imps.DeleteBuilderImpl.forPath(DeleteBuilderImpl.java:215) at org.apache.storm.curator.framework.imps.DeleteBuilderImpl.forPath(DeleteBuilderImpl.java:42) at storm.trident.topology.state.TransactionalState.delete(TransactionalState.java:90) ... 14 more Do I need to manually create zNode in zK server? how to do that? thanks AL On Wed, Feb 11, 2015 at 5:49 PM, Kosala Dissanayake <[email protected]> wrote: > Run the command following the words 'Launching worker with command: > > 'java' '-server' '-Xmx768m' '-Djava.net.preferIPv4Stack=true' > '-Djava.library.path=/srv/tmpvar/storm/data/supervisor/ > stormdist/KafkaIngresBasic-5-1423692389/resources/Linux- > amd64:/srv/tmpvar/storm/data/supervisor/stormdist/ > KafkaIngresBasic-5-1423692389/resources:/usr/lib/jvm/java-7-openjdk-amd64' > '-Dlogfile.name=worker-6703.log' '-Dstorm.home=/srv/storm/storm' > '-Dstorm.conf.file=' '-Dstorm.options=' > '-Dstorm.log.dir=/srv/storm/storm/logs' > '-Dlogback.configurationFile=/srv/storm/storm/logback/cluster.xml' > '-Dstorm.id=KafkaIngresBasic-5-1423692389' > '-Dworker.id=0b3efe86-751f-449f-b331-25a530e85101' > '-Dworker.port=6703' '-cp' '/srv/storm/storm/lib/jgrapht- > core-0.9.0.jar:/srv/storm/storm/lib/clj-stacktrace-0.2. > 2.jar:/srv/storm/storm/lib/disruptor-2.10.1.jar:/srv/ > storm/storm/lib/math.numeric-tower-0.0.1.jar:/srv/storm/ > storm/lib/minlog-1.2.jar:/srv/storm/storm/lib/jline-2.11. > jar:/srv/storm/storm/lib/ring-servlet-0.3.11.jar:/srv/storm/ > storm/lib/clojure-1.5.1.jar:/srv/storm/storm/lib/ring- > jetty-adapter-0.3.11.jar:/srv/storm/storm/lib/jetty-6.1.26. > jar:/srv/storm/storm/lib/clj-time-0.4.1.jar:/srv/storm/ > storm/lib/jetty-util-6.1.26.jar:/srv/storm/storm/lib/ > servlet-api-2.5.jar:/srv/storm/storm/lib/commons-exec- > 1.1.jar:/srv/storm/storm/lib/core.incubator-0.1.0.jar:/srv/ > storm/storm/lib/clout-1.0.1.jar:/srv/storm/storm/lib/ > snakeyaml-1.11.jar:/srv/storm/storm/lib/storm-core-0.9.3. > jar:/srv/storm/storm/lib/slf4j-api-1.7.5.jar:/srv/ > storm/storm/lib/tools.cli-0.2.4.jar:/srv/storm/storm/lib/ > joda-time-2.0.jar:/srv/storm/storm/lib/logback-classic-1.0. > 13.jar:/srv/storm/storm/lib/kryo-2.21.jar:/srv/storm/ > storm/lib/tools.logging-0.2.3.jar:/srv/storm/storm/lib/ > objenesis-1.2.jar:/srv/storm/storm/lib/commons-codec-1.6. > jar:/srv/storm/storm/lib/logback-core-1.0.13.jar:/srv/ > storm/storm/lib/ring-core-1.1.5.jar:/srv/storm/storm/lib/ > json-simple-1.1.jar:/srv/storm/storm/lib/carbonite-1.4. > 0.jar:/srv/storm/storm/lib/chill-java-0.3.5.jar:/srv/ > storm/storm/lib/log4j-over-slf4j-1.6.6.jar:/srv/storm/ > storm/lib/commons-fileupload-1.2.1.jar:/srv/storm/storm/ > lib/hiccup-0.3.6.jar:/srv/storm/storm/lib/ring-devel-0. > 3.11.jar:/srv/storm/storm/lib/commons-logging-1.1.3.jar:/ > srv/storm/storm/lib/tools.macro-0.1.0.jar:/srv/storm/ > storm/lib/asm-4.0.jar:/srv/storm/storm/lib/commons-io-2. > 4.jar:/srv/storm/storm/lib/compojure-1.1.3.jar:/srv/ > storm/storm/lib/commons-lang-2.5.jar:/srv/storm/storm/lib/ > reflectasm-1.07-shaded.jar:/srv/storm/storm/conf:/srv/ > tmpvar/storm/data/supervisor/stormdist/KafkaIngresBasic-5-1423692389/stormjar.jar' > 'backtype.storm.daemon.worker' 'KafkaIngresBasic-5-1423692389' > 'd2ff3ed7-2b84-45b7-99cc-63d859944591' '6703' '0b3efe86-751f-449f-b331- > 25a530e85101' > > > separately, manually, and see if you get any error messages. > > On Thu, Feb 12, 2015 at 9:39 AM, Sa Li <[email protected]> wrote: > >> Hi, Kosala >> >> Thank you for the reply, I have reconfigure the hostnames in above >> machines which are in my DEV cluster. Now I am able to run my topology in >> dev storm cluster with no problem. Now I am moving my code to production, >> problem coming again, it has no problem running on the localmode, but >> showing the error after submitting to storm cluster, see the attached UI >> screenshot. This the hosts file for each node >> >> 127.0.0.1 localhost >> 127.0.1.1 complicated-laugh complicated-laugh.master >> >> 10.100.98.100 exemplary-birds >> 10.100.98.101 voluminous-mass >> 10.100.98.102 harmful-jar >> >> 10.100.98.103 complicated-laugh >> 10.100.98.104 beloved-judge >> 10.100.98.105 visible-alley >> 10.100.98.106 aromatic-reward >> >> When I check the storm logs, it shows such errors >> >> logs in supervisor nodes: >> 2015-02-11T22:36:12.270+0000 b.s.d.supervisor [INFO] >> 197cc48d-8db6-45ed-bc05-2cc81351538f still hasn't started >> 2015-02-11T22:36:12.771+0000 b.s.d.supervisor [INFO] >> 197cc48d-8db6-45ed-bc05-2cc81351538f still hasn't started >> 2015-02-11T22:36:13.273+0000 b.s.d.supervisor [INFO] Shutting down and >> clearing state for id 197cc48d-8db6-45ed-bc05-2cc81351538f. Current >> supervisor time: 1423694173. State: :disallowed, Heartbeat: >> #backtype.storm.daemon.common.WorkerHeartbeat{:time-secs 1423694173, >> :storm-id "KafkaIngresBasic-5-1423692389", :executors #{[-1 -1]}, :port >> 6702} >> 2015-02-11T22:36:13.273+0000 b.s.d.supervisor [INFO] Shutting down >> d2ff3ed7-2b84-45b7-99cc-63d859944591:197cc48d-8db6-45ed-bc05-2cc81351538f >> 2015-02-11T22:36:14.276+0000 b.s.util [INFO] Error when trying to kill >> 5564. Process is probably already dead. >> 2015-02-11T22:36:14.276+0000 b.s.d.supervisor [INFO] Shut down >> d2ff3ed7-2b84-45b7-99cc-63d859944591:197cc48d-8db6-45ed-bc05-2cc81351538f >> 2015-02-11T22:36:14.277+0000 b.s.d.supervisor [INFO] Shutting down and >> clearing state for id b5237503-ab27-48af-a0f6-63d2e71da71a. Current >> supervisor time: 1423694173. State: :timed-out, Heartbeat: >> #backtype.storm.daemon.common.WorkerHeartbeat{:time-secs 1423694140, >> :storm-id "KafkaIngresBasic-5-1423692389", :executors #{[6 6] [14 14] [23 >> 23] [-1 -1]}, :port 6701} >> 2015-02-11T22:36:14.277+0000 b.s.d.supervisor [INFO] Shutting down >> d2ff3ed7-2b84-45b7-99cc-63d859944591:b5237503-ab27-48af-a0f6-63d2e71da71a >> 2015-02-11T22:36:14.278+0000 b.s.util [INFO] Error when trying to kill >> 5436. Process is probably already dead. >> 2015-02-11T22:36:15.280+0000 b.s.util [INFO] Error when trying to kill >> 5436. Process is probably already dead. >> 2015-02-11T22:36:15.280+0000 b.s.d.supervisor [INFO] Shut down >> d2ff3ed7-2b84-45b7-99cc-63d859944591:b5237503-ab27-48af-a0f6-63d2e71da71a >> 2015-02-11T22:36:15.281+0000 b.s.d.supervisor [INFO] Launching worker >> with assignment #backtype.storm.daemon.supervisor.LocalAssignment{:storm-id >> "KafkaIngresBasic-5-1423692389", :executors ([7 7] [16 16] [25 25])} for >> this supervisor d2ff3ed7-2b84-45b7-99cc-63d859944591 on port 6703 with id >> 0b3efe86-751f-449f-b331-25a530e85101 >> 2015-02-11T22:36:15.282+0000 b.s.d.supervisor [INFO] Launching worker >> with command: 'java' '-server' '-Xmx768m' '-Djava.net.preferIPv4Stack=true' >> '-Djava.library.path=/srv/tmpvar/storm/data/supervisor/stormdist/KafkaIngresBasic-5-1423692389/resources/Linux-amd64:/srv/tmpvar/storm/data/supervisor/stormdist/KafkaIngresBasic-5-1423692389/resources:/usr/lib/jvm/java-7-openjdk-amd64' >> '-Dlogfile.name=worker-6703.log' '-Dstorm.home=/srv/storm/storm' >> '-Dstorm.conf.file=' '-Dstorm.options=' >> '-Dstorm.log.dir=/srv/storm/storm/logs' >> '-Dlogback.configurationFile=/srv/storm/storm/logback/cluster.xml' >> '-Dstorm.id=KafkaIngresBasic-5-1423692389' >> '-Dworker.id=0b3efe86-751f-449f-b331-25a530e85101' '-Dworker.port=6703' >> '-cp' >> '/srv/storm/storm/lib/jgrapht-core-0.9.0.jar:/srv/storm/storm/lib/clj-stacktrace-0.2.2.jar:/srv/storm/storm/lib/disruptor-2.10.1.jar:/srv/storm/storm/lib/math.numeric-tower-0.0.1.jar:/srv/storm/storm/lib/minlog-1.2.jar:/srv/storm/storm/lib/jline-2.11.jar:/srv/storm/storm/lib/ring-servlet-0.3.11.jar:/srv/storm/storm/lib/clojure-1.5.1.jar:/srv/storm/storm/lib/ring-jetty-adapter-0.3.11.jar:/srv/storm/storm/lib/jetty-6.1.26.jar:/srv/storm/storm/lib/clj-time-0.4.1.jar:/srv/storm/storm/lib/jetty-util-6.1.26.jar:/srv/storm/storm/lib/servlet-api-2.5.jar:/srv/storm/storm/lib/commons-exec-1.1.jar:/srv/storm/storm/lib/core.incubator-0.1.0.jar:/srv/storm/storm/lib/clout-1.0.1.jar:/srv/storm/storm/lib/snakeyaml-1.11.jar:/srv/storm/storm/lib/storm-core-0.9.3.jar:/srv/storm/storm/lib/slf4j-api-1.7.5.jar:/srv/storm/storm/lib/tools.cli-0.2.4.jar:/srv/storm/storm/lib/joda-time-2.0.jar:/srv/storm/storm/lib/logback-classic-1.0.13.jar:/srv/storm/storm/lib/kryo-2.21.jar:/srv/storm/storm/lib/tools.logging-0.2.3.jar:/srv/storm/storm/lib/objenesis-1.2.jar:/srv/storm/storm/lib/commons-codec-1.6.jar:/srv/storm/storm/lib/logback-core-1.0.13.jar:/srv/storm/storm/lib/ring-core-1.1.5.jar:/srv/storm/storm/lib/json-simple-1.1.jar:/srv/storm/storm/lib/carbonite-1.4.0.jar:/srv/storm/storm/lib/chill-java-0.3.5.jar:/srv/storm/storm/lib/log4j-over-slf4j-1.6.6.jar:/srv/storm/storm/lib/commons-fileupload-1.2.1.jar:/srv/storm/storm/lib/hiccup-0.3.6.jar:/srv/storm/storm/lib/ring-devel-0.3.11.jar:/srv/storm/storm/lib/commons-logging-1.1.3.jar:/srv/storm/storm/lib/tools.macro-0.1.0.jar:/srv/storm/storm/lib/asm-4.0.jar:/srv/storm/storm/lib/commons-io-2.4.jar:/srv/storm/storm/lib/compojure-1.1.3.jar:/srv/storm/storm/lib/commons-lang-2.5.jar:/srv/storm/storm/lib/reflectasm-1.07-shaded.jar:/srv/storm/storm/conf:/srv/tmpvar/storm/data/supervisor/stormdist/KafkaIngresBasic-5-1423692389/stormjar.jar' >> 'backtype.storm.daemon.worker' 'KafkaIngresBasic-5-1423692389' >> 'd2ff3ed7-2b84-45b7-99cc-63d859944591' '6703' >> '0b3efe86-751f-449f-b331-25a530e85101' >> 2015-02-11T22:36:15.283+0000 b.s.d.supervisor [INFO] Launching worker >> with assignment #backtype.storm.daemon.supervisor.LocalAssignment{:storm-id >> "KafkaIngresBasic-5-1423692389", :executors ([6 6] [14 14] [23 23])} for >> this supervisor d2ff3ed7-2b84-45b7-99cc-63d859944591 on port 6701 with id >> 4c2284dc-b8b0-4ce8-86c1-26154c6f091e >> >> I am not sure if this is still the issue of hosts file. >> >> thanks >> >> AL >> >> On Tue, Feb 10, 2015 at 4:16 PM, Kosala Dissanayake <[email protected] >> > wrote: >> >>> Seems like a name resolution issue. Have you configured the IP addresses >>> for your supervisor machines in /etc/hosts? >>> >>> On Wed, Feb 11, 2015 at 5:36 AM, Sa Li <[email protected]> wrote: >>> >>>> I did some changes, now, I don't see any errors on Storm UI, but it >>>> won't work as it works in local mode, like writing things in DB, so I tail >>>> the logs again, it is still: >>>> >>>> 2015-02-10T10:34:36.989-0800 b.s.m.n.Client [INFO] Reconnect started >>>> for Netty-Client-pof-kstorm-dev2.pof.local:6702... [300] >>>> 2015-02-10T10:34:36.989-0800 b.s.m.n.StormClientErrorHandler [INFO] >>>> Connection failed Netty-Client-pof-kstorm-dev2.pof.local:6702 >>>> java.nio.channels.UnresolvedAddressException: null >>>> at sun.nio.ch.Net.checkAddress(Net.java:127) ~[na:1.7.0_72] >>>> at >>>> sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:644) >>>> ~[na:1.7.0_72] >>>> at >>>> org.apache.storm.netty.channel.socket.nio.NioClientSocketPipelineSink.connect(NioClientSocketPipelineSink.java:108) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.channel.socket.nio.NioClientSocketPipelineSink.eventSunk(NioClientSocketPipelineSink.java:70) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendDownstream(DefaultChannelPipeline.java:779) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.handler.codec.oneone.OneToOneEncoder.handleDownstream(OneToOneEncoder.java:54) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.channel.DefaultChannelPipeline.sendDownstream(DefaultChannelPipeline.java:591) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.channel.DefaultChannelPipeline.sendDownstream(DefaultChannelPipeline.java:582) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.channel.Channels.connect(Channels.java:634) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.channel.AbstractChannel.connect(AbstractChannel.java:207) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.bootstrap.ClientBootstrap.connect(ClientBootstrap.java:229) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> org.apache.storm.netty.bootstrap.ClientBootstrap.connect(ClientBootstrap.java:182) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> backtype.storm.messaging.netty.Client.connect(Client.java:152) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> backtype.storm.messaging.netty.Client.access$000(Client.java:43) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at backtype.storm.messaging.netty.Client$1.run(Client.java:107) >>>> [storm-core-0.9.3.jar:0.9.3] >>>> at >>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>>> [na:1.7.0_72] >>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>> [na:1.7.0_72] >>>> at >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178) >>>> [na:1.7.0_72] >>>> at >>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292) >>>> [na:1.7.0_72] >>>> at >>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>> [na:1.7.0_72] >>>> at >>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>> [na:1.7.0_72] >>>> at java.lang.Thread.run(Thread.java:745) [na:1.7.0_72] >>>> 2015-02-10T10:34:37.827-0800 b.s.m.n.Client [INFO] Closing Netty Client >>>> Netty-Client-pof-kstorm-dev2.pof.local:6701 >>>> 2015-02-10T10:34:37.827-0800 b.s.m.n.Client [INFO] Waiting for pending >>>> batchs to be sent with Netty-Client-pof-kstorm-dev2.pof.local:6701..., >>>> timeout: 600000ms, pendings: 0 >>>> 2015-02-10T10:34:37.828-0800 b.s.m.n.Client [INFO] Closing Netty Client >>>> Netty-Client-pof-kstorm-dev2.pof.local:6703 >>>> 2015-02-10T10:34:37.829-0800 b.s.m.n.Client [INFO] Waiting for pending >>>> batchs to be sent with Netty-Client-pof-kstorm-dev2.pof.local:6703..., >>>> timeout: 600000ms, pendings: 0 >>>> 2015-02-10T10:34:37.931-0800 b.s.m.n.Client [INFO] Closing Netty Client >>>> Netty-Client-pof-kstorm-dev2.pof.local:6702 >>>> 2015-02-10T10:34:37.931-0800 b.s.m.n.Client [INFO] Waiting for pending >>>> batchs to be sent with Netty-Client-pof-kstorm-dev2.pof.local:6702..., >>>> timeout: 600000ms, pendings: 0 >>>> >>>> >>>> Any idea how to fix this? seems it has connection issues to workers. >>>> >>>> thanks >>>> >>>> AL >>>> >>>> On Fri, Feb 6, 2015 at 11:09 AM, Sa Li <[email protected]> wrote: >>>> >>>>> Hi, All >>>>> >>>>> I have tested my topologies on local mode, it works fine. Now I like >>>>> to move forward to submit the topologies to storm cluster, here are the >>>>> problems on storm UI >>>>> >>>>> >>>>> $mastercoord-bg0 >>>>> <http://10.100.71.33:8080/component.html?id=%24mastercoord-bg0&topology_id=kstib001-2-1423182631> >>>>> 1 1 0 0 0.000 0 0 pof-kstorm-dev1.pof.local 6702 >>>>> <http://pof-kstorm-dev1.pof.local:8000/log?file=worker-6702.log> >>>>> java.lang.RuntimeException: >>>>> java.lang.NullPointerException at >>>>> storm.trident.topology.state.TransactionalState.<init>(TransactionalState.java:61) >>>>> at storm.trident.topology.state.TransactionalState.ne >>>>> >>>>> >>>>> >>>>> I check the storm logs, I see such errors on workers.log >>>>> >>>>> 2015-02-06T10:36:39.667-0800 b.s.m.n.Client [INFO] Reconnect started >>>>> for Netty-Client-pof-kstorm-dev2.pof.local:6700... [8] >>>>> 2015-02-06T10:36:39.668-0800 b.s.m.n.StormClientErrorHandler [INFO] >>>>> Connection failed Netty-Client-pof-kstorm-dev2.pof.local:6700 >>>>> java.nio.channels.UnresolvedAddressException: null >>>>> at sun.nio.ch.Net.checkAddress(Net.java:127) ~[na:1.7.0_65] >>>>> at >>>>> sun.nio.ch.SocketChannelImpl.connect(SocketChannelImpl.java:644) >>>>> ~[na:1.7.0_65] >>>>> at >>>>> org.apache.storm.netty.channel.socket.nio.NioClientSocketPipelineSink.connect(NioClientSocketPipelineSink.java:108) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.channel.socket.nio.NioClientSocketPipelineSink.eventSunk(NioClientSocketPipelineSink.java:70) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.channel.DefaultChannelPipeline$DefaultChannelHandlerContext.sendDownstream(DefaultChannelPipeline.java:779) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.handler.codec.oneone.OneToOneEncoder.handleDownstream(OneToOneEncoder.java:54) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.channel.DefaultChannelPipeline.sendDownstream(DefaultChannelPipeline.java:591) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.channel.DefaultChannelPipeline.sendDownstream(DefaultChannelPipeline.java:582) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.channel.Channels.connect(Channels.java:634) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.channel.AbstractChannel.connect(AbstractChannel.java:207) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.bootstrap.ClientBootstrap.connect(ClientBootstrap.java:229) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> org.apache.storm.netty.bootstrap.ClientBootstrap.connect(ClientBootstrap.java:182) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> backtype.storm.messaging.netty.Client.connect(Client.java:152) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> backtype.storm.messaging.netty.Client.access$000(Client.java:43) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> backtype.storm.messaging.netty.Client$1.run(Client.java:107) >>>>> [storm-core-0.9.3.jar:0.9.3] >>>>> at >>>>> java.util.concurrent.Executors$RunnableAdapter.call(Executors.java:471) >>>>> [na:1.7.0_65] >>>>> at java.util.concurrent.FutureTask.run(FutureTask.java:262) >>>>> [na:1.7.0_65] >>>>> at >>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.access$201(ScheduledThreadPoolExecutor.java:178) >>>>> [na:1.7.0_65] >>>>> at >>>>> java.util.concurrent.ScheduledThreadPoolExecutor$ScheduledFutureTask.run(ScheduledThreadPoolExecutor.java:292) >>>>> [na:1.7.0_65] >>>>> at >>>>> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >>>>> [na:1.7.0_65] >>>>> at >>>>> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >>>>> [na:1.7.0_65] >>>>> at java.lang.Thread.run(Thread.java:745) [na:1.7.0_65] >>>>> >>>>> >>>>> Is there something configuration I did wrong, here is my nimbus >>>>> storm.yaml >>>>> >>>>> storm.zookeeper.servers: >>>>> - "zkserver" >>>>> - "slave1" >>>>> - "slave2" >>>>> >>>>> nimbus.host: "nimbus" >>>>> supervisor.slots.ports: >>>>> - 6700 >>>>> - 6701 >>>>> - 6702 >>>>> - 6703 >>>>> >>>>> nimbus.childopts: "-Xmx1024m -Djava.net.preferIPv4Stack=true" >>>>> ui.childopts: "-Xmx768m -Djava.net.preferIPv4Stack=true" >>>>> supervisor.childopts: "-Djava.net.preferIPv4Stack=true" >>>>> worker.childopts: "-Xmx768m -Djava.net.preferIPv4Stack=true" >>>>> storm.local.dir: "/app/storm" >>>>> >>>>> >>>>> The supervisor storm.yaml >>>>> storm.zookeeper.servers: >>>>> - "zkserver" >>>>> - "slave1" >>>>> - "slave2" >>>>> >>>>> nimbus.host: "nimbus" >>>>> nimbus.childopts: "-Xmx1024m -Djava.net.preferIPv4Stack=true" >>>>> ui.childopts: "-Xmx768m -Djava.net.preferIPv4Stack=true" >>>>> supervisor.childopts: "-Djava.net.preferIPv4Stack=true" >>>>> worker.childopts: "-Xmx768m -Djava.net.preferIPv4Stack=true" >>>>> storm.local.dir: "/app/storm" >>>>> >>>>> >>>>> Thanks >>>>> >>>>> AL >>>>> >>>> >>>> >>> >> >
