Do you mean the config in the yaml file? I increased the worker memory and the spout is able to emit more tuples; the error is delayed but still there! The weird thing is that there are no tuple failures.. Il 04/Set/2014 20:08 "Vikas Agarwal" <[email protected]> ha scritto:
> I am not sure about it, however, looking at all possible config options > for storm would help. I did the same for one of my issues and found one > config option that was causing tuple failures. > > > On Thu, Sep 4, 2014 at 9:47 PM, Alberto Cordioli < > [email protected]> wrote: > >> That one is the full error log for the worker. No errors in >> supervisors and nimbus. >> That worker is associated with a spout that tries to make connection >> to HDFS to read avro files. Could be a problem related to this? >> >> >> On 4 September 2014 18:07, Vikas Agarwal <[email protected]> wrote: >> > Is it full error log? I mean we can look into source code where the >> worker >> > is trying to make some connection and may be we can guess what is wrong >> with >> > it. >> > >> > >> > On Thu, Sep 4, 2014 at 9:09 PM, Alberto Cordioli >> > <[email protected]> wrote: >> >> >> >> I've found this post describing the same problem. Unfortunately no >> >> answers: >> >> >> https://www.mail-archive.com/[email protected]/msg03623.html >> >> >> >> On 3 September 2014 18:58, Alberto Cordioli < >> [email protected]> >> >> wrote: >> >> > Hi all, >> >> > >> >> > I searched for similar problems without any luck. >> >> > I implemented a spout that continuously get this exception when >> >> > emitting "more than a certain number of tuples". I was not able to >> >> > understand how much this amount is, but I emit tuples in the order of >> >> > millions per seconds. >> >> > I've seen that other people had my same problem and resolved tuning >> >> > the ack executors parameter. In my case I don't have ackers (disabled >> >> > at spout level) and hence it couldn't related to this problem. >> >> > >> >> > The supervisor and nimbus logs look fine. The only problem I have is >> >> > in the spout worker: >> >> > >> >> > java.io.IOException: Connection reset by peer >> >> > at sun.nio.ch.FileDispatcherImpl.read0(Native Method) ~[na:1.7.0_65] >> >> > at sun.nio.ch.SocketDispatcher.read(SocketDispatcher.java:39) >> >> > ~[na:1.7.0_65] >> >> > at sun.nio.ch.IOUtil.readIntoNativeBuffer(IOUtil.java:223) >> >> > ~[na:1.7.0_65] >> >> > at sun.nio.ch.IOUtil.read(IOUtil.java:192) ~[na:1.7.0_65] >> >> > at sun.nio.ch.SocketChannelImpl.read(SocketChannelImpl.java:379) >> >> > ~[na:1.7.0_65] >> >> > at >> org.jboss.netty.channel.socket.nio.NioWorker.read(NioWorker.java:322) >> >> > ~[netty-3.2.2.Final.jar:na] >> >> > at >> >> > >> org.jboss.netty.channel.socket.nio.NioWorker.processSelectedKeys(NioWorker.java:281) >> >> > ~[netty-3.2.2.Final.jar:na] >> >> > at >> org.jboss.netty.channel.socket.nio.NioWorker.run(NioWorker.java:201) >> >> > ~[netty-3.2.2.Final.jar:na] >> >> > at >> >> > >> org.jboss.netty.util.internal.IoWorkerRunnable.run(IoWorkerRunnable.java:46) >> >> > [netty-3.2.2.Final.jar:na] >> >> > at >> >> > >> java.util.concurrent.ThreadPoolExecutor.runWorker(ThreadPoolExecutor.java:1145) >> >> > [na:1.7.0_65] >> >> > at >> >> > >> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:615) >> >> > [na:1.7.0_65] >> >> > at java.lang.Thread.run(Thread.java:745) [na:1.7.0_65] >> >> > >> >> > Does someone have an idea of why this happens? >> >> > >> >> > Thanks, >> >> > Alberto >> >> > >> >> > -- >> >> > Alberto Cordioli >> >> >> >> >> >> >> >> -- >> >> Alberto Cordioli >> > >> > >> > >> > >> > -- >> > Regards, >> > Vikas Agarwal >> > 91 – 9928301411 >> > >> > InfoObjects, Inc. >> > Execution Matters >> > http://www.infoobjects.com >> > 2041 Mission College Boulevard, #280 >> > Santa Clara, CA 95054 >> > +1 (408) 988-2000 Work >> > +1 (408) 716-2726 Fax >> >> >> >> -- >> Alberto Cordioli >> > > > > -- > Regards, > Vikas Agarwal > 91 – 9928301411 > > InfoObjects, Inc. > Execution Matters > http://www.infoobjects.com > 2041 Mission College Boulevard, #280 > Santa Clara, CA 95054 > +1 (408) 988-2000 Work > +1 (408) 716-2726 Fax > >
