Mike,

 

Thanks for the very helpful and in-depth explanation!  That will be very 
helpful in the future as I try to move this forward.

 

-Dave

 

From: Mike Percy [mailto:[email protected]] 
Sent: Thursday, October 11, 2012 6:09 PM
To: [email protected]
Subject: Re: Errors

 

You should consider how your system will act if there is a downstream failure. 
Even a capacity of 500 is extremely (orders of magnitude) too small in my 
opinion.

 

Consider setting a channel capacity equal to (average events per second 
ingested * # of seconds downtime you want to tolerate). So if you are ingesting 
1000 events/sec and you want to tolerate 1 hour of downtime without dropping 
events, you would want a channel capacity of 1000 * (60 * 60) = 3600000. Don't 
forget that the channel is a buffer that is intended to smooth out the 
latencies inherent in a complex network of storage systems. Even HDFS and HBase 
have latency hiccups sometimes, so try to avoid running close to your buffer 
capacity.

 

Regards,

Mike

 

On Thu, Oct 11, 2012 at 11:37 AM, Harish Mandala <[email protected]> 
wrote:

I've noticed in general that capacity = 100*transactionCapacity (or 
10*transactionCapacity) works well for me. 
 

        Regards,
        Harish 

         

        On Thu, Oct 11, 2012 at 2:34 PM, Cochran, David M (Contractor) 
<[email protected]> wrote:

        Trying that now...  set to 500 for each channel...  we'll see how it
        goes.
        
        For some reason 'channel capacity' didn't equate with the error msg.  In
        my figuring the part about the sinks not being able to keep up led me in
        another direction.... maybe I wasn't holding my head just right :)
        
        Thanks for the quick response!
        -Dave

        
        
        -----Original Message-----
        From: Brock Noland [mailto:[email protected]]
        Sent: Thursday, October 11, 2012 1:14 PM
        To: [email protected]
        Subject: Re: Errors
        
        Basically the channel is filling up. Have you increased the capacity of
        the channel?
        
        On Thu, Oct 11, 2012 at 1:08 PM, Cochran, David M (Contractor)
        <[email protected]> wrote:
        >
        > This error insists on making an appearance at least daily on my test
        > systems.
        >
        > Unable to put batch on required channel:
        > org.apache.flume.channel.MemoryChannel@555c07d8
        > Caused by: org.apache.flume.ChannelException: Space for commit to
        > queue couldn't be acquired Sinks are likely not keeping up with
        > sources, or the buffer size is too tight
        >
        > Changing the batch-size and batchSize around from default to 100's or
        > a 1000 doesn't seem to help
        >
        > increased the JAVA_OPTS="-Xms256m -Xmx512m"  still no change
        >
        > This shows up very intermittently, but daily, the logs being tailed
        > are not very big, and are not growing very quickly, actually very slow
        
        > in the grand scheme of things.
        >
        > Am I missing something to help balance things out here?
        >
        > Thanks
        > Dave
        
        
        
        --
        Apache MRUnit - Unit testing MapReduce -
        http://incubator.apache.org/mrunit/

         

 

Reply via email to