You haven't answered all questions yet :) Are you running this on EC2?
What instance types?

On Thu, Nov 18, 2010 at 12:12 AM, Lucas Nazário dos Santos
<nazario.lu...@gmail.com> wrote:
> It seems that newer Linux versions don't have the
> file /proc/sys/fs/epoll/max_user_instances, but instead
> /proc/sys/fs/epoll/max_user_watches. I'm not quite sure about what to do.
>
> Can I favor max_user_watches over max_user_instances? With what value?
>
> I also tried to play with the Xss argument and decreased it to 128K with no
> luck (xcievers at 4096).
>
> Lucas
>
>
>
> On Wed, Nov 17, 2010 at 8:02 PM, Lars George <lars.geo...@gmail.com> wrote:
>
>> That is what I was also thinking about, thanks for jumping in Todd.
>>
>> I was simply not sure if that is just on .27 or all after that one and
>> the defaults have never been increased.
>>
>> On Wed, Nov 17, 2010 at 8:24 PM, Todd Lipcon <t...@cloudera.com> wrote:
>> > On that new of a kernel you'll also need to increase your epoll limit.
>> Some
>> > tips about that here:
>> >
>> >
>> http://www.cloudera.com/blog/2009/03/configuration-parameters-what-can-you-just-ignore/
>> >
>> > Thanks
>> > -Todd
>> >
>> > On Wed, Nov 17, 2010 at 9:10 AM, Lars George <lars.geo...@gmail.com>
>> wrote:
>> >
>> >> Are you running on EC2? Couldn't you simply up the heap size for the
>> >> java processes?
>> >>
>> >> I do not think there is a hard and fast rule to how many xcievers you
>> >> need, trial and error is common. Or ifmyou have enough heap simply set
>> >> it too high, like 4096 and that usually works fine. It all depends on
>> >> how many regions and column families you have on each server.
>> >>
>> >> Lars
>> >>
>> >> On Wed, Nov 17, 2010 at 5:31 PM, Lucas Nazário dos Santos
>> >> <nazario.lu...@gmail.com> wrote:
>> >> > I'm using Linux, the Amazon beta version that they recently released.
>> I'm
>> >> > not very familiar with Linux, so I think the kernel version
>> >> > is 2.6.34.7-56.40.amzn1.x86_64. Hadoop version is 0.20.2 and HBase
>> >> version
>> >> > is 0.20.6. Hadoop and HBase have 2 GB each and they are not sawpping.
>> >> >
>> >> > Besides all other questions I posed, I have one more. How can I
>> calculate
>> >> > the maximum number of xcievers? Is there a formula?
>> >> >
>> >> > Lucas
>> >> >
>> >> >
>> >> >
>> >> > On Wed, Nov 17, 2010 at 2:12 PM, Lars George <lars.geo...@gmail.com>
>> >> wrote:
>> >> >
>> >> >> Hi Lucas,
>> >> >>
>> >> >> What OS are you on? What kernel version? What is your Hadoop and
>> HBase
>> >> >> version? How much heap do you assign to each Java process?
>> >> >>
>> >> >> Lars
>> >> >>
>> >> >> On Wed, Nov 17, 2010 at 3:05 PM, Lucas Nazário dos Santos
>> >> >> <nazario.lu...@gmail.com> wrote:
>> >> >> > Hi,
>> >> >> >
>> >> >> > This problem is widely know, but I'm not able to come up with a
>> decent
>> >> >> > solution for it.
>> >> >> >
>> >> >> > I'm scanning 1.000.000+ rows from one table in order to index their
>> >> >> content.
>> >> >> > Each row has around 100 KB. The problem is that I keep getting the
>> >> >> > exception:
>> >> >> >
>> >> >> > Exception in thread
>> >> >> "org.apache.hadoop.dfs.datanode$dataxceiveser...@82d37"
>> >> >> > java.lang.OutOfMemoryError: unable to create new native thread
>> >> >> >
>> >> >> > This is a Hadoop exception and it causes the DataNote to go down,
>> so I
>> >> >> > decreased the dfs.datanode.max.xcievers from 4048 to 512. Well,
>> that
>> >> led
>> >> >> me
>> >> >> > to another problem:
>> >> >> >
>> >> >> > java.io.IOException: xceiverCount 513 exceeds the limit of
>> concurrent
>> >> >> > xcievers 512
>> >> >> >
>> >> >> > This time the DataNode doesn't die, nor HBase, but my scan, and the
>> >> whole
>> >> >> > indexing process, suffers a lot.
>> >> >> >
>> >> >> > After reading different posts about this issue, I have the
>> impression
>> >> >> that
>> >> >> > HBase can't handle this limits transparently for the user. The
>> scanner
>> >> is
>> >> >> a
>> >> >> > sequential process, so I thought it would free Hadoop resources
>> >> already
>> >> >> used
>> >> >> > in order to make room for new requests for data under HDFS. What I
>> am
>> >> >> > missing? Should I slow down the scanning process? Should I scan
>> >> portions
>> >> >> of
>> >> >> > the table sequentially instead of doing a full scan in all
>> 1.000.000+
>> >> >> rows?
>> >> >> > Is there a timeout so unused Hadoop resources can be released?
>> >> >> >
>> >> >> > Thanks in advance,
>> >> >> > Lucas
>> >> >> >
>> >> >>
>> >> >
>> >>
>> >
>> >
>> >
>> > --
>> > Todd Lipcon
>> > Software Engineer, Cloudera
>> >
>>
>

Reply via email to