Josh,

Could you provide the parameters you used to configure the bloom filter?

Thanks.

---
Jim Kellerman, Senior Engineer; Powerset
[EMAIL PROTECTED]


> -----Original Message-----
> From: Josh Wills [mailto:[EMAIL PROTECTED]
> Sent: Sunday, October 21, 2007 7:28 PM
> To: hadoop-user@lucene.apache.org
> Subject: Re: A basic question on HBase
> 2)  I was running one of these batch-style uploads last night
> on an HTable that I configured w/BloomFilters on a couple of
> my column families.  During one of the compaction operations,
> I got the following exception--
>
> FATAL org.apache.hadoop.hbase.HRegionServer: Set stop flag in
> regionserver/0:0:0:0:0:0:0:0:60020.splitOrCompactChecker
> java.lang.ArrayIndexOutOfBoundsException
>         at java.lang.System.arraycopy(Native Method)
>         at
> sun.security.provider.DigestBase.engineUpdate(DigestBase.java:102)
>         at sun.security.provider.SHA.implDigest(SHA.java:94)
>         at
> sun.security.provider.DigestBase.engineDigest(DigestBase.java:161)
>         at
> sun.security.provider.DigestBase.engineDigest(DigestBase.java:140)
>         at
> java.security.MessageDigest$Delegate.engineDigest(MessageDiges
t.java:531)
>         at java.security.MessageDigest.digest(MessageDigest.java:309)
>         at org.onelab.filter.HashFunction.hash(HashFunction.java:125)
>         at org.onelab.filter.BloomFilter.add(BloomFilter.java:99)
>         at
> org.apache.hadoop.hbase.HStoreFile$BloomFilterMapFile$Writer.a
ppend(HStoreFile.java:895)
>         at org.apache.hadoop.hbase.HStore.compact(HStore.java:899)
>         at org.apache.hadoop.hbase.HStore.compact(HStore.java:728)
>         at
> org.apache.hadoop.hbase.HStore.compactHelper(HStore.java:632)
>         at
> org.apache.hadoop.hbase.HStore.compactHelper(HStore.java:564)
>         at org.apache.hadoop.hbase.HStore.compact(HStore.java:559)
>         at
> org.apache.hadoop.hbase.HRegion.compactStores(HRegion.java:717)
>         at
> org.apache.hadoop.hbase.HRegionServer$SplitOrCompactChecker.ch
eckForSplitsOrCompactions(HRegionServer.java:198)
>         at
> org.apache.hadoop.hbase.HRegionServer$SplitOrCompactChecker.ch
ore(HRegionServer.java:188)
>         at org.apache.hadoop.hbase.Chore.run(Chore.java:58)
>
> Note that this wasn't the first compaction that was run
> (there were others before it that ran successfully) and that
> the region hadn't been split at this point.  I defined the
> BloomFilterType.BLOOMFILTER on a couple of the
> columnfamilies, w/the largest one having ~100000 distinct
> entries.  I don't know which of these caused the failure, but
> I noticed that 100000 is quite a bit larger than the # of
> entries used in the testcases, so I'm wondering if that might
> be the problem.
>
> Thanks again, the 0.15.0 stuff looks very good- Josh
>
>
> On 10/19/07, edward yoon <[EMAIL PROTECTED]> wrote:
> >
> > You're welcome.
> > If you have any needs, questions, or comments in Hbase,
> please let us
> > know!
> >
> > Edward.
> > ----
> > B. Regards,
> > Edward yoon (Assistant Manager/R&D Center/NHN, corp.)
> > +82-31-600-6183, +82-10-7149-7856
> >
> >
> > > Date: Fri, 19 Oct 2007 14:33:45 +0800
> > > From: [EMAIL PROTECTED]
> > > To: hadoop-user@lucene.apache.org
> > > Subject: Re: A basic question on HBase
> > >
> > > Dear edward yoon & Michael Stack,
> > >
> > > After using the hadoop branch-0.15, hbase runs correctly.
> > >
> > > Thank you very much!
> > >
> > > Best wishes,
> > > Bin YANG
> > >
> > > On 10/19/07, Bin YANG  wrote:
> > >> Thank you! I can download it now!
> > >>
> > >> On 10/19/07, edward yoon  wrote:
> > >>>
> > >>> Run the following on the command-line:
> > >>>
> > >>> $ svn co http://svn.apache.org/repos/asf/lucene/hadoop/trunk
> > >>> hadoop
> > >>>
> > >>> See also for more information about the Hbase and Hbase
> Shell client program:
> > >>>
> > >>> - http://wiki.apache.org/lucene-hadoop/Hbase
> > >>> - http://wiki.apache.org/lucene-hadoop/Hbase/HbaseShell
> > >>>
> > >>>
> > >>> Edward.
> > >>> ----
> > >>> B. Regards,
> > >>> Edward yoon (Assistant Manager/R&D Center/NHN, corp.)
> > >>> +82-31-600-6183, +82-10-7149-7856
> > >>>
> > >>>
> > >>>> Date: Fri, 19 Oct 2007 13:46:51 +0800
> > >>>> From: [EMAIL PROTECTED]
> > >>>> To: hadoop-user@lucene.apache.org
> > >>>> Subject: Re: A basic question on HBase
> > >>>>
> > >>>> Dear Michael Stack:
> > >>>>
> > >>>> I am afraid that I cannot connect to the svn,
> > >>>>
> > >>>> Error: PROPFIND request failed on '/viewvc/lucene/hadoop/trunk'
> > >>>> Error: PROPFIND of '/viewvc/lucene/hadoop/trunk': 302 Found
> > >>>> (http://svn.apache.org)
> > >>>>
> > >>>> and
> > >>>>
> > >>>> Error: PROPFIND request failed on
> '/viewvc/lucene/hadoop/branches/branch-0.15'
> > >>>> Error: PROPFIND of
> '/viewvc/lucene/hadoop/branches/branch-0.15':
> > >>>> 302 Found (http://svn.apache.org)
> > >>>>
> > >>>> Would you please send me a 0.15 version of hadoop, or
> give some
> > >>>> information on how to connect to the svn successfully?
> > >>>>
> > >>>> Best wishes,
> > >>>> Bin YANG
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>>
> > >>>> On 10/19/07, Michael Stack wrote:
> > >>>>> (Ignore my last message. I had missed your back and
> forth with Edward).
> > >>>>>
> > >>>>> Regards step 3. below, you are starting both
> mapreduce and dfs daemons.
> > >>>>> You only need dfs daemons running hbase so you could do
> > >>>>> ./bin/start-dfs.sh instead.
> > >>>>>
> > >>>>> Are you using hadoop 0.14.x? (It looks like it going by the
> > >>>>> commands and log excerpt below). If so, please use
> TRUNK or the
> > >>>>> 0.15.0 candidate (Branch is here
> > >>>>>
> http://svn.apache.org/viewvc/lucene/hadoop/branches/branch-0.15/).
> > >>>>> There is a big difference between hbase 0.14.0 and
> 0.15.0 (The
> > >>>>> 0.15.0 candidate contains the first hbase release).
> For example
> > >>>>> vestige log files are properly split and distributed in later
> > >>>>> hbases where before they threw the "Can not start
> region server because..." exception.
> > >>>>>
> > >>>>> As Edward points out, the master does not seem to be
> getting the
> > >>>>> region server 'report-for-duty' message (which
> doesn't jibe with
> > >>>>> the region server log that says -ROOT- has been
> deployed because
> > >>>>> master assigns regions).
> > >>>>>
> > >>>>> Regards your not being able to reformat -- presuming
> no valuable
> > >>>>> data in your hdfs, that all is running on localhost, and that
> > >>>>> you are moving from hadoop 0.14.0 to 0.15.0 -- just
> remove /tmp/hadoop-hadoop dir.
> > >>>>>
> > >>>>> St.Ack
> > >>>>>
> > >>>>>
> > >>>>>
> > >>>>>
> > >>>>> Bin YANG wrote:
> > >>>>>> Dear edward,
> > >>>>>>
> > >>>>>> I will show you the steps what I have done:
> > >>>>>>
> > >>>>>> 1. hadoop-site.xml
> > >>>>>>
> > >>>>>>
> > >>>>>> fs.default.name
> > >>>>>> localhost:9000
> > >>>>>> Namenode
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> mapred.job.tracker
> > >>>>>> localhost:9001
> > >>>>>> JobTracker
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> dfs.replication
> > >>>>>> 1
> > >>>>>>
> > >>>>>>
> > >>>>>> 2. /hadoop-0.14.2$ bin/hadoop namenode -format 3.
> > >>>>>> bin/start-all.sh 4. hbase.site.xml
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> hbase.master
> > >>>>>> localhost:60000
> > >>>>>> The host and port that the HBase master runs at.
> > >>>>>> TODO: Support 'local' (All running in single context).
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> hbase.regionserver
> > >>>>>> localhost:60010
> > >>>>>> The host and port a HBase region server runs at.
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>>
> > >>>>>> 5. bin/hbase-start.sh
> > >>>>>>
> > >>>>>> The log:
> > >>>>>> 1. hbase-hadoop-regionserver-yangbin.log
> > >>>>>>
> > >>>>>> 2007-10-18 15:40:58,588 INFO
> org.apache.hadoop.util.NativeCodeLoader:
> > >>>>>> Loaded the native-hadoop library
> > >>>>>> 2007-10-18 15:40:58,592 INFO
> > >>>>>> org.apache.hadoop.io.compress.zlib.ZlibFactory: Successfully
> > >>>>>> loaded & initialized native-zlib library
> > >>>>>> 2007-10-18 15:40:58,690 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server listener on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,692 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 3 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,694 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 4 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,692 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 2 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,691 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 1 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,696 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 5 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,691 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 0 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,696 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 6 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,697 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 7 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,698 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 8 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,699 INFO
> org.apache.hadoop.hbase.HRegionServer:
> > >>>>>> HRegionServer started at: 127.0.1.1:60010
> > >>>>>> 2007-10-18 15:40:58,709 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 9 on 60010: starting
> > >>>>>> 2007-10-18 15:40:58,867 INFO org.apache.hadoop.hbase.HStore:
> > >>>>>> HStore online for --ROOT--,,0/info
> > >>>>>> 2007-10-18 15:40:58,872 INFO
> org.apache.hadoop.hbase.HRegion:
> > >>>>>> region --ROOT--,,0 available
> > >>>>>> 2007-10-18 18:21:55,558 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server:
> localhost/127.0.0.1:60000. Already tried 1 time(s).
> > >>>>>> 2007-10-18 18:21:56,577 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server:
> localhost/127.0.0.1:60000. Already tried 2 time(s).
> > >>>>>> 2007-10-18 18:21:57,585 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server:
> localhost/127.0.0.1:60000. Already tried 3 time(s).
> > >>>>>> 2007-10-18 18:21:58,593 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server:
> localhost/127.0.0.1:60000. Already tried 4 time(s).
> > >>>>>> 2007-10-18 18:22:05,874 ERROR
> org.apache.hadoop.hbase.HRegionServer:
> > >>>>>> Can not start region server because
> > >>>>>> org.apache.hadoop.hbase.RegionServerRunningException: region
> > >>>>>> server already running at 127.0.1.1:60010 because logdir
> > >>>>>> /tmp/hadoop-hadoop/hbase/log_yangbin_60010 exists at
> > >>>>>>
> org.apache.hadoop.hbase.HRegionServer.(HRegionServer.java:482)
> > >>>>>> at
> > >>>>>>
> org.apache.hadoop.hbase.HRegionServer.(HRegionServer.java:407)
> > >>>>>> at
> > >>>>>>
> org.apache.hadoop.hbase.HRegionServer.main(HRegionServer.java:1
> > >>>>>> 357)
> > >>>>>>
> > >>>>>> 2007-10-18 19:57:40,243 INFO
> org.apache.hadoop.util.NativeCodeLoader:
> > >>>>>> Loaded the native-hadoop library
> > >>>>>> 2007-10-18 19:57:40,274 INFO
> > >>>>>> org.apache.hadoop.io.compress.zlib.ZlibFactory: Successfully
> > >>>>>> loaded & initialized native-zlib library
> > >>>>>> 2007-10-18 19:57:40,364 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server listener on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,366 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 0 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,367 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 1 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,368 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 2 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,368 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 3 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,369 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 4 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,370 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 5 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,371 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 6 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,371 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 7 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,372 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 8 on 60010: starting
> > >>>>>> 2007-10-18 19:57:40,373 INFO
> org.apache.hadoop.hbase.HRegionServer:
> > >>>>>> HRegionServer started at: 127.0.1.1:60010
> > >>>>>> 2007-10-18 19:57:40,384 INFO
> org.apache.hadoop.ipc.Server: IPC
> > >>>>>> Server handler 9 on 60010: starting
> > >>>>>> 2007-10-18 19:57:41,118 INFO org.apache.hadoop.hbase.HStore:
> > >>>>>> HStore online for --ROOT--,,0/info
> > >>>>>> 2007-10-18 19:57:41,125 INFO
> org.apache.hadoop.hbase.HRegion:
> > >>>>>> region --ROOT--,,0 available
> > >>>>>>
> > >>>>>> 2. hbase-hadoop-master-yangbin.log
> > >>>>>>
> > >>>>>> There is a lot of the below statement
> > >>>>>>
> > >>>>>> 2007-10-18 15:52:52,885 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 1 time(s).
> > >>>>>> 2007-10-18 15:52:53,892 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 2 time(s).
> > >>>>>> 2007-10-18 15:52:54,900 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 3 time(s).
> > >>>>>> 2007-10-18 15:52:55,904 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 4 time(s).
> > >>>>>> 2007-10-18 15:52:56,912 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 5 time(s).
> > >>>>>> 2007-10-18 15:52:57,924 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 6 time(s).
> > >>>>>> 2007-10-18 15:52:58,928 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 7 time(s).
> > >>>>>> 2007-10-18 15:52:59,932 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 8 time(s).
> > >>>>>> 2007-10-18 15:53:00,936 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 9 time(s).
> > >>>>>> 2007-10-18 15:53:01,939 INFO org.apache.hadoop.ipc.Client:
> > >>>>>> Retrying connect to server: /127.0.1.1:60010.
> Already tried 10 time(s).
> > >>>>>> 2007-10-18 15:53:02,943 INFO
> org.apache.hadoop.ipc.RPC: Server
> > >>>>>> at /127.0.1.1:60010 not available yet, Zzzzz...
> > >>>>>>
> > >>>>>>
> > >>>>>
> > >>>>>
> > >>>>
> > >>>>
> > >>>> --
> > >>>> Bin YANG
> > >>>> Department of Computer Science and Engineering Fudan
> University
> > >>>> Shanghai, P. R. China
> > >>>> EMail: [EMAIL PROTECTED]
> > >>>
> > >>>
> _________________________________________________________________
> > >>> Windows Live Hotmail and Microsoft Office Outlook -
> together at last. Get it now.
> > >>>
> http://office.microsoft.com/en-us/outlook/HA102225181033.aspx?pid=
> > >>> CL100626971033
> > >>
> > >>
> > >> --
> > >> Bin YANG
> > >> Department of Computer Science and Engineering Fudan University
> > >> Shanghai, P. R. China
> > >> EMail: [EMAIL PROTECTED]
> > >>
> > >
> > >
> > > --
> > > Bin YANG
> > > Department of Computer Science and Engineering
> > > Fudan University
> > > Shanghai, P. R. China
> > > EMail: [EMAIL PROTECTED]
> >
> > _________________________________________________________________
> > Windows Live Hotmail and Microsoft Office Outlook -
> together at last. Get it now.
> >
> http://office.microsoft.com/en-us/outlook/HA102225181033.aspx?
> pid=CL100626971033
>

Reply via email to