Josh, Could you provide the parameters you used to configure the bloom filter?
Thanks. --- Jim Kellerman, Senior Engineer; Powerset [EMAIL PROTECTED] > -----Original Message----- > From: Josh Wills [mailto:[EMAIL PROTECTED] > Sent: Sunday, October 21, 2007 7:28 PM > To: hadoop-user@lucene.apache.org > Subject: Re: A basic question on HBase > 2) I was running one of these batch-style uploads last night > on an HTable that I configured w/BloomFilters on a couple of > my column families. During one of the compaction operations, > I got the following exception-- > > FATAL org.apache.hadoop.hbase.HRegionServer: Set stop flag in > regionserver/0:0:0:0:0:0:0:0:60020.splitOrCompactChecker > java.lang.ArrayIndexOutOfBoundsException > at java.lang.System.arraycopy(Native Method) > at > sun.security.provider.DigestBase.engineUpdate(DigestBase.java:102) > at sun.security.provider.SHA.implDigest(SHA.java:94) > at > sun.security.provider.DigestBase.engineDigest(DigestBase.java:161) > at > sun.security.provider.DigestBase.engineDigest(DigestBase.java:140) > at > java.security.MessageDigest$Delegate.engineDigest(MessageDiges t.java:531) > at java.security.MessageDigest.digest(MessageDigest.java:309) > at org.onelab.filter.HashFunction.hash(HashFunction.java:125) > at org.onelab.filter.BloomFilter.add(BloomFilter.java:99) > at > org.apache.hadoop.hbase.HStoreFile$BloomFilterMapFile$Writer.a ppend(HStoreFile.java:895) > at org.apache.hadoop.hbase.HStore.compact(HStore.java:899) > at org.apache.hadoop.hbase.HStore.compact(HStore.java:728) > at > org.apache.hadoop.hbase.HStore.compactHelper(HStore.java:632) > at > org.apache.hadoop.hbase.HStore.compactHelper(HStore.java:564) > at org.apache.hadoop.hbase.HStore.compact(HStore.java:559) > at > org.apache.hadoop.hbase.HRegion.compactStores(HRegion.java:717) > at > org.apache.hadoop.hbase.HRegionServer$SplitOrCompactChecker.ch eckForSplitsOrCompactions(HRegionServer.java:198) > at > org.apache.hadoop.hbase.HRegionServer$SplitOrCompactChecker.ch ore(HRegionServer.java:188) > at org.apache.hadoop.hbase.Chore.run(Chore.java:58) > > Note that this wasn't the first compaction that was run > (there were others before it that ran successfully) and that > the region hadn't been split at this point. I defined the > BloomFilterType.BLOOMFILTER on a couple of the > columnfamilies, w/the largest one having ~100000 distinct > entries. I don't know which of these caused the failure, but > I noticed that 100000 is quite a bit larger than the # of > entries used in the testcases, so I'm wondering if that might > be the problem. > > Thanks again, the 0.15.0 stuff looks very good- Josh > > > On 10/19/07, edward yoon <[EMAIL PROTECTED]> wrote: > > > > You're welcome. > > If you have any needs, questions, or comments in Hbase, > please let us > > know! > > > > Edward. > > ---- > > B. Regards, > > Edward yoon (Assistant Manager/R&D Center/NHN, corp.) > > +82-31-600-6183, +82-10-7149-7856 > > > > > > > Date: Fri, 19 Oct 2007 14:33:45 +0800 > > > From: [EMAIL PROTECTED] > > > To: hadoop-user@lucene.apache.org > > > Subject: Re: A basic question on HBase > > > > > > Dear edward yoon & Michael Stack, > > > > > > After using the hadoop branch-0.15, hbase runs correctly. > > > > > > Thank you very much! > > > > > > Best wishes, > > > Bin YANG > > > > > > On 10/19/07, Bin YANG wrote: > > >> Thank you! I can download it now! > > >> > > >> On 10/19/07, edward yoon wrote: > > >>> > > >>> Run the following on the command-line: > > >>> > > >>> $ svn co http://svn.apache.org/repos/asf/lucene/hadoop/trunk > > >>> hadoop > > >>> > > >>> See also for more information about the Hbase and Hbase > Shell client program: > > >>> > > >>> - http://wiki.apache.org/lucene-hadoop/Hbase > > >>> - http://wiki.apache.org/lucene-hadoop/Hbase/HbaseShell > > >>> > > >>> > > >>> Edward. > > >>> ---- > > >>> B. Regards, > > >>> Edward yoon (Assistant Manager/R&D Center/NHN, corp.) > > >>> +82-31-600-6183, +82-10-7149-7856 > > >>> > > >>> > > >>>> Date: Fri, 19 Oct 2007 13:46:51 +0800 > > >>>> From: [EMAIL PROTECTED] > > >>>> To: hadoop-user@lucene.apache.org > > >>>> Subject: Re: A basic question on HBase > > >>>> > > >>>> Dear Michael Stack: > > >>>> > > >>>> I am afraid that I cannot connect to the svn, > > >>>> > > >>>> Error: PROPFIND request failed on '/viewvc/lucene/hadoop/trunk' > > >>>> Error: PROPFIND of '/viewvc/lucene/hadoop/trunk': 302 Found > > >>>> (http://svn.apache.org) > > >>>> > > >>>> and > > >>>> > > >>>> Error: PROPFIND request failed on > '/viewvc/lucene/hadoop/branches/branch-0.15' > > >>>> Error: PROPFIND of > '/viewvc/lucene/hadoop/branches/branch-0.15': > > >>>> 302 Found (http://svn.apache.org) > > >>>> > > >>>> Would you please send me a 0.15 version of hadoop, or > give some > > >>>> information on how to connect to the svn successfully? > > >>>> > > >>>> Best wishes, > > >>>> Bin YANG > > >>>> > > >>>> > > >>>> > > >>>> > > >>>> > > >>>> On 10/19/07, Michael Stack wrote: > > >>>>> (Ignore my last message. I had missed your back and > forth with Edward). > > >>>>> > > >>>>> Regards step 3. below, you are starting both > mapreduce and dfs daemons. > > >>>>> You only need dfs daemons running hbase so you could do > > >>>>> ./bin/start-dfs.sh instead. > > >>>>> > > >>>>> Are you using hadoop 0.14.x? (It looks like it going by the > > >>>>> commands and log excerpt below). If so, please use > TRUNK or the > > >>>>> 0.15.0 candidate (Branch is here > > >>>>> > http://svn.apache.org/viewvc/lucene/hadoop/branches/branch-0.15/). > > >>>>> There is a big difference between hbase 0.14.0 and > 0.15.0 (The > > >>>>> 0.15.0 candidate contains the first hbase release). > For example > > >>>>> vestige log files are properly split and distributed in later > > >>>>> hbases where before they threw the "Can not start > region server because..." exception. > > >>>>> > > >>>>> As Edward points out, the master does not seem to be > getting the > > >>>>> region server 'report-for-duty' message (which > doesn't jibe with > > >>>>> the region server log that says -ROOT- has been > deployed because > > >>>>> master assigns regions). > > >>>>> > > >>>>> Regards your not being able to reformat -- presuming > no valuable > > >>>>> data in your hdfs, that all is running on localhost, and that > > >>>>> you are moving from hadoop 0.14.0 to 0.15.0 -- just > remove /tmp/hadoop-hadoop dir. > > >>>>> > > >>>>> St.Ack > > >>>>> > > >>>>> > > >>>>> > > >>>>> > > >>>>> Bin YANG wrote: > > >>>>>> Dear edward, > > >>>>>> > > >>>>>> I will show you the steps what I have done: > > >>>>>> > > >>>>>> 1. hadoop-site.xml > > >>>>>> > > >>>>>> > > >>>>>> fs.default.name > > >>>>>> localhost:9000 > > >>>>>> Namenode > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> mapred.job.tracker > > >>>>>> localhost:9001 > > >>>>>> JobTracker > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> dfs.replication > > >>>>>> 1 > > >>>>>> > > >>>>>> > > >>>>>> 2. /hadoop-0.14.2$ bin/hadoop namenode -format 3. > > >>>>>> bin/start-all.sh 4. hbase.site.xml > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> hbase.master > > >>>>>> localhost:60000 > > >>>>>> The host and port that the HBase master runs at. > > >>>>>> TODO: Support 'local' (All running in single context). > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> hbase.regionserver > > >>>>>> localhost:60010 > > >>>>>> The host and port a HBase region server runs at. > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> > > >>>>>> 5. bin/hbase-start.sh > > >>>>>> > > >>>>>> The log: > > >>>>>> 1. hbase-hadoop-regionserver-yangbin.log > > >>>>>> > > >>>>>> 2007-10-18 15:40:58,588 INFO > org.apache.hadoop.util.NativeCodeLoader: > > >>>>>> Loaded the native-hadoop library > > >>>>>> 2007-10-18 15:40:58,592 INFO > > >>>>>> org.apache.hadoop.io.compress.zlib.ZlibFactory: Successfully > > >>>>>> loaded & initialized native-zlib library > > >>>>>> 2007-10-18 15:40:58,690 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server listener on 60010: starting > > >>>>>> 2007-10-18 15:40:58,692 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 3 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,694 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 4 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,692 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 2 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,691 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 1 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,696 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 5 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,691 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 0 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,696 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 6 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,697 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 7 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,698 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 8 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,699 INFO > org.apache.hadoop.hbase.HRegionServer: > > >>>>>> HRegionServer started at: 127.0.1.1:60010 > > >>>>>> 2007-10-18 15:40:58,709 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 9 on 60010: starting > > >>>>>> 2007-10-18 15:40:58,867 INFO org.apache.hadoop.hbase.HStore: > > >>>>>> HStore online for --ROOT--,,0/info > > >>>>>> 2007-10-18 15:40:58,872 INFO > org.apache.hadoop.hbase.HRegion: > > >>>>>> region --ROOT--,,0 available > > >>>>>> 2007-10-18 18:21:55,558 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: > localhost/127.0.0.1:60000. Already tried 1 time(s). > > >>>>>> 2007-10-18 18:21:56,577 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: > localhost/127.0.0.1:60000. Already tried 2 time(s). > > >>>>>> 2007-10-18 18:21:57,585 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: > localhost/127.0.0.1:60000. Already tried 3 time(s). > > >>>>>> 2007-10-18 18:21:58,593 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: > localhost/127.0.0.1:60000. Already tried 4 time(s). > > >>>>>> 2007-10-18 18:22:05,874 ERROR > org.apache.hadoop.hbase.HRegionServer: > > >>>>>> Can not start region server because > > >>>>>> org.apache.hadoop.hbase.RegionServerRunningException: region > > >>>>>> server already running at 127.0.1.1:60010 because logdir > > >>>>>> /tmp/hadoop-hadoop/hbase/log_yangbin_60010 exists at > > >>>>>> > org.apache.hadoop.hbase.HRegionServer.(HRegionServer.java:482) > > >>>>>> at > > >>>>>> > org.apache.hadoop.hbase.HRegionServer.(HRegionServer.java:407) > > >>>>>> at > > >>>>>> > org.apache.hadoop.hbase.HRegionServer.main(HRegionServer.java:1 > > >>>>>> 357) > > >>>>>> > > >>>>>> 2007-10-18 19:57:40,243 INFO > org.apache.hadoop.util.NativeCodeLoader: > > >>>>>> Loaded the native-hadoop library > > >>>>>> 2007-10-18 19:57:40,274 INFO > > >>>>>> org.apache.hadoop.io.compress.zlib.ZlibFactory: Successfully > > >>>>>> loaded & initialized native-zlib library > > >>>>>> 2007-10-18 19:57:40,364 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server listener on 60010: starting > > >>>>>> 2007-10-18 19:57:40,366 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 0 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,367 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 1 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,368 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 2 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,368 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 3 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,369 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 4 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,370 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 5 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,371 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 6 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,371 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 7 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,372 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 8 on 60010: starting > > >>>>>> 2007-10-18 19:57:40,373 INFO > org.apache.hadoop.hbase.HRegionServer: > > >>>>>> HRegionServer started at: 127.0.1.1:60010 > > >>>>>> 2007-10-18 19:57:40,384 INFO > org.apache.hadoop.ipc.Server: IPC > > >>>>>> Server handler 9 on 60010: starting > > >>>>>> 2007-10-18 19:57:41,118 INFO org.apache.hadoop.hbase.HStore: > > >>>>>> HStore online for --ROOT--,,0/info > > >>>>>> 2007-10-18 19:57:41,125 INFO > org.apache.hadoop.hbase.HRegion: > > >>>>>> region --ROOT--,,0 available > > >>>>>> > > >>>>>> 2. hbase-hadoop-master-yangbin.log > > >>>>>> > > >>>>>> There is a lot of the below statement > > >>>>>> > > >>>>>> 2007-10-18 15:52:52,885 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 1 time(s). > > >>>>>> 2007-10-18 15:52:53,892 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 2 time(s). > > >>>>>> 2007-10-18 15:52:54,900 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 3 time(s). > > >>>>>> 2007-10-18 15:52:55,904 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 4 time(s). > > >>>>>> 2007-10-18 15:52:56,912 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 5 time(s). > > >>>>>> 2007-10-18 15:52:57,924 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 6 time(s). > > >>>>>> 2007-10-18 15:52:58,928 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 7 time(s). > > >>>>>> 2007-10-18 15:52:59,932 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 8 time(s). > > >>>>>> 2007-10-18 15:53:00,936 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 9 time(s). > > >>>>>> 2007-10-18 15:53:01,939 INFO org.apache.hadoop.ipc.Client: > > >>>>>> Retrying connect to server: /127.0.1.1:60010. > Already tried 10 time(s). > > >>>>>> 2007-10-18 15:53:02,943 INFO > org.apache.hadoop.ipc.RPC: Server > > >>>>>> at /127.0.1.1:60010 not available yet, Zzzzz... > > >>>>>> > > >>>>>> > > >>>>> > > >>>>> > > >>>> > > >>>> > > >>>> -- > > >>>> Bin YANG > > >>>> Department of Computer Science and Engineering Fudan > University > > >>>> Shanghai, P. R. China > > >>>> EMail: [EMAIL PROTECTED] > > >>> > > >>> > _________________________________________________________________ > > >>> Windows Live Hotmail and Microsoft Office Outlook - > together at last. Get it now. > > >>> > http://office.microsoft.com/en-us/outlook/HA102225181033.aspx?pid= > > >>> CL100626971033 > > >> > > >> > > >> -- > > >> Bin YANG > > >> Department of Computer Science and Engineering Fudan University > > >> Shanghai, P. R. China > > >> EMail: [EMAIL PROTECTED] > > >> > > > > > > > > > -- > > > Bin YANG > > > Department of Computer Science and Engineering > > > Fudan University > > > Shanghai, P. R. China > > > EMail: [EMAIL PROTECTED] > > > > _________________________________________________________________ > > Windows Live Hotmail and Microsoft Office Outlook - > together at last. Get it now. > > > http://office.microsoft.com/en-us/outlook/HA102225181033.aspx? > pid=CL100626971033 >