Yes, it is. Thanks.

Best,
Arber



On Tue, Apr 17, 2012 at 12:05 AM, Jonathan Hsieh <j...@cloudera.com> wrote:

> Arber,
>
> Good to hear! Just to confirm, the bug/patch the same as HBASE-5488?
>
> Jon.
>
> On Sun, Apr 15, 2012 at 4:36 AM, Yabo Xu <arber.resea...@gmail.com> wrote:
>
> > Hi Jon:
> >
> > Please ignore my last email. We found it was a bug, fix it by a patch and
> > rebuild, and it works now. Data are back! Thanks.
> >
> > Best,
> > Arber
> >
> >
> >
> > On Sun, Apr 15, 2012 at 12:47 PM, Yabo Xu <arber.resea...@gmail.com>
> > wrote:
> >
> > > Dear Jon:
> > >
> > > We just ran OfflineMetaRepair, while getting the following exceptions.
> > > Checked online...it seems that is bug. Any suggestions on how to check
> > out
> > > the most-updated version of OfflineMetaRepair to work with our version
> of
> > > HBase? Thanks in advance.
> > >
> > > 12/04/15 12:28:35 INFO util.HBaseFsck: Loading HBase regioninfo from
> > > HDFS...
> > > 12/04/15 12:28:39 ERROR util.HBaseFsck: Bailed out due to:
> > > java.lang.IllegalArgumentException: Wrong FS: hdfs://
> > >
> >
> n4.example.com:12345/hbase/summba.yeezhao.content/03cde9116662fade27545d86ea71a372/.regioninfo
> > ,
> > > expected: file:///
> > >  at org.apache.hadoop.fs.FileSystem.checkPath(FileSystem.java:310)
> > > at
> > >
> >
> org.apache.hadoop.fs.RawLocalFileSystem.pathToFile(RawLocalFileSystem.java:47)
> > >  at
> > >
> >
> org.apache.hadoop.fs.RawLocalFileSystem.getFileStatus(RawLocalFileSystem.java:357)
> > > at
> > >
> >
> org.apache.hadoop.fs.FilterFileSystem.getFileStatus(FilterFileSystem.java:245)
> > >  at
> > >
> >
> org.apache.hadoop.fs.ChecksumFileSystem$ChecksumFSInputChecker.<init>(ChecksumFileSystem.java:125)
> > > at
> > >
> org.apache.hadoop.fs.ChecksumFileSystem.open(ChecksumFileSystem.java:283)
> > >  at org.apache.hadoop.fs.FileSystem.open(FileSystem.java:356)
> > > at
> > org.apache.hadoop.hbase.util.HBaseFsck.loadMetaEntry(HBaseFsck.java:256)
> > >  at
> > >
> org.apache.hadoop.hbase.util.HBaseFsck.loadTableInfo(HBaseFsck.java:284)
> > > at
> org.apache.hadoop.hbase.util.HBaseFsck.rebuildMeta(HBaseFsck.java:402)
> > >  at
> > >
> >
> org.apache.hadoop.hbase.util.hbck.OfflineMetaRepair.main(OfflineMetaRepair.java:90)
> > >
> > > We checked on hdfs, and the files shown in exception are available. Any
> > > point
> > >
> > > Best,
> > > Arber
> > >
> > >
> > > On Sun, Apr 15, 2012 at 11:48 AM, Yabo Xu <arber.resea...@gmail.com
> > >wrote:
> > >
> > >> Thanks, St. Ack & Jon. To answer St. Ack's question, we are using
> HBase
> > >> 0.90.6, and the data corruption happens when some data nodes are lost
> > due
> > >> to the power issue. We've tried hbck and it reports that ROOT is not
> > found,
> > >> and hfsk reports two blocks of ROOT and META are CORUPT status.
> > >>
> > >> Jon: We just checked OfflineMetaRepair, it seems to be the right tool,
> > >> and is trying it now. Just want to confirm: is it compatible with
> > 0.90.6?
> > >>
> > >> Best,
> > >> Arber
> > >>
> > >>
> > >> On Sun, Apr 15, 2012 at 8:55 AM, Jonathan Hsieh <j...@cloudera.com>
> > wrote:
> > >>
> > >>> There is a two tools that can try to help you (unfortunately, I
> haven't
> > >>> written the user documentation for either yet)
> > >>>
> > >>> One is called OfflineMetaRepair.  This assumes that hbase is offline
> > >>> reads
> > >>> the data in HDFS  to create a new ROOT and new META.  If you data is
> in
> > >>> good shape, this should work for you. Depending  on which version of
> > >>> hadoop
> > >>> you are using, you may need to apply HBASE-5488.
> > >>>
> > >>> On the latest branches of hbase (0.90/0.92/0.94/trunk) the hbck tool
> > has
> > >>> been greatly enhanced and may be able to help out as well once an
> > initial
> > >>> META table is built, and your hbase is able to get online.  This will
> > >>> currently will require a patch HBASE-5781 to be applied to be useful.
> > >>>
> > >>> Jon.
> > >>>
> > >>>
> > >>> On Sat, Apr 14, 2012 at 1:35 PM, Yabo Xu <arber.resea...@gmail.com>
> > >>> wrote:
> > >>>
> > >>> > Hi all:
> > >>> >
> > >>> > Just had a desperate  night....We had a small production hbase
> > >>> cluster( 8
> > >>> > nodes), and due to the accident crash of a few nodes, ROOT and META
> > are
> > >>> > corrupted, while the rest of tables are mostly there. Are there any
> > >>> way to
> > >>> > restore ROOT and META?
> > >>> >
> > >>> > Any of the hints would be appreciated very much! Waiting on line...
> > >>> >
> > >>> > Best,
> > >>> > Arber
> > >>> >
> > >>>
> > >>>
> > >>>
> > >>> --
> > >>> // Jonathan Hsieh (shay)
> > >>> // Software Engineer, Cloudera
> > >>> // j...@cloudera.com
> > >>>
> > >>
> > >>
> > >
> >
>
>
>
> --
> // Jonathan Hsieh (shay)
> // Software Engineer, Cloudera
> // j...@cloudera.com
>

Reply via email to