Gagandeep: Take a look at https://issues.apache.org/jira/browse/HBASE-2789
I think you need to propagate hbase-site.xml to region servers. On Thu, Sep 2, 2010 at 3:34 AM, Gagandeep Singh <[email protected]>wrote: > Hi Daniel > > I have downloaded hadoop-0.20.2+320.tar.gz from this location > http://archive.cloudera.com/cdh/3/ > And also changed the *dfs.support.append* flag to *true* in your * > hdfs-site.xml* as mentioned here > http://wiki.apache.org/hadoop/Hbase/HdfsSyncSupport. > > But data loss is still happening. Am I using the right version? > Is there any other settings that I need to make so that data gets flushed > to > HDFS. > > Thanks, > Gagan > > > > On Thu, Aug 26, 2010 at 11:57 PM, Jean-Daniel Cryans <[email protected] > >wrote: > > > That, or use CDH3b2. > > > > J-D > > > > On Thu, Aug 26, 2010 at 11:22 AM, Gagandeep Singh > > <[email protected]> wrote: > > > Thanks Daniel > > > > > > It means I have to checkout the code from branch and build it on my > local > > > machine. > > > > > > Gagan > > > > > > > > > On Thu, Aug 26, 2010 at 9:51 PM, Jean-Daniel Cryans < > [email protected] > > >wrote: > > > > > >> Then I would expect some form of dataloss yes, because stock hadoop > > >> 0.20 doesn't have any form of fsync so HBase doesn't know whether the > > >> data made it to the datanodes when appending to the WAL. Please use > > >> the 0.20-append hadoop branch with HBase 0.89 or cloudera's CDH3b2. > > >> > > >> J-D > > >> > > >> On Thu, Aug 26, 2010 at 7:22 AM, Gagandeep Singh > > >> <[email protected]> wrote: > > >> > HBase - 0.20.5 > > >> > Hadoop - 0.20.2 > > >> > > > >> > Thanks, > > >> > Gagan > > >> > > > >> > > > >> > > > >> > On Thu, Aug 26, 2010 at 7:11 PM, Jean-Daniel Cryans < > > [email protected] > > >> >wrote: > > >> > > > >> >> Hadoop and HBase version? > > >> >> > > >> >> J-D > > >> >> > > >> >> On Aug 26, 2010 5:36 AM, "Gagandeep Singh" < > > [email protected]> > > >> >> wrote: > > >> >> > > >> >> Hi Group, > > >> >> > > >> >> I am checking HBase/HDFS fail over. I am inserting 1M records from > my > > >> HBase > > >> >> client application. I am clubbing my Put operation such that 10 > > records > > >> get > > >> >> added into the List<Put> and then I call the table.put(). I have > not > > >> >> modified the default setting of Put operation which means all data > is > > >> >> written in WAL and in case of server failure my data should not be > > lost. > > >> >> > > >> >> But I noticed somewhat strange behavior, while adding records if I > > kill > > >> my > > >> >> Region Server then my application waits till the time region data > is > > >> moved > > >> >> to another region. But I noticed while doing so all my data is lost > > and > > >> my > > >> >> table is emptied. > > >> >> > > >> >> Could you help me understand the behavior. Is there some kind of > > Cache > > >> also > > >> >> involved while writing because of which my data is lost. > > >> >> > > >> >> > > >> >> Thanks, > > >> >> Gagan > > >> >> > > >> > > > >> > > > > > >
