hp proliant raid 10 with 4 sas. 15k. smartarray 6i. 2cpu/4core.
2010/9/22 Ryan Rawson <[email protected]>: > generally checksum errors are due to hardware faults of one kind or another. > > what is your hardware like? > > On Wed, Sep 22, 2010 at 2:08 AM, Andrey Stepachev <[email protected]> wrote: >> But why it is bad? Split/compaction? I make my own RetryResultIterator >> which reopen scanner on timeout. But what is best way to reopen scanner. >> Can you point me where i can find all this exceptions? Or may be >> here already some sort for recoveratble iterator? >> >> 2010/9/22 Ryan Rawson <[email protected]>: >>> ah ok i think i get it... basically at this point your scanner is bad >>> and iterating on it again won't work. the scanner should probably >>> self close itself so you get tons of additional exceptions but instead >>> we dont. >>> >>> there is probably a better fix for this, i'll ponder >>> >>> On Wed, Sep 22, 2010 at 1:57 AM, Ryan Rawson <[email protected]> wrote: >>>> very strange... looks like a bad block ended up in your scanner and >>>> subsequent nexts were failing due to that short read. >>>> >>>> did you have to kill the regionserver or did things recover and >>>> continue normally? >>>> >>>> -ryan >>>> >>>> On Wed, Sep 22, 2010 at 1:37 AM, Andrey Stepachev <[email protected]> wrote: >>>>> Hi All. >>>>> >>>>> I get org.apache.hadoop.fs.ChecksumException for a table on heavy >>>>> write in standalone mode. >>>>> table tmp.bsn.main created 2010-09-22 10:42:28,860 and then 5 threads >>>>> writes data to it. >>>>> At some moment exception thrown. >>>>> >>>>> Andrey. >>>>> >>>> >>> >> >
