I don't know.  Can you dump the curl output to a file or STDOUT and
take a look at it?
St.Ack

On Tue, Sep 21, 2010 at 10:36 PM, Jack Levin <[email protected]> wrote:
> Thanks, patched version is now running (actually SU's version was
> patched, ty).  Another question is in regards to REST doing base64,
> when you send a header  "Accept: application/octet-stream", I get just
> byte stream, e.g. no base64, does this mean internally the cell is
> split our byte by byte without any conversion to base64? Hence no
> overhead?
>
> -Jack
>
> On Tue, Sep 21, 2010 at 10:09 PM, Stack <[email protected]> wrote:
>> Given the log snippet, I'd guess its because your hbase doesn't have 
>> HBASE-2643.
>>
>> The above makes it so we continue through an EOF exception when
>> splitting logs where before we'd fail the splitting, requeue, split,
>> then fail again.
>>
>> Here is comment recently added to our little hbase book at 
>> src/docbkx/book.xml:
>>
>>      <section>
>>        <title>How EOFExceptions are treated when splitting a crashed
>>        RegionServers' WALs</title>
>>
>>        <para>If we get an EOF while splitting logs, we proceed with the split
>>        even when <varname>hbase.hlog.split.skip.errors</varname> ==
>>        <constant>false</constant>. An EOF while reading the last log in the
>>        set of files to split is near-guaranteed since the RegionServer likely
>>        crashed mid-write of a record. But we'll continue even if we got an
>>        EOF reading other than the last file in the set.<footnote>
>>            <para>For background, see <link
>>            
>> xlink:href="https://issues.apache.org/jira/browse/HBASE-2643";>HBASE-2643
>>            Figure how to deal with eof splitting logs</link></para>
>>          </footnote></para>
>>      </section>
>>
>> St.Ack
>>
>> On Tue, Sep 21, 2010 at 3:00 PM, Jack Levin <[email protected]> wrote:
>>> First, I saw:
>>>
>>>
>>> 2010-09-21 11:30:05,122 DEBUG
>>> org.apache.hadoop.hbase.master.RegionServerOperationQueue: Put
>>> ProcessServerShutdown of 10.103.2.5,60020,1285042335711 back on queue
>>> 2010-09-21 11:30:05,122 DEBUG
>>> org.apache.hadoop.hbase.master.RegionServerOperationQueue: Processing
>>> todo: ProcessServerShutdown of 10.103.2.5,60020,1285042335711
>>> 2010-09-21 11:30:05,122 INFO
>>> org.apache.hadoop.hbase.master.RegionServerOperation: Process shutdown
>>> of server 10.103.2.5,60020,1285042335711: logSplit: false,
>>> rootRescanned: false, n
>>> umberOfMetaRegions: 1, onlineMetaRegions.size(): 0
>>>
>>> repeated rapidly for 20 mins or so.
>>>
>>> Then:
>>>
>>> Bunch of regions got unassigned:
>>>
>>>
>>> 2010-09-21 12:00:07,782 DEBUG
>>> org.apache.hadoop.hbase.master.RegionManager: Unassigning 66 regions
>>> from 10.103.2.3,60020,1285042333293
>>> 2010-09-21 12:00:07,782 DEBUG
>>> org.apache.hadoop.hbase.master.RegionManager: Going to close region
>>> img816,img2103r.jpg,1285003791610.1592893332
>>> 2010-09-21 12:00:07,782 DEBUG
>>> org.apache.hadoop.hbase.master.RegionManager: Going to close region
>>> img534,92166039.jpg,1284949117852.1009352950
>>> 2010-09-21 12:00:07,782 DEBUG
>>> org.apache.hadoop.hbase.master.RegionManager: Going to close region
>>> img36,abcwu.jpg,1285001278990.272235177
>>>
>>>
>>> Restarting master did not help.  Ultimately what brought the cluster
>>> back up, is full shutdown of regionservers, and masters, and then
>>> bring all up.
>>>
>>> Any ideas what might have happened here?
>>>
>>> We are running:
>>>
>>> HBase Version   0.89.20100726, r979826
>>> Hadoop Version  0.20.2+320, r9b72d268a0b590b4fd7d13aca17c1c453f8bc957
>>> Regions On FS   5057
>>>
>>> 3 zookeepers and 13 regionservers.
>>>
>>> -Jack
>>>
>>
>

Reply via email to