If you scan meta you need two commas in your start key. Meta is fine.
The format is like:
"table name,region start key,timestamp"

If you want to start at the first of a table say "table,,". Not that
tables cannot have comms in the name.

On Monday, April 12, 2010, Al Lias <al.l...@gmx.de> wrote:
> And this is what I get when looking for this region in the shell:
>
> hbase(main):002:0> scan '.META.', { STARTROW =>
> 'emailmd5,69678443e89be6e657549ef440186b' ,LIMIT => 2}
>
> Caused by: java.lang.IllegalArgumentException: No 44 in
> <3'emailmd5,69678443e89be6e657549ef440186��������>,
> length=30, offset=19
>         at
> org.apache.hadoop.hbase.KeyValue.getRequiredDelimiterInReverse(KeyValue.java:1210)
>         at
> org.apache.hadoop.hbase.KeyValue$MetaKeyComparator.compareRows(KeyValue.java:1663)
>         at
> org.apache.hadoop.hbase.KeyValue$KeyComparator.compare(KeyValue.java:1702)
>
> ...
>
> Does this indicate that a) I put a bad key somewhere or b) is the key
> not properly escaped somewhere or c) is the data corrupted otherwise?
>
> Thx
>
>         Al
>
>
>
> Am 10.04.2010 21:08, schrieb Al Lias:
>> Thanks looking into it, Todd,
>>
>> Am 09.04.2010 17:16, schrieb Todd Lipcon:
>>> Hi,
>>>
>>> This is likely a multiple assignment bug.
>>>
>>
>> I tried again, this time I grep'ed for the the region that a client
>> could not find. Locks like something with "mutliple assigment".
>>
>> http://pastebin.com/CHD0KSPH
>>
>>> Can you grep the NN log for the block ID 991235084167234271 ? This should
>>> tell you which file it was originally allocated to, as well as what IP wrote
>>> it. You should also see a deletion later. Also, the filename should give you
>>> a clue as to which region the block is from. You can then consult those
>>> particular RS and master logs to see which servers deleted the file and why.
>>>
>>
>> PLS help; http://pastebin.com/zUxqyyfU (not sorted by time)
>> I can only see that the Master adviced to delete....
>>
>> (This error is a different instance of the same problem than the one above)
>>
>> Thx,
>>
>>       Al
>>
>>> -Todd
>>>
>>> On Fri, Apr 9, 2010 at 12:56 AM, Al Lias <al.l...@gmx.de> wrote:
>>>
>>>> I repeatedly have the following problem with
>>>> 0.20.3/dfs.datanode.socket.write.timeout=0: Some RS is requested for
>>>> some data, the DFS can not find it, client hangs until timeout.
>>>>
>>>> Grepping the cluster logs, I can see this:
>>>>
>>>> 1. at some time the DFS is asked to delete a block, blocks are deleted
>>>> from the datanodes
>>>>
>>>> 2. some minutes later, a RS seems to ask for exactly this block...DFS
>>>> says "Block blk_.. is not valid." and then "No live nodes contain
>>>> current block".
>>>>
>>>> (I have xceivers and file desc limit high,
>>>> dfs.datanode.handler.count=10, No particulary high load, 17 Servers with
>>>> 24G/4Core)
>>>>
>>>> More log here: http://pastebin.com/cdqsy8Ae
>>>>
>>>> ?
>>>>
>>>> Thx, Al
>>>>
>>>>
>>>>
>>>>
>>>
>>>
>
>

Reply via email to