Hi,

I store a lot of columns for one row key and if the size become to big the
relevant Region Server crashs if I try to get or scan the row. For example
if I try to get the relevant row I got this error:

2013-09-11 12:46:43,696 WARN org.apache.hadoop.ipc.HBaseServer:
(operationTooLarge): {"processingtimems":3091,"client":"192.168.0.34:52488
","ti$

If I try to load the relevant row via Apache Pig and the HBaseStorage
Loader (use the scan operation) I got this message and after that the
Region Servers crashs:

2013-09-11 10:30:23,542 WARN org.apache.hadoop.ipc.HBaseServer:
(responseTooLarge): {"processingtimems":1851,"call":"next(-588368116791418695,
1), rpc version=1, client version=29,$

I'm using Cloudera 4.4.0 with 0.94.6-cdh4.4.0

Any clues?

regards

Reply via email to