Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change 
notification.

The "Hbase/FAQ" page has been changed by JeffHammerbacher.
http://wiki.apache.org/hadoop/Hbase/FAQ?action=diff&rev1=61&rev2=62

--------------------------------------------------

   1. [[#16|Can HBase development be done on windows?]]
   1. [[#17|Please explain HBase version numbering?]]
   1. [[#18|What version of Hadoop do I need to run HBase?]]
-  1. [[#10|Any other troubleshooting pointers for me?]]
+  1. [[#19|Any other troubleshooting pointers for me?]]
   1. [[#20|Are there any schema design examples?]]
   1. [[#21|How do I add/remove a node?]]
   1. [[#22|Why do servers have start codes?]]
+  1. [[#23|What is the maximum recommended cell size?]]
  
  == Answers ==
  
@@ -224, +225 @@

  
  If a region server crashes and recovers, it cannot be given work until its 
lease times out. If the lease is identified only by an IP address and port 
number, then that server can't do any progress until the lease times out. A 
start code is added so that the restarted server can begin doing work 
immediately upon recovery. For more, see 
https://issues.apache.org/jira/browse/HBASE-1156.
  
+ '''23. <<Anchor(23)>> What is the maximum recommended cell size?'''
+ 
+ A rough rule of thumb, with little empirical validation, is to keep the data 
in HDFS and store pointers to the data in HBase if you expect the cell size to 
be consistently above 10 MB. If you do expect large cell values and you still 
plan to use HBase for the storage of cell contents, you'll want to increase the 
block size and the maximum region size for the table to keep the index size 
reasonable and the split frequency acceptable.
+ 

Reply via email to