Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change 
notification.

The "Hbase/FAQ" page has been changed by JeffHammerbacher.
http://wiki.apache.org/hadoop/Hbase/FAQ?action=diff&rev1=62&rev2=63

--------------------------------------------------

   1. [[#21|How do I add/remove a node?]]
   1. [[#22|Why do servers have start codes?]]
   1. [[#23|What is the maximum recommended cell size?]]
+  1. [[#24|Why can't I iterate through the rows a table in reverse order?]]
  
  == Answers ==
  
@@ -229, +230 @@

  
  A rough rule of thumb, with little empirical validation, is to keep the data 
in HDFS and store pointers to the data in HBase if you expect the cell size to 
be consistently above 10 MB. If you do expect large cell values and you still 
plan to use HBase for the storage of cell contents, you'll want to increase the 
block size and the maximum region size for the table to keep the index size 
reasonable and the split frequency acceptable.
  
+ '''24. <<Anchor(24)>> Why can't I iterate through the rows a table in reverse 
order?'''
+ 
+ Because of the way 
[[http://hbase.apache.org/docs/current/api/org/apache/hadoop/hbase/io/hfile/HFile.html|HFile]]
 works: for efficiency, column values are put on disk with the length of the 
value written first and then the bytes of the actual value written second. To 
navigate through these values in reverse order, these length values would need 
to be stored twice (at the end as well) or in a side file. A robust secondary 
index implementation is the likely solution here to ensure the primary use case 
remains fast.
+ 

Reply via email to