Dear Wiki user,

You have subscribed to a wiki page or wiki category on "Hadoop Wiki" for change 
notification.

The following page has been changed by stack:
http://wiki.apache.org/hadoop/Hbase/FAQ

The comment on the change is:
Fixup on file limit spec.

------------------------------------------------------------------------------
  
  Currently Hbase is a file handle glutton.  Running an Hbase loaded w/ more 
than a few regions, its possible to blow past the common 1024 default file 
handle limit for the user running the process.  Running out of file handles is 
like an OOME, things start to fail in strange ways.  To up the users' file 
handles, edit '''/etc/security/limits.conf''' on all nodes and restart your 
cluster.
  
+ {{{# Each line describes a limit for a user in the form:
+ #
+ # domain    type    item    value
+ #
+ hbase     hard    nofile  32768}}}
+ 
  The math runs roughly as follows: Per column family, there is at least one 
mapfile and possibly up to 5 or 6 if a region is under load (lets say 3 per 
column family on average).  Multiply by the number of regions per region 
server.  So, for example, say you have a schema of 3 column familes per region 
and that you have 100 regions per regionserver, the JVM will open 3 * 3 * 100 
mapfiles -- 900 file descriptors not counting open jar files, conf files, etc 
(Run 'lsof -p REGIONSERVER_PID' to see for sure).
  
  '''7. [[Anchor(7)]] What can I do to improve hbase performance?'''

Reply via email to