Ko o, On Sep 19, 2015 2:12 PM, "Zhudacai" <[email protected]> wrote:
> Hi,ALL > > > > We get a problem about enable UseNUMA flag for my hadoop framework. > > > > We've tried to specify JVM flags during hadoop daemon's starts, > > e.g. export HADOOP_NAMENODE_OPTS="-XX:+UseNUMA > -Dcom.sun.management.jmxremote $HADOOP_NAMENODE_OPTS", > > export HADOOP_SECONDARYNAMENODE_OPTS="-XX:+UseNUMA > -Dcom.sun.management.jmxremote $HADOOP_SECONDARYNAMENODE_OPTS", etc. > > But the ratio between local and remote memory access is 2:1, just remains > as same as before. > > > > Then we find that hadoop MapReduce start child JVM processes to run task > in containers. So we passes -XX:+UseNUMA to JVMs by set theting > configuration parameter child.java.opts. But hadoop starts to throw > ExitCodeExceptionException (exitCode=1), seems that hadoop does not support > this JVM parameter. > > > > What should we do to enable UseNUMA flag for my hadoop? Or what should we > do to decrease the local/remote memory access in NUMA framework? Should we > just change Hadoop script or resorts to source code? And how to do it? > > > > The hadoop version is 2.6.0. > > > > Best Regards. > > > > Dacai >
