[ https://issues.apache.org/jira/browse/HADOOP-6029?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12720484#action_12720484 ]
Jothi Padmanabhan commented on HADOOP-6029: ------------------------------------------- Here are a couple of links that explain -Xmx value and maxMemory() will be different. http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=4391499 http://bugs.sun.com/bugdatabase/view_bug.do?bug_id=4686462 A comment from the second link -- " ... freeMemory() and totalMemory() report the amount of memory _inside_ the jvm while maxMemory() reports on the amount of memory _outside_ the jvm, i.e. the amount the whole jvm uses as seen from the OS." > TestReduceFetch failed. > ----------------------- > > Key: HADOOP-6029 > URL: https://issues.apache.org/jira/browse/HADOOP-6029 > Project: Hadoop Core > Issue Type: Bug > Components: mapred > Reporter: Tsz Wo (Nicholas), SZE > Attachments: > FAILING-PARTIALMEM-TEST-org.apache.hadoop.mapred.TestReduceFetch.txt, > TEST-org.apache.hadoop.mapred.TestReduceFetch.txt, > TEST-org.apache.hadoop.mapred.TestReduceFetch.txt > > > {noformat} > Testcase: testReduceFromMem took 23.625 sec > FAILED > Non-zero read from local: 83 > junit.framework.AssertionFailedError: Non-zero read from local: 83 > at > org.apache.hadoop.mapred.TestReduceFetch.testReduceFromMem(TestReduceFetch.java:289) > at junit.extensions.TestDecorator.basicRun(TestDecorator.java:24) > at junit.extensions.TestSetup$1.protect(TestSetup.java:23) > at junit.extensions.TestSetup.run(TestSetup.java:27) > {noformat} > Ran TestReduceFetch a few times on a clean trunk. It failed consistently. -- This message is automatically generated by JIRA. - You can reply to this email to add a comment to the issue online.