It looks like the root cause of this is negative tests (i.e., teststhat are supposed to fail) are not actually quitting. They still keep running, even after their HNP is gone -- spinning endlessly, consuming CPU cycles.
Hence, the CPU load on a bunch of nodes in my cluster is in the hundreds (which then even ends up causing positive tests to fail). Let me dig into this a bit and see if I can get some stack traces, etc., and open up a github issue. > On Feb 18, 2016, at 12:15 PM, Howard Pritchard <hpprit...@gmail.com> wrote: > > Hi Folks, > > I noticed cisco-community MTT results are really red/pink today. > If I try to view some of the ibm test results though, something goes > south with mtt and this is what get's posted back to my browser: > > Fatal error: Allowed memory size of 67108864 bytes exhausted (tried to > allocate 71 bytes) in > /nfs/data/osl/www/mtt.open-mpi.org/reporter/dashboard.inc on line 271 > > So, I guess the first priority is do we know what's happened > with cisco MTT? > > Second, is this a known problem with the mtt reporter? > Is there a way to work around it? > > Thanks, > > Howard > > _______________________________________________ > devel mailing list > de...@open-mpi.org > Subscription: http://www.open-mpi.org/mailman/listinfo.cgi/devel > Link to this post: > http://www.open-mpi.org/community/lists/devel/2016/02/18609.php -- Jeff Squyres jsquy...@cisco.com For corporate legal information go to: http://www.cisco.com/web/about/doing_business/legal/cri/