On Wed, May/21/2008 06:46:06PM, Pavel Shamis (Pasha) wrote:
> I sent it directly to your email. Please check.
> Thanks,
> Pasha

Got it. Thanks. It's a PHP memory overload issue.
(Apparently I didn't look far back enough in the httpd
error_logs.) See below.

> 
> Ethan Mallove wrote:
> > On Wed, May/21/2008 05:19:44PM, Pavel Shamis (Pasha) wrote:
> >   
> >> Jeff Squyres wrote:
> >>     
> >>> Are we running into http max memory problems or http max upload size 
> >>> problems again?
> >>>       
> >> I guess it is some server side issue, you need to check the 
> >> /var/log/httpd/* log on the server.
> >>     
> >
> > The only thing I found in the httpd logs
> > (/var/log/httpd/www.open-mpi.org/error_log*) was this PHP
> > warning, which I don't think would result in lost results:
> >
> >  PHP Warning:  array_shift(): The argument should be an array in 
> > .../submit/index.php on line 1683
> >
> > I haven't received any emailed Postgres errors either. When
> > were these results submitted? I searched for "mellanox" over
> > the past four days. It seem the results aren't buried in
> > here, because I see no test run failures ...
> >
> >   http://www.open-mpi.org/mtt/index.php?do_redir=659
> >
> > I'm assuming you're running with two Reporter INI sections:
> > Textfile and MTTDatabase? Can you send some MTT client
> > --verbose/--debug output from the below runs?
> >
> > Thanks,
> > Ethan
> >
> >   
> >>> On May 21, 2008, at 5:28 AM, Pavel Shamis (Pasha) wrote:
> >>>
> >>>       
> >>>> Hi,
> >>>>
> >>>> Here is test result from my last mtt run:
> >>>> +-------------+----------------+------+------+----------+------+
> >>>> | Phase       | Section        | Pass | Fail | Time out | Skip |
> >>>> +-------------+----------------+------+------+----------+------+
> >>>> | MPI install | ompi/gcc       | 1    | 0    | 0        | 0    |
> >>>> | MPI install | ompi/intel-9.0 | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | trivial        | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | trivial        | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | intel-suite    | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | intel-suite    | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | imb            | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | imb            | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | presta         | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | presta         | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | osu_benchmarks | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | osu_benchmarks | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | netpipe        | 1    | 0    | 0        | 0    |
> >>>> | Test Build  | netpipe        | 1    | 0    | 0        | 0    |
> >>>> | Test Run    | trivial        | 64   | 0    | 0        | 0    |
> >>>> | Test Run    | trivial        | 64   | 0    | 0        | 0    |
> >>>> | Test Run    | intel-suite    | 3179 | 165  | 400      | 0    |
> >>>> | Test Run    | intel-suite    | 492  | 0    | 0        | 0    |
> >>>> +-------------+----------------+------+------+----------+------+
> >>>>
> >>>> In the OMPI MTT DB (http://www.open-mpi.org/mtt) I found the follow
> >>>> "test run" results:
> >>>> | Test Run    | trivial        | 64   | 0    | 0        | 0    |
> >>>> | Test Run    | trivial        | 64   | 0    | 0        | 0    |
> >>>> | Test Run    | intel-suite    | 492  | 0    | 0        | 0    |
> >>>>
> >>>> And I can not find this one:
> >>>> | Test Run    | intel-suite    | 3179 | 165  | 400      | 0    |


Some missing results are in mttdb_debug_file.16.txt (and
17.txt), which are the largest .txt files of the bunch. 8
variants isn't that much, but maybe it causes a problem when
there's lots of stderr/stdout? I'm surprised
submit/index.php barfs on files this size:

  $ ls -l
  ...
  -rw-r--r--    1 em162155 staff      956567 May 21 14:21 
mttdb_debug_file.16.inc.gz
  -rw-r--r--    1 em162155 staff     9603132 May 21 14:09 
mttdb_debug_file.16.txt
  ...

  $ client/mtt-submit -h www.open-mpi.org -f mttdb_debug_file.16.txt -z -u sun 
-p sun4sun -d
  LWP::UserAgent::new: ()
  LWP::UserAgent::proxy: http

  Filelist: $VAR1 = 'mttdb_debug_file.16.txt';
  LWP::MediaTypes::read_media_types: Reading media types from 
/ws/ompi-tools/lib/perl5/5.8.8/LWP/media.types
  LWP::MediaTypes::read_media_types: Reading media types from 
/usr/perl5/site_perl/5.8.4/LWP/media.types
  LWP::MediaTypes::read_media_types: Reading media types from 
/home/em162155/.mime.types
  LWP::UserAgent::request: ()
  LWP::UserAgent::send_request: POST 
http://www.open-mpi.org/mtt/submit/index.php
  LWP::UserAgent::_need_proxy: Not proxied
  LWP::Protocol::http::request: ()
  LWP::UserAgent::request: Simple response: OK

  $ tail -f /var/log/httpd/www.open-mpi.org/error_log | grep -w submit
  ...
  [client 192.18.128.5] PHP Fatal error:  Allowed memory size of 33554432 bytes 
exhausted (tried to allocate 14 bytes) in 
/nfs/rontok/xraid/data/osl/www/www.open-mpi.org/mtt/submit/index.php on line 
1559
  ...

We'll have to somehow be more efficient on these loops.
E.g., line 1559:

   foreach (array_keys($_POST) as $k) {

Maybe if we broke $_POST up into multiple parts (e.g.,
$_POST_1, $_POST_2, ... $_POST_N)? Maybe we could do
something more efficient than array_keys here? I'm not sure.

The only workaround on the client side would be to break up
the runs. Maybe do a single MPI Install at a time? Do
ompi/gcc then ompi/intel-9.0 as seperate invocations of the
MTT client.

Sorry :-(

-Ethan


> >>>>
> >>>> From the log I see that all tests results were submitted successfully.
> >>>>
> >>>> Can you please check ?
> >>>>
> >>>> Thanks,
> >>>>
> >>>> Pasha
> >>>> _______________________________________________
> >>>> mtt-users mailing list
> >>>> mtt-us...@open-mpi.org
> >>>> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-users
> >>>>         
> >>>       
> >> _______________________________________________
> >> mtt-users mailing list
> >> mtt-us...@open-mpi.org
> >> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-users
> >>     
> >
> >   
> 
> _______________________________________________
> mtt-users mailing list
> mtt-us...@open-mpi.org
> http://www.open-mpi.org/mailman/listinfo.cgi/mtt-users

Reply via email to