Gilles,

Thank you for your information.

We'll use GNU compiler and may use Fujitsu compiler too.
In both case, we'll compile on SPARC at the first stage
because cross-compiling is messy a little. Cross-compiling
Open MPI requires a file for --with-cross configure option
and cross-compiling MPI programs needs both opal_wrapper
on x86-64 and libraries on SPARC. (or prepare special
wrapper-data.txt to compile MPI programs on SPARC)

Thanks,
Takahiro Kawashima,
MPI development team,
Fujitsu

> Dear Kawashima-san,
> 
> 
> the mimimum test on each PR is to run configure/make
> 
> if you are not cross compiling, you can also run make check
> 
> a bit more sophisticated test is to run make distcheck
> 
> ideally, if you have several nodes available, you can run some mpi tests 
> in a custom script
> 
> (see 
> https://github.com/mellanox-hpc/jenkins_scripts/blob/master/jenkins/ompi/ompi_jenkins.sh)
> 
> 
> out of curiosity, do you use Fujitsu or GNU compilers ? if Fujitsu 
> compilers, do you compile on Sparc or cross compile on x86_64 ?
> 
> 
> Cheers,
> 
> 
> Gilles
> 
> 
> On 5/19/2016 10:05 AM, Kawashima, Takahiro wrote:
> > Jeff,
> >
> > I hope we do in the future. But currently we don't have enough
> > machine time and direct Internet connectivity (especially from
> > testing machines).
> >
> > What type of test do you expect? Building Open MPI binary and
> > running some short test programs on some x86-64 machines are
> > realistic if the connectivity problem is resolved.
> > But running many or long test programs on many SPARC machines
> > per GitHub pull request is not realistic for us.
> > (daily or weekly run may be realistic)
> >
> >> Great!
> >>
> >> Will you be able to also do some continuous integration type testing 
> >> (i.e., run some basic tests for each Github pull request)?  Josh/IBM is 
> >> going to post some information about their Jenkins/Github pull request 
> >> setup shortly.
> >>
> >>
> >>> On May 18, 2016, at 9:50 AM, Kawashima, Takahiro 
> >>> <t-kawash...@jp.fujitsu.com> wrote:
> >>>
> >>> Jeff,
> >>>
> >>> Thank you. It's very useful information.
> >>> I'll plan our run based on your information.
> >>>
> >>> Once we (Fujitsu) come to be able to run the test suites regularly,
> >>> we'll prepare to upload the reports to the server and push our test 
> >>> suites.
> >>>
> >>> Thanks,
> >>> Takahiro Kawashima,
> >>> MPI development team,
> >>> Fujitsu
> >>>
> >>>>> Fujitsu started to try MTT + ompi-tests on our machines.
> >>>>> With the sample .ini file, we wrote our .ini file and some
> >>>>> test suites are run.
> >>>>>
> >>>>> I have two questions.
> >>>>>
> >>>>> (a) There are many test suites (directories) in ompi-tests.
> >>>>>    ibm, onesided, sun, ...
> >>>>>    Which test suites should I use to participate in
> >>>>>    OMPI MTT daily/weekly run?
> >>>> The general guidance is: run as many tests as you have resources for.  
> >>>> Meaning: we'll take any testing you can give.  :-)
> >>>>
> >>>> Have a look in ompi-tests:cisco/mtt/community/*.ini and 
> >>>> cisco/mtt/usnic/*.ini.  Those are the ini files I use every night for 
> >>>> Cisco usNIC-specific testing and community-wide testing. You can see the 
> >>>> results of them in the MTT community reporter:
> >>>>
> >>>>     http://mtt.open-mpi.org/
> >>>>
> >>>> I generally aim for about 20-24 hours of testing.  It's a little fuzzy, 
> >>>> because Cisco's MTT will only fire for a given version (I'm currently 
> >>>> testing the master, v1.10, and v2.x branches) if there were new commits 
> >>>> that day (i.e., if there's a new nightly snapshot tarball since the last 
> >>>> run).
> >>>>
> >>>> If you run too many tests such that your testing is more than 24 hours, 
> >>>> then your resources quickly get behind and you're testing tarballs from 
> >>>> days ago -- and that's not too useful.
> >>>>
> >>>>> (b) What is the recommended `np` value (number of processes)?
> >>>>>    Should I use the largest `np` I can run?
> >>>> Yes, subject to what I mentioned above: you want to aim for a total of 
> >>>> ~24 hours of testing so that you can start the next cycle with the next 
> >>>> night's snapshot tarball.
> >>>>
> >>>> You can pack this in however you want -- do lots of small-np-value tests 
> >>>> and a few large-np-value tests (just to sanity test large np values, 
> >>>> etc.), etc.
> >>>>
> >>>> You can also take into account that little development is done on the 
> >>>> weekends.  For example, you might want to aim for ~24 hours of testing 
> >>>> on Monday-Thursday evenings, and then aim for a 3-day run on Friday 
> >>>> evening (because there might not be new tarballs generated over the 
> >>>> weekend).
> >>>>
> >>>>>    Does it depend on test suites?
> >>>> Yes.  Some test suites have upper-bounds on the number of processes they 
> >>>> can run.  IIRC, the Intel test suite, for example, can only run up to 64 
> >>>> processes (because of some hard-coded array sizes) unless you use a 
> >>>> specific -D to compile it (that increases the size of these arrays).

Reply via email to