On Sun, Jan 16, 2011 at 07:39:48PM -0600, Elena Pourmal wrote: > Hi Leigh, > > It looks like that collective write fails but independent succeeds. Which > version did you build? We will try to reproduce it.
Hi Elena: Do you think this could be a regression of a bug I reported against 1.8.4 about a year ago? I sent a test case to the hdf5 help desk 2 December 2009 with the subject "bug in HDF5 type handling", though that was in the read path. Sorry, I don't have a ticket id or anything like that to help track it down. ==rob > Thank you! > > Elena > > > On Jan 16, 2011, at 3:47 PM, Leigh Orf wrote: > > > I managed to build pHDF5 on blueprint.ncsa.uiuc.edu (IBM AIX Power 6). I > > compiled the hyperslab_by_chunk.f90 test program found at > > http://www.hdfgroup.org/HDF5/Tutor/phypechk.html without error. When I run > > it, however, I get the following output: > > > > ATTENTION: 0031-408 4 tasks allocated by LoadLeveler, continuing... > > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > > MPI_Type_free, task 0 > > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > > MPI_Type_free, task 1 > > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > > MPI_Type_free, task 2 > > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > > MPI_Type_free, task 3 > > HDF5: infinite loop closing library > > > > D,S,T,D,S,F,D,G,S,T,F,AC,FD,P,FD,P,FD,P,E,E,SL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL > > HDF5: infinite loop closing library > > > > The line which causes the grief is: > > > > CALL h5dwrite_f(dset_id, H5T_NATIVE_INTEGER, data, dimsfi, error, & > > file_space_id = filespace, mem_space_id = memspace, > > xfer_prp = plist_id) > > > > If I replace that call with the one that is commented out in the program, > > it runs without a problem. That line is: > > > > CALL h5dwrite_f(dset_id, H5T_NATIVE_INTEGER, data, dimsfi,error, & > > file_space_id = filespace, mem_space_id = memspace) > > > > Any ideas? I definitely want to take advantage of doing collective I/O if > > possible. > > > > Leigh > > > > -- > > Leigh Orf > > Associate Professor of Atmospheric Science > > Department of Geology and Meteorology > > Central Michigan University > > Currently on sabbatical at the National Center for Atmospheric Research in > > Boulder, CO > > NCAR office phone: (303) 497-8200 > > > > _______________________________________________ > > Hdf-forum is for HDF software users discussion. > > [email protected] > > http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org > > _______________________________________________ > Hdf-forum is for HDF software users discussion. > [email protected] > http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org -- Rob Latham Mathematics and Computer Science Division Argonne National Lab, IL USA _______________________________________________ Hdf-forum is for HDF software users discussion. [email protected] http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org
