Elena, That is good news, indeed this was with 1.8.5-patch1.
Is code written with using independent IO structured significantly different than with collective IO? I would like to get moving with pHDF5 and as I am currently not too familiar with it, want to make sure that I am not going to have to do a rewrite after the collective code works. It does seem to all occur behind the scenes with the h5dwrite command, so I presume I am safe. Thanks, Leigh On Mon, Jan 17, 2011 at 4:59 PM, Elena Pourmal <[email protected]>wrote: > Leigh, > > I am writing to confirm that the bug you reported does exist in > 1.8.5-patch1, but is fixed in 1.8.6 (coming soon). > > Elena > On Jan 16, 2011, at 3:47 PM, Leigh Orf wrote: > > I managed to build pHDF5 on blueprint.ncsa.uiuc.edu (IBM AIX Power 6). I > compiled the hyperslab_by_chunk.f90 test program found at > http://www.hdfgroup.org/HDF5/Tutor/phypechk.html without error. When I run > it, however, I get the following output: > > ATTENTION: 0031-408 4 tasks allocated by LoadLeveler, continuing... > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > MPI_Type_free, task 0 > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > MPI_Type_free, task 1 > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > MPI_Type_free, task 2 > ERROR: 0032-110 Attempt to free a predefined datatype (2) in > MPI_Type_free, task 3 > HDF5: infinite loop closing library > > > D,S,T,D,S,F,D,G,S,T,F,AC,FD,P,FD,P,FD,P,E,E,SL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL,FL > HDF5: infinite loop closing library > > The line which causes the grief is: > > CALL h5dwrite_f(dset_id, H5T_NATIVE_INTEGER, data, dimsfi, error, & > file_space_id = filespace, mem_space_id = memspace, > xfer_prp = plist_id) > > If I replace that call with the one that is commented out in the program, > it runs without a problem. That line is: > > CALL h5dwrite_f(dset_id, H5T_NATIVE_INTEGER, data, dimsfi,error, & > file_space_id = filespace, mem_space_id = memspace) > > Any ideas? I definitely want to take advantage of doing collective I/O if > possible. > > Leigh > > -- > Leigh Orf > Associate Professor of Atmospheric Science > Department of Geology and Meteorology > Central Michigan University > Currently on sabbatical at the National Center for Atmospheric Research > in Boulder, CO > NCAR office phone: (303) 497-8200 > > _______________________________________________ > Hdf-forum is for HDF software users discussion. > [email protected] > http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org > > > > _______________________________________________ > Hdf-forum is for HDF software users discussion. > [email protected] > http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org > > -- Leigh Orf Associate Professor of Atmospheric Science Department of Geology and Meteorology Central Michigan University Currently on sabbatical at the National Center for Atmospheric Research in Boulder, CO NCAR office phone: (303) 497-8200
_______________________________________________ Hdf-forum is for HDF software users discussion. [email protected] http://mail.hdfgroup.org/mailman/listinfo/hdf-forum_hdfgroup.org
