Hi Ulrik,
On 7/10/2013 12:08 PM, [email protected] wrote:
Hi Mohamad,
What I'm trying to achieve is a graceful error handling in case
'something' wrong happens. My parallel hdf5 writer application is the
back end of a system which at the front reads data off a detector. The
front end system will then pass on the data along with some metadata
to describe where the 2D frame sits in the full dataset. So each MPI
process sits on a separate server and each server is connected to one
piece of readout-electronic (front-end) which reads a 2D strip off the
full detector/camera.
Because errors just happens for various reasons in complex systems --
especially ones in development (and in this case we are talking about
several software and hardware subsystems working in supposedly
beautiful synchronisation) -- the writer must be able to recover even
from erroneous use -- for example if the frontend is sending an the
writer a bit of data with wrongly configured offsets, causing us to be
writing outside datasets (wrong offsets is just an example -- of
course I could make sanity checks for this particular case...)
I still think you are not making a distinction between programmable
errors and system/hardware errors.
But if you want to recover from the case that one process fails in a
collective write call, you will need to add fault tolerance yourself, by
checking the return status of every process, and communicating it to all
the ranks in the communicator.
The bottom line (or question) is really just: is there a way to
recover somewhat gracefully if an error has happened on one or more
nodes -- and what would be the consequences? (corrupt file?)
I do not know how to answer this question. There are many range of
errors, some of which can be recovered from; others well not so much.
Add to that there is still no fault tolerance in MPI, so recovering from
MPI failures is hard.
As for file corruption, again it depends on the error. If a failure
prevents processes to close the file and flush the metadata cache, then
yes you will end up with a possible corrupt file.
Perhaps I need to switch off the collective mode? Would that allow me
to close the file without having done an equal number of extend/write
on each node?
switching off collective mode means that dataset access operations
(H5Dread/H5Dwrite) can be done independently. Other operations are still
required to be collective. see:
http://www.hdfgroup.org/HDF5/doc/RM/CollectiveCalls.html
Thanks,
Mohamad
Cheers,
Ulrik
*From:*Hdf-forum [mailto:[email protected]] *On
Behalf Of *Mohamad Chaarawi
*Sent:* 10 July 2013 17:46
*To:* 'HDF Users Discussion List'
*Subject:* Re: [Hdf-forum] Bail out on parallel hdf5 write or extend error
Hi Ulrik,
*From:*Hdf-forum [mailto:[email protected]] *On
Behalf Of *[email protected]
<mailto:[email protected]>
*Sent:* Wednesday, July 10, 2013 11:01 AM
*To:* [email protected] <mailto:[email protected]>
*Subject:* [Hdf-forum] Bail out on parallel hdf5 write or extend error
Hello,
I am writing an application to stream data from multiple imaging
detectors, operating in a synchronised fashion into a single dataset
in one HDF5 file. The dataset is a chunked 3D dataset where 2D (X,Y)
images gets appended on the 3^rd dimension (dimensions are defined as:
[Z, Y, X] -- so as the 2D frames are received I extend dimension Z.
At this point I need to work out how to deal with errors -- if one
node for some reason does something wrong like trying to write outside
the dataset dimensions or whatever, I need to be able to close the
current file and return to the initial state, ready to create and
write to another file.
Hmm, I'm not quite sure I understand what you are trying to achieve
here. When you say that a node does something wrong like trying to
write outside the dataset dimensions, that implies an erroneous
program and should be corrected, and not try and recover from it. From
what I understand, you are attempting to use HDF5 erroneously, and
continue to do that expecting a certain behaviour. This is not possible.
I might have misunderstood you because I'm not aware of the full
details about your use case here, i.e. why would you write outside the
dataset dimensions.
For performance reasons I am using collective IO and so I think I need
the same number of extend, write, etc -- or the H5Fclose call will
hang. Is this correct?
The hang would most probably not happen in H5Fclose if you don't call
extend and write (if you set collective I/O) collectively. It will
happen in the extend or write itself, because a collective operation
expects all processes to be there at some point in time. If one
process does not call the operation and other processes attempt to
talk to that process, then your program will hang.
Can I use the H5Pset_fclose_degree to set H5F_CLOSE_STRONG safely in
the parallel hdf5 case or will that cause a crash/hang/corrupt file?
I do not think this is relevant to what you are asking/require. Sure
you can set to H5F_CLOSE_STRONG, but that does not mean you can avoid
calling collective operations on all processes, or use the API in an
erroneous manner.
Thanks,
Mohamad
Cheers,
Ulrik
---------------------------------------------------------------------
Ulrik Kofoed Pedersen
Senior Software Engineer
Diamond Light Source Ltd
Phone: 01235 77 8580
--
This e-mail and any attachments may contain confidential, copyright
and or privileged material, and are for the use of the intended
addressee only. If you are not the intended addressee or an authorised
recipient of the addressee please notify us of receipt by returning
the e-mail and do not use, copy, retain, distribute or disclose the
information in or attached to the e-mail.
Any opinions expressed within this e-mail are those of the individual
and not necessarily of Diamond Light Source Ltd.
Diamond Light Source Ltd. cannot guarantee that this e-mail or any
attachments are free from viruses and we cannot accept liability for
any damage which you may sustain as a result of software viruses which
may be transmitted in or with the message.
Diamond Light Source Limited (company no. 4375679). Registered in
England and Wales with its registered office at Diamond House, Harwell
Science and Innovation Campus, Didcot, Oxfordshire, OX11 0DE, United
Kingdom
--
This e-mail and any attachments may contain confidential, copyright
and or privileged material, and are for the use of the intended
addressee only. If you are not the intended addressee or an authorised
recipient of the addressee please notify us of receipt by returning
the e-mail and do not use, copy, retain, distribute or disclose the
information in or attached to the e-mail.
Any opinions expressed within this e-mail are those of the individual
and not necessarily of Diamond Light Source Ltd.
Diamond Light Source Ltd. cannot guarantee that this e-mail or any
attachments are free from viruses and we cannot accept liability for
any damage which you may sustain as a result of software viruses which
may be transmitted in or with the message.
Diamond Light Source Limited (company no. 4375679). Registered in
England and Wales with its registered office at Diamond House, Harwell
Science and Innovation Campus, Didcot, Oxfordshire, OX11 0DE, United
Kingdom
_______________________________________________
Hdf-forum is for HDF software users discussion.
[email protected]
http://mail.lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org
_______________________________________________
Hdf-forum is for HDF software users discussion.
[email protected]
http://mail.lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org