This looks like the result if the hdf5 library and tool is not all using shared libraries.
Allen ________________________________________ From: Hdf-forum <hdf-forum-boun...@lists.hdfgroup.org> on behalf of Francesc Alted <fal...@gmail.com> Sent: Friday, July 7, 2017 10:58:41 AM To: HDF Users Discussion List Subject: Re: [Hdf-forum] Dynamic filters on Windows Hi Werner, Yeah, the different in the number of parameters is curious. And no, all the hdf5-blosc plugin versions take 7 params (4 reserved, hence the 4 0s) and 3 for the actual compression parameters (complevel, shuffle, codec), so I don't quite understand why 8 (with 5 0s) works for me, but it is indeed the required number on Unix boxes (using HDF5 1.8.x there, so maybe the difference is just because HDF5 1.8.x and 1.10.x). I have just tried with 7 params on Windows, but I am having the same problem again: λ h5repack -v -f UD=32001,7,0,0,0,0,4,2,1 carray1-zlib.h5 carray1-repack.h5 Objects to modify layout are... Objects to apply filter are... User Defined 32001 Making file <carray1-repack.h5>... ----------------------------------------- Type Filter (Compression) Name ----------------------------------------- group / HDF5-DIAG: Error detected in HDF5 (1.10.1) thread 0: #000: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Pocpl.c line 1037 in H5Pget_filter_by_id2(): can't find object for ID major: Object atom minor: Unable to find atom information (already closed?) #001: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Pint.c line 3815 in H5P_object_verify(): property list is not a member of the class major: Property lists minor: Unable to register new atom #002: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Pint.c line 3765 in H5P_isa_class(): not a property list major: Invalid arguments to routine minor: Inappropriate type warning: could not create dataset </carray>. Applying original settings dset /carray <warning: could not apply the filter to /carray> Anyway, for the record, I ended using ptrepack for converting HDF5 files into Blosc-compressed ones: λ ptrepack --complib blosc:lz4 --complevel=5 --shuffle 1 carray1-zlib.h5 carray1-repack.h5 -o λ ptdump carray1-repack.h5 / (RootGroup) '' /carray (CArray(200, 300), shuffle, blosc:lz4(5)) '' Anyways, thanks for all the help, Francesc 2017-07-07 16:18 GMT+02:00 Werner Benger <wer...@cct.lsu.edu<mailto:wer...@cct.lsu.edu>>: Hi Francesc, I got it to run with 1.10.1 and GCC 7.1 under MSYS2, kind of, but the 8-parameter makes it to hang forever, it only works with giving 7 parameters on h5repack for me: /bin/arch-x86_64-w64-mingw32-OptDeb/h5repack -v -f UD=32001,7,0,0,0,0,4,2,1 myRGBA.f5 myRGBA-repack.f5 Objects to modify layout are... Objects to apply filter are... User Defined 32001 Making file <myRGBA-repack.f5>... ----------------------------------------- Type Filter (Compression) Name ... dset UD (1.582:1) /t=000000000.0000000000/RGBA/Points/StandardCartesianChart3D/Height/Fragment[0x0x0] ... The file looks all fine: ./bin/arch-x86_64-w64-mingw32-OptDeb/h5ls.exe -rv myRGBA-repack.f5/t=000000000.0000000000/RGBA/Points/StandardCartesianChart3D/ReturnNumber/Fragment[0x0x0] Opened "myRGBA-repack.f5" with sec2 driver. t=000000000.0000000000/RGBA/Points/StandardCartesianChart3D/ReturnNumber/Fragment[0x0x0] Dataset {1/1, 256/256, 256/256} Data: Contiguous Location: 1:4283257 Links: 1 Chunks: {1, 256, 256} 65536 bytes Storage: 65536 logical bytes, 21807 allocated bytes, 300.53% utilization Filter-0: blosc-32001 {2, 2, 1, 65536, 4, 2, 1} Type: native unsigned char whereas the setting with 8 parameters hangs forever: ./bin/arch-x86_64-w64-mingw32-OptDeb/h5repack -v -f UD=32001,8,0,0,0,0,0,4,2,1 myRGBA.f5 myRGBA-repack.f5 If you want to try my configuration, try this in MSYS2: svn co https://subversion.assembla.com/svn/FiberHDF5/hdf5; cd hdf5; mingw32-make -j12 Maybe I am using an older version of blosc that doesn't like 8 parameters? Werner On 04.07.2017 15:50, Werner Benger wrote: Hi, I have not tried HDF 1.10.1 yet with Mingw, but wanted to give it a try in the next days anyway and can let you know how it goes. I'm not using cmake, but my own makefiles, I think you have access to it anyway, so for now with the 1.8.17 version. I can let you know once I got to check with 1.10.1 . Werner On 04.07.2017 14:35, Francesc Alted wrote: Hi, Today I spent some more time on this issue, and I recompiled everything (the Blosc library and its plugin) against HDF5 1.10.1 (the binaries from hdfgroup.org<http://hdfgroup.org>) in 64 bit mode. Again, I can read Blosc-compressed HDF5 files with h5ls and h5dump (so I am pretty sure that the plugin is correctly compiled), but h5repack is still failing: """ C:\tmp λ h5repack -v -f UD=32001,8,0,0,0,0,0,9,1,2 carray1.h5 carray1-repack.h5 Objects to modify layout are... Objects to apply filter are... User Defined 32001 Making file <carray1-repack.h5>... ----------------------------------------- Type Filter (Compression) Name ----------------------------------------- group / HDF5-DIAG: Error detected in HDF5 (1.10.1) thread 0: #000: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Pocpl.c line 1037 in H5Pget_filter_by_id2(): can't find object for ID major: Object atom minor: Unable to find atom information (already closed?) #001: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Pint.c line 3815 in H5P_object_verify(): property list is not a member of the class major: Property lists minor: Unable to register new atom #002: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Pint.c line 3765 in H5P_isa_class(): not a propert y list major: Invalid arguments to routine minor: Inappropriate type warning: could not create dataset </carray>. Applying original settings dset /carray <warning: could not apply the filter to /carray> """ At this point, I don't know what to do. I am using VS2017 Community Edition; could it be possible that I need VS2015 for compiling the plugin? Also, I have tried to use MingW64 to compile the plugin, but cmake has difficulties in finding the HDF5 installation: """ faltet@FRANCESCALT476E MINGW64 /c/Users/faltet/hdf5-blosc/build $ HDF5_ROOT="/c/Program Files/HDF_Group/HDF5/1.10.1" cmake .. System is unknown to cmake, create: Platform/MINGW64_NT-10.0 to use this system, please send your config file to cma k...@www.cmake.org<mailto:k...@www.cmake.org> so it can be added to cmake Your CMakeCache.txt file was copied to CopyOfCMakeCache.txt. Please send that fi le to cm...@www.cmake.org<mailto:cm...@www.cmake.org>. BLOSC_PREFIX='/c/Users/faltet/hdf5-blosc/build/blosc' BLOSC_INSTALL_DIR='/c/Users/faltet/hdf5-blosc/build/blosc' BLOSC_CMAKE_ARGS='-DCMAKE_INSTALL_PREFIX=/c/Users/faltet/hdf5-blosc/build/blosc' GIT_EXECUTABLE='/usr/bin/git.exe' CMake Error at /usr/share/cmake-3.6.2/Modules/FindPackageHandleStandardArgs.cmak e:148 (message): Could NOT find HDF5 (missing: HDF5_LIBRARIES HDF5_HL_LIBRARIES) (found version "1.10.1") Call Stack (most recent call first): /usr/share/cmake-3.6.2/Modules/FindPackageHandleStandardArgs.cmake:388 (_FPHSA _FAILURE_MESSAGE) /usr/share/cmake-3.6.2/Modules/FindHDF5.cmake:820 (find_package_handle_standar d_args) CMakeLists.txt:48 (find_package) -- Configuring incomplete, errors occurred! See also "/c/Users/faltet/hdf5-blosc/build/CMakeFiles/CMakeOutput.log". """ Thanks! 2017-06-28 22:25 GMT+02:00 Werner Benger <wer...@cct.lsu.edu<mailto:wer...@cct.lsu.edu>>: Francesc, I am getting this message, which is correct since I don't have it compiled with zstd. h5repack -f UD=32001,8,0,0,0,0,0,9,1,5 myRGBA.f5 myRGBA-out.f5 HDF5-DIAG: Error detected in HDF5 (1.8.17) thread 10592: #000: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5D.c line 426 in H5Dclose(): can't decrement count on dataset ID major: Dataset minor: Unable to decrement reference count #001: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5I.c line 1549 in H5I_dec_app_ref_always_close(): can't decrement ID ref count major: Object atom minor: Unable to decrement reference count #002: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5I.c line 1491 in H5I_dec_app_ref(): can't decrement ID ref count major: Object atom minor: Unable to decrement reference count #003: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dint.c line 1544 in H5D_close(): unable to destroy chunk cache major: Dataset minor: Unable to release object #004: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dchunk.c line 5164 in H5D__chunk_dest(): unable to flush one or more raw data chunks major: Low-level I/O minor: Unable to flush data from cache #005: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dchunk.c line 2641 in H5D__chunk_cache_evict(): cannot flush indexed storage buffer major: Low-level I/O minor: Write failed #006: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dchunk.c line 2528 in H5D__chunk_flush_entry(): output pipeline failed major: Data filters minor: Filter operation failed #007: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Z.c line 1412 in H5Z_pipeline(): filter returned failure major: Data filters minor: Write failed #008: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/h5zblosc/blosc_filter.c line 190 in blosc_filter(): this Blosc library does not have support for the 'zstd' compressor, but only for: blosclz,lz4,lz4hc,zlib major: Data filters minor: Callback failed #009: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dint.c line 1507 in H5D_close(): unable to flush cached dataset info major: Dataset minor: Write failed #010: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dint.c line 2438 in H5D__flush_real(): unable to flush raw data major: Dataset minor: Unable to flush data from cache #011: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dchunk.c line 2112 in H5D__chunk_flush(): unable to flush one or more raw data chunks major: Dataset minor: Unable to flush data from cache #012: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Dchunk.c line 2528 in H5D__chunk_flush_entry(): output pipeline failed major: Data filters minor: Filter operation failed #013: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/hdf5-1.8.17/H5Z.c line 1412 in H5Z_pipeline(): filter returned failure major: Data filters minor: Write failed #014: C:/msys64/home/Wrk/vish/fish/fiber/FiberHDF5/h5zblosc/blosc_filter.c line 190 in blosc_filter(): this Blosc library does not have support for the 'zstd' compressor, but only for: blosclz,lz4,lz4hc,zlib major: Data filters minor: Callback failed h5repack error: <myRGBA.f5>: Could not copy data to: myRGBA-out.f5 But when using LZ4HC as filter it works fine: Wrk $ h5repack -f UD=32001,8,0,0,0,0,0,9,1,2 myRGBA.f5 myRGBA-out.f5 Wrk $ ll -rw-r--r-- 1 Wrk None 12M Jan 25 16:43 myRGBA.f5 -rw-r--r-- 1 Wrk None 5.0M Jun 28 22:18 myRGBA-out.f5 I'm using 1.8.17, evidently. Btw, the restriction with the lib-prefix has the advantage that the HDF5 plugin loader does not look at DLL's such as "blosc.dll", it would not even try to load it as a plugin. Just something better than "lib" would probably be preferable, and independent of platform and compiler. I guess in your version it even tries to load the blosc.dll itself as a plugin - which could lead to issues if such a plugin would contain some startup code like global c++ constructors or code in DllMain which does "something". That's unlikely to be related to your problem though, I don't see from the output what the problem could be, it works for me. Werner On 28.06.2017 19:01, Francesc Alted wrote: Werner, I think the lib- prefix does not apply to my case because I am using VS2015 for compiling the libraries. But agreed that HDF5 taking different conventions depending on whether a MSVC or GCC-based is a bit confusing. BTW, although the h5dump can read blosc files with the plugin, ptrepack cannot create them on Windows: C:\Users\faltet>h5repack -f UD=32001,8,0,0,0,0,0,9,1,5 carray1-blosc.h5 carray1-repack.h5 HDF5-DIAG: Error detected in HDF5 (1.8.19) thread 0: #000: C:\Users\buildbot\Buildbot\hdf518-StdRelease-code-vs14\build\hdfsrc\src\H5Pocpl.c line 1100 in H5Pget_filter_by_id2(): can't find object for ID major: Object atom minor: Unable to find atom information (already closed?) #001: C:\Users\buildbot\Buildbot\hdf518-StdRelease-code-vs14\build\hdfsrc\src\H5Pint.c line 3379 in H5P_object_verify(): property list is not a member of the class major: Property lists minor: Unable to register new atom #002: C:\Users\buildbot\Buildbot\hdf518-StdRelease-code-vs14\build\hdfsrc\src\H5Pint.c line 3329 in H5P_isa_class(): not a property list major: Invalid arguments to routine minor: Inappropriate type HDF5-DIAG: Error detected in HDF5 (1.8.19) thread 0: #000: C:\Users\buildbot\Buildbot\hdf518-StdRelease-code-vs14\build\hdfsrc\src\H5Pocpl.c line 1100 in H5Pget_filter_by_id2(): can't find object for ID major: Object atom minor: Unable to find atom information (already closed?) #001: C:\Users\buildbot\Buildbot\hdf518-StdRelease-code-vs14\build\hdfsrc\src\H5Pint.c line 3379 in H5P_object_verify(): property list is not a member of the class major: Property lists minor: Unable to register new atom #002: C:\Users\buildbot\Buildbot\hdf518-StdRelease-code-vs14\build\hdfsrc\src\H5Pint.c line 3329 in H5P_isa_class(): not a property list major: Invalid arguments to routine minor: Inappropriate type h5repack error: <carray1-blosc.h5>: Could not copy data to: carray1-repack.h5 I have tested with both 1.10.1 and 1.8.19 (above). whereas it works perfectly well on a Mac box: Francescs-MacBook-Pro:PyTables faltet$ h5repack -f UD=32001,8,0,0,0,0,0,9,1,5 carray1-blosc.h5 carray1-repack.h5 Francescs-MacBook-Pro:PyTables faltet$ h5ls -v carray1-repack.h5 Opened "carray1-repack.h5" with sec2 driver. carray Dataset {200/218, 300/300} Location: 1:800 Links: 1 Chunks: {218, 300} 65400 bytes Storage: 60000 logical bytes, 58 allocated bytes, 103448.28% utilization Filter-0: blosc-32001 {2, 2, 1, 65400, 9, 1, 5, 5} Type: native unsigned char Any hints? 2017-06-28 15:15 GMT+02:00 Werner Benger <wer...@cct.lsu.edu<mailto:wer...@cct.lsu.edu>>: Hi Francesc, it took me a while to figure out this detail, too, I had the blosc.dll in the PATH at first, and it seems to work under some circumstances but not others... not entirely clear to me when it needs to be in the PATH or in the calling's DLL's directory outside of the PATH. There's some code in H5PL.c that checks for a lib prefix: #ifndef __CYGWIN__ if(!HDstrncmp(dp->d_name, "lib", (size_t)3) && (HDstrstr(dp->d_name, ".so") || HDstrstr(dp->d_name, ".dylib"))) { #else ... #endif So I guess in your case that prefix check did not apply, whereas for me using gcc under windows it did. Thus it seems best for compatibility to keep the plugins named with lib- starting under Windows, once compiled that C code should not matter whether it was produced by GCC or MSC. Otherwise yes, the name does not matter indeed. Werner On 28.06.2017 15:00, Francesc Alted wrote: Hi Werner, Right, what I was missing was basically moving the blosc.dll library into the HDF5 plugin directory. I thought that putting blosc.dll in a directory in the %PATH% was going to be enough, but apparently this is not the case (I suppose this is a Windows trickery). For what is worth, the name of the DLL plugin is not that important as I have tried with libHDF5blosc.dll, libH5Zblosc.dll, H5Zblosc.dll and even blosc_plugin.dll and all of them work. Apparently the HDF5 library tries all the DLLs in the plugin directory and uses the one that registers the necessary filter ID; pretty smart. Thanks so much! 2017-06-28 14:31 GMT+02:00 Werner Benger <wer...@cct.lsu.edu<mailto:wer...@cct.lsu.edu>>: Just to add: under windows, the plugins need to start with lib as prefix. It's technically not required, but the HDF5 plugin loader expects that. So the liblz4.dll in my plugins dir is the LZ4 plugin,the libHDF5blosc.dll the blosc-filter; the blosc.dll is used by the blosc filter and needs to be in the same directory if compiled as dynamic library as well. Werner On 28.06.2017 14:24, Werner Benger wrote: Hi, I am using HDF5 with blosc dynamic plugin loading regularly under Windows, and this are the files as needed in the plugin directory that HDF5 looks for: HDF5Plugins $ ll total 1.2M -rwxr-xr-x 1 Wrk None 382K Jun 26 23:30 blosc.dll* -rwxr-xr-x 1 Wrk None 330K Jun 26 23:30 libHDF5blosc.dll* -rwxr-xr-x 1 Wrk None 410K Jun 26 23:30 liblz4.dll* The blosc.dll is the blosc library, libHDF5blosc.dll is the HDF5 plugin using that library. I'm using Mingw64, which is gcc 6.3 under MSYS2 . Werner On 28.06.2017 13:44, Francesc Alted wrote: Hi, I have been trying to give support to HDF5 on Windows for dynamically loading the Blosc plugin (https://github.com/Blosc/hdf5-blosc), but no success so far. I basically have compiled the plugin with: > cl /I"C:\Program Files (x86)\blosc\include" > /I"C:\HDF_Group\HDF5\1.8.12\include" libH5Zblosc.c blosc_filter.c > /OUT:libH5Zblosc.dll /LD /link /LIBPATH:"C:\Program Files (x86)\blosc\lib" > /LIBPATH:"C:\HDF_Group\HDF5\1.8.12\lib" blosc.lib hdf5.lib and then copied the libH5Zblosc.dll to "%ALLUSERSPROFILE%/hdf5/lib/plugin", but when I try to use a h5dump for getting the data of an HDF5 with some datasets compressed with Blosc, I am getting: >h5dump \tmp\carray1-blosc.h5 HDF5 "\tmp\carray1-blosc.h5" { GROUP "/" { DATASET "carray" { DATATYPE H5T_STD_U8LE DATASPACE SIMPLE { ( 200, 300 ) / ( 218, 300 ) } DATA {h5dump error: unable to print data } } } } When asking for more errors, I am getting a more informative message: >h5dump --enable-error-stack \tmp\carray1-blosc.h5 HDF5 "\tmp\carray1-blosc.h5" { GROUP "/" { DATASET "carray" { DATATYPE H5T_STD_U8LE DATASPACE SIMPLE { ( 200, 300 ) / ( 218, 300 ) } DATA {HDF5-DIAG: Error detected in HDF5 (1.10.1) thread 0: #000: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Dio.c line 171 in H5Dread(): can't read data major: Dataset minor: Read failed #001: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Dio.c line 544 in H5D__read(): can't read data major: Dataset minor: Read failed #002: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Dchunk.c line 2050 in H5D__chunk_read(): unable to read raw data chunk major: Low-level I/O minor: Read failed #003: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Dchunk.c line 3405 in H5D__chunk_lock(): data pipeline read failed major: Data filters minor: Filter operation failed #004: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5Z.c line 1349 in H5Z_pipeline(): required filter 'blosc' is not registered major: Data filters minor: Read failed #005: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5PL.c line 389 in H5PL_load(): search in paths failed major: Plugin for dynamically loaded library minor: Can't get value #006: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\src\H5PL.c line 812 in H5PL__find(): can't open directory major: Plugin for dynamically loaded library minor: Can't open directory or file h5dump error: unable to print data } } } } H5tools-DIAG: Error detected in HDF5:tools (1.10.1) thread 0: #000: C:\autotest\hdf5110-StdRelease-code-10vs14\build\hdfsrc\tools\lib\h5tools_dump.c line 1639 in h5tools_dump_simple_dset(): H5Dread failed major: Failure in tools library minor: error in function I suppose the meaningful part is here: 10vs14\build\hdfsrc\src\H5PL.c line 812 in H5PL__find(): can't open directory b ut I have no clues on 1) what directory the dynamic plugin machinery in HDF5 is looking at and 2) which name the DLL should have (as per manual, I am using libH5Zblosc.dll, but I have tried with H5Zblosc.dll and blosc.dll with no success). I have also tried to use the HDF5_PLUGIN_PATH environment variable, so as to direct the plugin machinery in HDF5 to look into the specific %ALLUSERSPROFILE%/hdf5/lib/plugin place, but no luck. I tried with HDF5 1.8.12 and 1.10.1, with same results. Finally, if it is of any help, I am using Windows 10 64 bit. As a suggestion, it may help to debug situations like this if the HDF5 backtrace of the plugin machinery would provide info about 1) which directory it is looking at for finding the HDF5 plugin and 2) the name of the DLL that it is trying to load. Thanks in advance, -- Francesc Alted _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- ___________________________________________________________________________ Dr. Werner Benger Visualization Research Center for Computation & Technology at Louisiana State University (CCT/LSU) 2019 Digital Media Center, Baton Rouge, Louisiana 70803 Tel.: +1 225 578 4809<tel:%28225%29%20578-4809> Fax.: +1 225 578-5362<tel:%28225%29%20578-5362> _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- ___________________________________________________________________________ Dr. Werner Benger Visualization Research Center for Computation & Technology at Louisiana State University (CCT/LSU) 2019 Digital Media Center, Baton Rouge, Louisiana 70803 Tel.: +1 225 578 4809<tel:%28225%29%20578-4809> Fax.: +1 225 578-5362<tel:%28225%29%20578-5362> _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- Francesc Alted _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- ___________________________________________________________________________ Dr. Werner Benger Visualization Research Center for Computation & Technology at Louisiana State University (CCT/LSU) 2019 Digital Media Center, Baton Rouge, Louisiana 70803 Tel.: +1 225 578 4809<tel:%28225%29%20578-4809> Fax.: +1 225 578-5362<tel:%28225%29%20578-5362> _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- Francesc Alted _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- ___________________________________________________________________________ Dr. Werner Benger Visualization Research Center for Computation & Technology at Louisiana State University (CCT/LSU) 2019 Digital Media Center, Baton Rouge, Louisiana 70803 Tel.: +1 225 578 4809<tel:%28225%29%20578-4809> Fax.: +1 225 578-5362<tel:%28225%29%20578-5362> _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- Francesc Alted _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- ___________________________________________________________________________ Dr. Werner Benger Visualization Research Center for Computation & Technology at Louisiana State University (CCT/LSU) 2019 Digital Media Center, Baton Rouge, Louisiana 70803 Tel.: +1 225 578 4809<tel:(225)%20578-4809> Fax.: +1 225 578-5362<tel:(225)%20578-5362> _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- ___________________________________________________________________________ Dr. Werner Benger Visualization Research Center for Computation & Technology at Louisiana State University (CCT/LSU) 2019 Digital Media Center, Baton Rouge, Louisiana 70803 Tel.: +1 225 578 4809<tel:(225)%20578-4809> Fax.: +1 225 578-5362<tel:(225)%20578-5362> _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org<mailto:Hdf-forum@lists.hdfgroup.org> http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5 -- Francesc Alted _______________________________________________ Hdf-forum is for HDF software users discussion. Hdf-forum@lists.hdfgroup.org http://lists.hdfgroup.org/mailman/listinfo/hdf-forum_lists.hdfgroup.org Twitter: https://twitter.com/hdf5