ok -- quick update: As far as I can see issue boiled down to lack of support for such large data "files" in zlib (compression library used by libnifti). I am not sure how quickly issue would get resolved since it requires breaking an ABI of the library (and may be more). I've emailed the authors, waiting for reply.
On a good side: there is a pure Python reimplementation of pynifti by Matthew Brett and Michael Hanke (yes yes -- the same PyMVPA Michael) -- it is coming to replace current pynifti, and is already supported in our development (0.5) branch of PyMVPA. On my quick trials nibabel (and new mri_dataset) seems to support large files without any problem ;) Both nibabel and development snapshot of PyMVPA are available from our NeuroDebian repository if you are using Debian-derived system, so you could give them a try if you desperately need to work with compressed large nifti files. Be warned though, that development version of PyMVPA, although quite usable and somewhat stable, is still in a fluid state and API might change (not to say that it is already different from 0.4.x versions PyMVPA). But you can use dev.pymvpa.org as the ultimate source of documentation and inspiration. With best regards, Yarik On Sun, 27 Jun 2010, McKell Carter wrote: > Wow, very quick work tracking it! When I load the uncompressed file > I'm able to access the full array. I'm running a couple tests with > pymvpa now to make sure, but that seems a good way to get things > working for now. Please do let me know of any updates. > Cheers, > McKell -- .-. =------------------------------ /v\ ----------------------------= Keep in touch // \\ (yoh@|www.)onerussian.com Yaroslav Halchenko /( )\ ICQ#: 60653192 Linux User ^^-^^ [175555] _______________________________________________ Pkg-ExpPsy-PyMVPA mailing list Pkg-ExpPsy-PyMVPA@lists.alioth.debian.org http://lists.alioth.debian.org/mailman/listinfo/pkg-exppsy-pymvpa