Does this capture the issue well enough: https://github.com/numenta/nupic/issues/1231 ? --------- Matt Taylor OS Community Flag-Bearer Numenta
On Thu, Aug 21, 2014 at 7:46 AM, Ryan Belcher <[email protected]> wrote: > I don't think it has anything to do with the size of the data (input) but > rather the size of the region(s) you're trying to serialize. For example, I > had a SP region with 12288 columns encoding an SDR that was probably 40k > bits. I think if your SP/TP is big enough, you can just feed it 1 record > and it will fail when saving. > > > On Thu, Aug 21, 2014 at 9:50 AM, Marek Otahal <[email protected]> wrote: >> >> I think python 3 is not realistic for us now. >> >> what they suggest is to pickle at smaller parts, and we do have a code for >> that >> "workaround is to pickle your array yourself in smaller parts, or use some >> other file format than Python pickles." >> >> if you can make a test-case with large data, it would be helpful. >> >> Cheers, >> >> _______________________________________________ >> nupic mailing list >> [email protected] >> http://lists.numenta.org/mailman/listinfo/nupic_lists.numenta.org >> > > > _______________________________________________ > nupic mailing list > [email protected] > http://lists.numenta.org/mailman/listinfo/nupic_lists.numenta.org > _______________________________________________ nupic mailing list [email protected] http://lists.numenta.org/mailman/listinfo/nupic_lists.numenta.org
