hello, I met a trouble when I used XSSF to read the 07 file which has 75W
records. And its size is about 30M. When I use XSSFWork to load the file. The
error would appear : java.lang.OutOfMemory : Java heap space. I think that the
problem is the size of file too large. And JVM can't support enough space.
I don't want to simply increase the memory allocation for JVM. Because it's
not always useful , if the file which I operate is too much large to over the
size of RAM.
I would like to know the solution how to load a large amount of records by
chunks for downstream processing. Those which api can solve this. Or anyone
have better solutions.
Thanks
Harry Yu