hello,     I met a trouble when I used XSSF to read the 07 file which has 75W 
records. And its size is about 30M. When I use XSSFWork to load the file. The 
error would appear : java.lang.OutOfMemory : Java heap space.  I think that the 
problem is the size of file too large. And JVM can't support enough space.
    I don't want to simply increase the memory allocation for JVM. Because it's 
not always useful , if the file which I operate is too much large to over the 
size of RAM.
    I would like to know the solution how to load a large amount of records by 
chunks for downstream processing. Those which api can solve this.  Or anyone 
have better solutions.
  
   Thanks
  Harry Yu

Reply via email to