Reading point is keeping everything in memory which might have crashed it. pointList.add(record.getSecond());
Your dataset size is 40 MB but the vectors might be too large. How many dimensions are you having in your Vector?
On 04-11-2011 10:57, gaurav redkar wrote:
Hello, I am in a fix with the Clusterdumper utility. The clusterdump utility crashes when it tries to output the clusters by outputting an out of memory exception: java heap space. when i checked the error stack, it seems that the program crashed in readPoints() function. i guess it is unable to build the "result" map. Any idea how do i fix this.?? I am working on a dataset of size 40mb. I had tried increaseing the heap space but with no luck. Thanks Gaurav ----- No virus found in this message. Checked by AVG - www.avg.com Version: 10.0.1411 / Virus Database: 2092/3994 - Release Date: 11/03/11
