Hi :

I am trying to use your tool SamtoFastq on a very large full genome BAM file ( 127GB ) to extract R1 and R2. But every time i try to run the tool it gives me this error: (after a long time)

INFO 2014-12-04 05:53:44 SamToFastq Processed 1,332,000,000 records. Elapsed time: 04:29:26s. Time for last 1,000,000: 132s. Last read position: hs37d5:21,480,730 INFO 2014-12-04 05:57:06 SamToFastq Processed 1,333,000,000 records. Elapsed time: 04:32:48s. Time for last 1,000,000: 202s. Last read position: hs37d5:21,715,030 INFO 2014-12-04 06:01:47 SamToFastq Processed 1,334,000,000 records. Elapsed time: 04:37:29s. Time for last 1,000,000: 280s. Last read position: hs37d5:22,206,196 [Thu Dec 04 06:08:35 CET 2014] picard.sam.SamToFastq done. Elapsed time: 284.29 minutes.
Runtime.totalMemory()=8485076992
To get help, see http://broadinstitute.github.io/picard/index.html#GettingHelp Exception in thread "main" java.lang.OutOfMemoryError: GC overhead limit exceeded at htsjdk.samtools.BAMRecord.decodeBaseQualities(BAMRecord.java:317)
        at htsjdk.samtools.BAMRecord.getBaseQualities(BAMRecord.java:279)
        at htsjdk.samtools.SAMRecord.isValid(SAMRecord.java:1636)
at htsjdk.samtools.BAMFileReader$BAMFileIterator.advance(BAMFileReader.java:642) at htsjdk.samtools.BAMFileReader$BAMFileIterator.next(BAMFileReader.java:628) at htsjdk.samtools.BAMFileReader$BAMFileIterator.next(BAMFileReader.java:598) at htsjdk.samtools.SamReader$AssertingIterator.next(SamReader.java:514) at htsjdk.samtools.SamReader$AssertingIterator.next(SamReader.java:488)
        at picard.sam.SamToFastq.doWork(SamToFastq.java:149)
at picard.cmdline.CommandLineProgram.instanceMain(CommandLineProgram.java:185)
        at picard.sam.SamToFastq.main(SamToFastq.java:137)

In the beginning of my project i had some files of 80-100 GB which already gave me problems with the GCoverhead limit. There for i changed the setting of GCoverhead to 0.98 instead of your default which was +- 0.50. This helped for the somewhat smaller files. Can you please give me an indication on how to fix this? I read on some previous post: *Re: [Samtools-help] Out of memory samtofastq <http://sourceforge.net/p/samtools/mailman/message/27904292/>*
From: Joachim Jacob <joachim.jacob@vi...> - 2011-08-04 08:28:35
*Attachments:*Message as HTML <http://sourceforge.net/p/samtools/mailman/attachment/4E3A582B.1010809%40vib.be/1/>

Hi all,

I have apparently fixed the issue by using these parameters (24 CPU
machine, with 96GB RAM):

*java -Xmx40g -jar -XX:-UseGCOverheadLimit -XX:-UseParallelGC *

Steadily but firmly the fastq file is being filled.

Joachim

This still doesnt work for my big BAM file. I am currently retrying it with more cores and higher memory. Keep you updated if that works.

Best,

Sander
------------------------------------------------------------------------------
Download BIRT iHub F-Type - The Free Enterprise-Grade BIRT Server
from Actuate! Instantly Supercharge Your Business Reports and Dashboards
with Interactivity, Sharing, Native Excel Exports, App Integration & more
Get technology previously reserved for billion-dollar corporations, FREE
http://pubads.g.doubleclick.net/gampad/clk?id=164703151&iu=/4140/ostg.clktrk
_______________________________________________
Samtools-help mailing list
Samtools-help@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/samtools-help

Reply via email to