Christian No I can run with debug and let you know result
Sent from my iPhone > On Jun 19, 2014, at 6:23 PM, Christian Grün <[email protected]> wrote: > > Did you use debugging (-d)? > > Am 19.06.2014 22:01 schrieb "Erol Akarsu" <[email protected]>: >> Chrsitian, >> >> It does not show any stack trace. >> But it shows only "Out Of Main Memor" exception in "Query Info" page >> I think it is impossible to process so large XML file 9GM like this. >> >> I had to index xml file first and processed it fine. >> >> Erol Akarsu >> >> >> >>> On Wed, Jun 18, 2014 at 8:02 PM, Christian Grün <[email protected]> >>> wrote: >>> How does the stack trace look like? >>> >>> Am 18.06.2014 16:09 schrieb "Erol Akarsu" <[email protected]>: >>> >>>> I have about 9G xml file and like to process with basex. >>>> But I am getting out of memory erroe (with setting -Xmx7G). Yesterday, I >>>> was able to process like this where $rawProductsFile is the name of big >>>> xml file. XML file consists of "Products" elements and process them one by >>>> one and write result into a file. >>>> >>>> But today, it is out of memory exception. >>>> >>>> let $res := >>>> >>>> for $doc in fn:doc($rawProductsFile)//Products >>>> >>>> let $list:= >>>> >>>> for $tname in $tables >>>> >>>> let $rows := ($doc[fn:name() eq $tname]|$doc//table[@name eq >>>> $tname]) >>>> >>>> let $list2 := local:processOneProd ($tname,$rows,$tableMap) >>>> >>>> let $fw := >>>> file:append(fn:concat($tablesFolder,$tname,".xml"),$list2) >>>> >>>> return $fw >>>> >>>> return $list >>>> >>>> Which way is good to process large xml file like this, process each >>>> element one be one? >>>> >>>> I appreciate your help >>>> >>>> Erol Akarsu

