Chrsitian,

It does not show any stack trace.
But it shows only "Out Of Main Memor" exception in "Query Info" page
I think it is impossible to process so large XML file 9GM like this.

I had to index xml file first and processed it fine.

Erol Akarsu



On Wed, Jun 18, 2014 at 8:02 PM, Christian Grün <[email protected]>
wrote:

> How does the stack trace look like?
> Am 18.06.2014 16:09 schrieb "Erol Akarsu" <[email protected]>:
>
> I have about 9G xml file and like to process with basex.
>> But I am getting out of memory erroe (with setting -Xmx7G). Yesterday, I
>> was able to process like this where $rawProductsFile is the name of big xml
>> file. XML file consists of "Products" elements and process them one by one
>> and write result into a file.
>>
>> But today, it is out of memory exception.
>>
>>
>> let $res :=
>>
>>     for $doc in fn:doc($rawProductsFile)//Products
>>
>>       let $list:=
>>
>>         for $tname in $tables
>>
>>             let $rows := ($doc[fn:name() eq $tname]|$doc//table[@name eq
>> $tname])
>>
>>             let $list2 := local:processOneProd ($tname,$rows,$tableMap)
>>
>>             let $fw :=
>> file:append(fn:concat($tablesFolder,$tname,".xml"),$list2)
>>
>>             return $fw
>>       return $list
>>
>> Which way is good to process large xml file like this, process each
>> element one be one?
>>
>> I appreciate your help
>>
>> Erol Akarsu
>>
>

Reply via email to