Hello there, I am trying to parse a large CSV file into XML by the following xquery:
let $file := fetch:text("D:\BPLAN\tlk.txt") let $convert := csv:parse($file, map { 'header' : true(), 'separator' : 'tab'}) return fn:put( <tlks> {for $row in $convert/csv/record return <tlk>{$row/*}</tlk> }</tlks>, "D:\BPLAN\tlk.xml" ) Using the GUI, it runs out of memory -- when I click on the bottom right hand corner (where the memory usage is shown), it says to increase memory, restart using -Xmx <size>. I do this through the MS DOS prompt, but -Xmx does not appear to be a parameter any more, Is there a better method for parsing large CSV files? I then want to add the resulting file tlk.xml to a new database. Kindest Regards Shaun Connelly-Flynn