Hello there,
I am trying to parse a large CSV file into XML by the following xquery:
let $file := fetch:text("D:\BPLAN\tlk.txt") let $convert := csv:parse($file, map { 'header' : true(), 'separator' : 'tab'}) return fn:put( <tlks> {for $row in $convert/csv/record return <tlk>{$row/*}</tlk> }</tlks>, "D:\BPLAN\tlk.xml" )
Using the GUI, it runs out of memory -- when I click on the bottom right hand corner (where the memory usage is shown), it says to increase memory, restart using -Xmx <size>.
I do this through the MS DOS prompt, but -Xmx does not appear to be a parameter any more,
Is there a better method for parsing large CSV files? I then want to add the resulting file tlk.xml to a new database.
Kindest Regards Shaun Connelly-Flynn