What makes you think saving to a binary and loading from a binary would gain
much efficiency. Right now parsing an in-memory XML string with libxml2 is
close to 2 times faster than trying to copy the same preparsed xmlDocPtr
subtree.

I can achieve 3x efficiency in loading raw data from disk compared to parsing the document by storing raw SAX events to a BerkeleyDB. The gain for small records is zero, but for 1 megabyte plus records the savings become significant.


Note that this isn't the DocPtr tree, I'm not debating that! :)

-- Rob

,'/:. Dr Robert Sanderson ([EMAIL PROTECTED])
,'-/::::. http://www.csc.liv.ac.uk/~azaroth/
,'--/::(@)::. Dept. of Computer Science, Room 805
,'---/::::::::::. University of Liverpool
____/:::::::::::::. I L L U M I N A T I Cheshire3 IR System: http://www.cheshire3.org/
_______________________________________________
xml mailing list, project page http://xmlsoft.org/
[email protected]
http://mail.gnome.org/mailman/listinfo/xml

Reply via email to