[jira] [Updated] (XALANJ-2607) Improve performance for large documents using ID attributes
[ https://issues.apache.org/jira/browse/XALANJ-2607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matthias Urban updated XALANJ-2607: --- Description: XalanJ gets very slow for large XML documents using ID attributes often used in article lists. If, for instance, an article list with 1.000.000 entries is parsed, then it takes 6 minutes (on my machine) just to build the DTM. This is due to a design decision in DTMStringPool utilizing a fixed size hash table of 101 entries. This works astoundingly well for documents with less than 10.000 different attribute values. Then it starts to get slower and slower. I've tested multiple solutions, like to increase the hash table size to 100.000 and 1.000.000 (overkill for everyday documents), or make it configurable leaving the current size the default (needs to introduce a 'backdoor' system property). In the end the best solution was to use a simple HashMap for the string lookup. It proved to be as fast as the fixed size table combined with a very good scalability. See the patch attached to this issue. It was created using the current trunk version of DTMStringPool.java. Here is an example for testing. Without the patch applied it takes about 6 minutes to finish. With the patch applied it only takes 3 seconds! {code:title=Test.java} import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.util.Date; import javax.xml.transform.Transformer; import javax.xml.transform.TransformerFactory; import javax.xml.transform.stream.StreamResult; import javax.xml.transform.stream.StreamSource; public class Test { public static void main(String[] args) { try { File xmlfile = new File("doc.xml"); FileWriter fout = new FileWriter(xmlfile); fout.write("\n"); for (int i = 0; i < 100; i++) { fout.append("articlename\n"); } fout.write(""); fout.close(); System.out.println("Start : " + new Date()); TransformerFactory factory = TransformerFactory.newInstance(); Transformer transformer = factory.newTransformer(new StreamSource(new File("script.xsl"))); transformer.transform(new StreamSource(xmlfile), new StreamResult(new FileOutputStream("out.txt"))); System.out.println("End : " + new Date()); } catch (Exception e) { e.printStackTrace(); } } } {code} {code:xml|title=script.xsl} http://www.w3.org/1999/XSL/Transform;> {code} was: XalanJ gets very slow for large XML documents using ID attributes often used in article lists. If, for instance, an article list with 1.000.000 entries is parsed, then it takes 6 minutes (on my machine) just to build the DTM. This is due to a design decision in DTMStringPool utilizing a fixed size hash table of 101 entries. This works astoundingly well for documents with less than 100.000 different attribute values (at least on my machine). Then it starts to get slower and slower. I've tested multiple solutions, like to increase the hash table size to 100.000 and 1.000.000 (overkill for everyday documents), or make it configurable leaving the current size the default (needs to introduce a 'backdoor' system property). In the end the best solution is to use a simple HashMap for the string lookup. It proved to be as fast as the fixed size table combined with a very good scalability. See the patch attached to this issue. It was created using the current trunk version of DTMStringPool.java. Here is an example for testing. {code:title=Test.java} import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.util.Date; import javax.xml.transform.Transformer; import javax.xml.transform.TransformerFactory; import javax.xml.transform.stream.StreamResult; import javax.xml.transform.stream.StreamSource; public class Test { public static void main(String[] args) { try { File xmlfile = new File("doc.xml"); FileWriter fout = new FileWriter(xmlfile); fout.write("\n"); for (int i = 0; i < 100; i++) { fout.append("articlename\n"); } fout.write(""); fout.close(); System.out.println("Start : " + new Date()); TransformerFactory factory = TransformerFactory.newInstance(); Transformer transformer = factory.newTransformer(new StreamSource(new File("script.xsl"))); transformer.transform(new StreamSource(xmlfile), new StreamResult(new FileOutputStream("out.txt"))); System.out.println("End : " + new Date()); } catch (Exception e) { e.printStackTrace(); } } } {code} {code:xml|title=script.xsl} http://www.w3.org/1999/XSL/Transform;> {code} > Improve performance for large documents using ID attributes > --- > > Key: XALANJ-2607 > URL:
[jira] [Updated] (XALANJ-2607) Improve performance for large documents using ID attributes
[ https://issues.apache.org/jira/browse/XALANJ-2607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matthias Urban updated XALANJ-2607: --- Attachment: DTMStringPool.patch > Improve performance for large documents using ID attributes > --- > > Key: XALANJ-2607 > URL: https://issues.apache.org/jira/browse/XALANJ-2607 > Project: XalanJ2 > Issue Type: Improvement > Security Level: No security risk; visible to anyone(Ordinary problems in > Xalan projects. Anybody can view the issue.) > Components: DTM >Affects Versions: 2.7.2 > Environment: Tested on but not limited to: Windows, x64, JRE 1.8 >Reporter: Matthias Urban >Assignee: Steven J. Hathaway > Attachments: DTMStringPool.patch > > > XalanJ gets very slow for large XML documents using ID attributes often used > in article lists. If, for instance, an article list with 1.000.000 entries is > parsed, then it takes 6 minutes (on my machine) just to build the DTM. This > is due to a design decision in DTMStringPool utilizing a fixed size hash > table of 101 entries. This works astoundingly well for documents with less > than 100.000 different attribute values (at least on my machine). Then it > starts to get slower and slower. > I've tested multiple solutions, like to increase the hash table size to > 100.000 and 1.000.000 (overkill for everyday documents), or make it > configurable leaving the current size the default (needs to introduce a > 'backdoor' system property). In the end the best solution is to use a simple > HashMap for the string lookup. It proved to be as fast as the fixed size > table combined with a very good scalability. See the patch attached to this > issue. It was created using the current trunk version of DTMStringPool.java. > Here is an example for testing. > {code:title=Test.java} > import java.io.File; > import java.io.FileOutputStream; > import java.io.FileWriter; > import java.util.Date; > import javax.xml.transform.Transformer; > import javax.xml.transform.TransformerFactory; > import javax.xml.transform.stream.StreamResult; > import javax.xml.transform.stream.StreamSource; > public class Test { > public static void main(String[] args) { > try { > File xmlfile = new File("doc.xml"); > FileWriter fout = new FileWriter(xmlfile); > fout.write("\n"); > for (int i = 0; i < 100; i++) { > fout.append("articlename\n"); > } > fout.write(""); > fout.close(); > System.out.println("Start : " + new Date()); > TransformerFactory factory = TransformerFactory.newInstance(); > Transformer transformer = factory.newTransformer(new StreamSource(new > File("script.xsl"))); > transformer.transform(new StreamSource(xmlfile), new StreamResult(new > FileOutputStream("out.txt"))); > System.out.println("End : " + new Date()); > } > catch (Exception e) { > e.printStackTrace(); > } > } > } > {code} > {code:xml|title=script.xsl} > > xmlns:xsl="http://www.w3.org/1999/XSL/Transform;> > > > > > > > > > > > {code} -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: dev-unsubscr...@xalan.apache.org For additional commands, e-mail: dev-h...@xalan.apache.org
[jira] [Updated] (XALANJ-2607) Improve performance for large documents using ID attributes
[ https://issues.apache.org/jira/browse/XALANJ-2607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matthias Urban updated XALANJ-2607: --- Description: XalanJ gets very slow for large XML documents using ID attributes often used in article lists. If, for instance, an article list with 1.000.000 entries is parsed, then it takes 6 minutes (on my machine) just to build the DTM. This is due to a design decision in DTMStringPool utilizing a fixed size hash table of 101 entries. This works astoundingly well for documents with less than 100.000 different attribute values (at least on my machine). Then it starts to get slower and slower. I've tested multiple solutions, like to increase the hash table size to 100.000 and 1.000.000 (overkill for everyday documents), or make it configurable leaving the current size the default (needs to introduce a 'backdoor' system property). In the end the best solution is to use a simple HashMap for the string lookup. It proved to be as fast as the fixed size table combined with a very good scalability. See the patch attached to this issue. It was created using the current trunk version of DTMStringPool.java. Here is an example for testing. {code:title=Test.java} import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.util.Date; import javax.xml.transform.Transformer; import javax.xml.transform.TransformerFactory; import javax.xml.transform.stream.StreamResult; import javax.xml.transform.stream.StreamSource; public class Test { public static void main(String[] args) { try { File xmlfile = new File("doc.xml"); FileWriter fout = new FileWriter(xmlfile); fout.write("\n"); for (int i = 0; i < 100; i++) { fout.append("articlename\n"); } fout.write(""); fout.close(); System.out.println("Start : " + new Date()); TransformerFactory factory = TransformerFactory.newInstance(); Transformer transformer = factory.newTransformer(new StreamSource(new File("script.xsl"))); transformer.transform(new StreamSource(xmlfile), new StreamResult(new FileOutputStream("out.txt"))); System.out.println("End : " + new Date()); } catch (Exception e) { e.printStackTrace(); } } } {code} {code:xml|title=script.xsl} http://www.w3.org/1999/XSL/Transform;> {code} was: XalanJ gets very slow for large XML documents using ID attributes often used in article lists. If, for instance, an article list with 1.000.000 entries is parsed, then it takes 6 minutes (on my machine) just to build the DTM. This is due to a design decision in DTMStringPool utilizing a fixed size hash table of 101 entries. This works astoundingly well for documents with less than 100.000 different attribute values (at least on my machine). Then it starts to get slower and slower. A minimally invasive solution is to make the hash table size configurable leaving the current size the default. If an application is expected to work with large documents, then it can increase the hash table size. For all others nothing changes. See the patch attached to this issue. It was created using the current trunk version of DTMStringPool.java. Here is an example for testing. {code:title=Test.java} import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.util.Date; import javax.xml.transform.Transformer; import javax.xml.transform.TransformerFactory; import javax.xml.transform.stream.StreamResult; import javax.xml.transform.stream.StreamSource; public class Test { public static void main(String[] args) { try { File xmlfile = new File("doc.xml"); FileWriter fout = new FileWriter(xmlfile); fout.write("\n"); for (int i = 0; i < 100; i++) { fout.append("articlename\n"); } fout.write(""); fout.close(); // reduce the time needed to 3 seconds! otherwise it will take 6 minutes! System.setProperty("org.apache.xml.dtm.ref.DTMStringPool.hashPoolSize", "10"); System.out.println("Start : " + new Date()); TransformerFactory factory = TransformerFactory.newInstance(); Transformer transformer = factory.newTransformer(new StreamSource(new File("script.xsl"))); transformer.transform(new StreamSource(xmlfile), new StreamResult(new FileOutputStream("out.txt"))); System.out.println("End : " + new Date()); } catch (Exception e) { e.printStackTrace(); } } } {code} {code:xml|title=script.xsl} http://www.w3.org/1999/XSL/Transform;> {code} > Improve performance for large documents using ID attributes > --- > > Key: XALANJ-2607 > URL: https://issues.apache.org/jira/browse/XALANJ-2607 > Project: XalanJ2 >
[jira] [Updated] (XALANJ-2607) Improve performance for large documents using ID attributes
[ https://issues.apache.org/jira/browse/XALANJ-2607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matthias Urban updated XALANJ-2607: --- Attachment: (was: DTMStringPool.patch) > Improve performance for large documents using ID attributes > --- > > Key: XALANJ-2607 > URL: https://issues.apache.org/jira/browse/XALANJ-2607 > Project: XalanJ2 > Issue Type: Improvement > Security Level: No security risk; visible to anyone(Ordinary problems in > Xalan projects. Anybody can view the issue.) > Components: DTM >Affects Versions: 2.7.2 > Environment: Tested on but not limited to: Windows, x64, JRE 1.8 >Reporter: Matthias Urban >Assignee: Steven J. Hathaway > > XalanJ gets very slow for large XML documents using ID attributes often used > in article lists. If, for instance, an article list with 1.000.000 entries is > parsed, then it takes 6 minutes (on my machine) just to build the DTM. This > is due to a design decision in DTMStringPool utilizing a fixed size hash > table of 101 entries. This works astoundingly well for documents with less > than 100.000 different attribute values (at least on my machine). Then it > starts to get slower and slower. > A minimally invasive solution is to make the hash table size configurable > leaving the current size the default. If an application is expected to work > with large documents, then it can increase the hash table size. For all > others nothing changes. See the patch attached to this issue. It was created > using the current trunk version of DTMStringPool.java. > Here is an example for testing. > {code:title=Test.java} > import java.io.File; > import java.io.FileOutputStream; > import java.io.FileWriter; > import java.util.Date; > import javax.xml.transform.Transformer; > import javax.xml.transform.TransformerFactory; > import javax.xml.transform.stream.StreamResult; > import javax.xml.transform.stream.StreamSource; > public class Test { > public static void main(String[] args) { > try { > File xmlfile = new File("doc.xml"); > FileWriter fout = new FileWriter(xmlfile); > fout.write("\n"); > for (int i = 0; i < 100; i++) { > fout.append("articlename\n"); > } > fout.write(""); > fout.close(); > // reduce the time needed to 3 seconds! otherwise it will take 6 > minutes! > System.setProperty("org.apache.xml.dtm.ref.DTMStringPool.hashPoolSize", > "10"); > System.out.println("Start : " + new Date()); > TransformerFactory factory = TransformerFactory.newInstance(); > Transformer transformer = factory.newTransformer(new StreamSource(new > File("script.xsl"))); > transformer.transform(new StreamSource(xmlfile), new StreamResult(new > FileOutputStream("out.txt"))); > System.out.println("End : " + new Date()); > } > catch (Exception e) { > e.printStackTrace(); > } > } > } > {code} > {code:xml|title=script.xsl} > > xmlns:xsl="http://www.w3.org/1999/XSL/Transform;> > > > > > > > > > > > {code} -- This message was sent by Atlassian JIRA (v6.3.15#6346) - To unsubscribe, e-mail: dev-unsubscr...@xalan.apache.org For additional commands, e-mail: dev-h...@xalan.apache.org
[jira] [Updated] (XALANJ-2607) Improve performance for large documents using ID attributes
[ https://issues.apache.org/jira/browse/XALANJ-2607?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel ] Matthias Urban updated XALANJ-2607: --- Description: XalanJ gets very slow for large XML documents using ID attributes often used in article lists. If, for instance, an article list with 1.000.000 entries is parsed, then it takes 6 minutes (on my machine) just to build the DTM. This is due to a design decision in DTMStringPool utilizing a fixed size hash table of 101 entries. This works astoundingly well for documents with less than 100.000 different attribute values (at least on my machine). Then it starts to get slower and slower. A minimally invasive solution is to make the hash table size configurable leaving the current size the default. If an application is expected to work with large documents, then it can increase the hash table size. For all others nothing changes. See the patch attached to this issue. It was created using the current trunk version of DTMStringPool.java. Here is an example for testing. {code:title=Test.java} import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.util.Date; import javax.xml.transform.Transformer; import javax.xml.transform.TransformerFactory; import javax.xml.transform.stream.StreamResult; import javax.xml.transform.stream.StreamSource; public class Test { public static void main(String[] args) { try { File xmlfile = new File("doc.xml"); FileWriter fout = new FileWriter(xmlfile); fout.write("\n"); for (int i = 0; i < 100; i++) { fout.append("articlename\n"); } fout.write(""); fout.close(); // reduce the time needed to 3 seconds! otherwise it will take 6 minutes! System.setProperty("org.apache.xml.dtm.ref.DTMStringPool.hashPoolSize", "10"); System.out.println("Start : " + new Date()); TransformerFactory factory = TransformerFactory.newInstance(); Transformer transformer = factory.newTransformer(new StreamSource(new File("script.xsl"))); transformer.transform(new StreamSource(xmlfile), new StreamResult(new FileOutputStream("out.txt"))); System.out.println("End : " + new Date()); } catch (Exception e) { e.printStackTrace(); } } } {code} {code:xml|title=script.xsl} http://www.w3.org/1999/XSL/Transform;> {code} was: XalanJ gets very slow for large XML documents using ID attributes often used in article lists. If, for instance, an article list with 1.000.000 entries is parsed, then it takes 6 minutes (on my machine) just to build the DTM. This is due to a design decision in DTMStringPool utilizing a fixed size hash table of 101 entries. This works astoundingly well for documents with less than 100.000 different attribute values (at least on my machine). Then it starts to get slower and slower. A minimally invasive solution is to make the hash table size configurable leaving the current size the default. If an application is expected to work with large documents, then it can increase the hash table size. For all others nothing changes. See the patch attached to this issue. It was created using the current trunk version of DTMStringPool.java. Here is an example for testing. {code:title=Test.java} import java.io.File; import java.io.FileOutputStream; import java.io.FileWriter; import java.util.Date; import javax.xml.transform.Transformer; import javax.xml.transform.TransformerFactory; import javax.xml.transform.stream.StreamResult; import javax.xml.transform.stream.StreamSource; public class Test { public static void main(String[] args) { try { File xmlfile = new File("doc.xml"); FileWriter fout = new FileWriter(xmlfile); fout.write("\n"); for (int i = 0; i < 100; i++) { fout.append("articlename\n"); } fout.write(""); fout.close(); // reduce the time needed to 3 seconds! otherwise it will take 6 minutes! System.setProperty("org.apache.xml.dtm.ref.DTMStringPool.hashPoolSize", "10"); System.out.println("Start : " + new Date()); TransformerFactory factory = TransformerFactory.newInstance(); Transformer transformer = factory.newTransformer(new StreamSource(new File("script.xsl"))); transformer.transform(new StreamSource(xmlfile), new StreamResult(new FileOutputStream("out.txt"))); System.out.println("End : " + new Date()); } catch (Exception e) { e.printStackTrace(); } } } {code}