[ 
https://issues.apache.org/jira/browse/FOP-2860?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Simon Steiner resolved FOP-2860.
--------------------------------
    Fix Version/s: trunk
       Resolution: Fixed

http://svn.apache.org/viewvc?view=revision&revision=1904062

> [PATCH] BreakingAlgorithm causes high memory consumption
> --------------------------------------------------------
>
>                 Key: FOP-2860
>                 URL: https://issues.apache.org/jira/browse/FOP-2860
>             Project: FOP
>          Issue Type: Bug
>    Affects Versions: 2.3
>            Reporter: Raman Katsora
>            Assignee: Simon Steiner
>            Priority: Critical
>             Fix For: trunk
>
>         Attachments: fop.xconf, image-2019-04-16-10-07-53-502.png, 
> memory6.patch, test-1500000.fo, test-250000.fo, test-300000.fo
>
>
> when a single element (e.g. {{<fo:block>}}) contains a sufficiently large 
> amount of text, the fo-to-pdf transformation causes very high memory 
> consumption.
> For instance, transforming a document with {{<fo:block>}} containing 1.5 
> million characters (~1.5Mb [^test-1500000.fo]) requires about 3Gb of RAM.
> The heapdump shows 27.5 million 
> {{org.apache.fop.layoutmgr.BreakingAlgorithm.KnuthNode}} (~2.6Gb).
> We start observing this issue, having about 300 thousand characters in a 
> single element ([^test-300000.fo]). But the high memory consumption isn't 
> observed when processing 250 thousand characters ([^test-250000.fo]).
> Add <simple-line-breaking>true</simple-line-breaking> to fop.xconf for a 
> lower memory method



--
This message was sent by Atlassian Jira
(v8.20.10#820010)

Reply via email to