[ 
https://issues.apache.org/jira/browse/CONNECTORS-1489?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16464726#comment-16464726
 ] 

Karl Wright commented on CONNECTORS-1489:
-----------------------------------------

Hi [~shashank.raj], this is actually a result of your setup, and is not a bug 
with ManifoldCF.

You need to be aware of the following things when sizing the agents process:
(1) If you are not using the extracting update handler, the ENTIRE document is 
fetched into memory.
(2) This can occur at the same time for every worker thread.
(3) In order to size your JVM properly with your configuration, you must 
therefore multiply the number of worker threads by the document maximum size 
(as you configure in your Solr connection), and add 100MB for overhead.

Thanks!


> ManifoldCF stops running with GC Overhead Limit Exceeded
> --------------------------------------------------------
>
>                 Key: CONNECTORS-1489
>                 URL: https://issues.apache.org/jira/browse/CONNECTORS-1489
>             Project: ManifoldCF
>          Issue Type: Bug
>          Components: Lucene/SOLR connector
>    Affects Versions: ManifoldCF 2.9.1
>            Reporter: Shashank Raj
>            Assignee: Karl Wright
>            Priority: Critical
>
> Hello Karl,
> GC Overhead heap error occurs each time and tomcat closes. Heap allocated is 
> 7Gb(Xmx). Is there any other reason this issue is coming up? I am using 
> ManifoldCF's tika. 
> I have Unchecked "Use Update Extract" and max doc size as 50mb.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

Reply via email to