[ 
https://issues.apache.org/jira/browse/SOLR-1911?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12867758#action_12867758
 ] 

Hoss Man commented on SOLR-1911:
--------------------------------

Simon: can you please provide a few more details:

1) what mechanism are you using to index content?  ie: POSTing XML from a 
remote client? using the stream.url or stream.file params? Using SolrCell? 
using DIH? (and if you are using DIH, from what source? DB? HTTP? File? .. and 
with what transformers?)

2) what files does lsof show are open after each successive commit until the 
limit is reached?  seeing how the file list grows -- specifically which files 
are never getting closed -- over time is really the only way to track down what 
code isn't closing files

(The stack trace you posted shows what it's doing when it runs out of file 
handles, but there isn't necessarily any correlation between that and what code 
should be closing files but isn't )

> File descriptor leak while indexing, may cause index corruption
> ---------------------------------------------------------------
>
>                 Key: SOLR-1911
>                 URL: https://issues.apache.org/jira/browse/SOLR-1911
>             Project: Solr
>          Issue Type: Bug
>          Components: update
>    Affects Versions: 1.5
>         Environment:         Ubuntu Linux, Java build 1.6.0_16-b01
>         Solr Specification Version: 3.0.0.2010.05.12.16.17.46
>       Solr Implementation Version: 4.0-dev exported - simon - 2010-05-12 
> 16:17:46  -- bult from updated trunk
>       Lucene Specification Version: 4.0-dev
>       Lucene Implementation Version: 4.0-dev exported - 2010-05-12 16:18:26
>       Current Time: Thu May 13 12:21:12 EDT 2010
>       Server Start Time:Thu May 13 11:45:41 EDT 2010
>            Reporter: Simon Rosenthal
>            Priority: Critical
>
> While adding documents to an already existing index using this build, the 
> number of open file descriptors increases dramatically until the open file 
> per-process limit is reached (1024) , at which point there are error messages 
> in the log to that effect. If the server is restarted the index may be corrupt
> commits are handled by autocommit every 60 seconds or 500 documents (usually 
> the time limit is reached first). 
> mergeFactor is 10.
> It looks as though each time a commit takes place, the number of open files  
> (obtained from " lsof -p `cat solr.pid` | egrep ' [0-9]+r ' ") increases by 
> 40, There are several open file descriptors associated with each file in the 
> index.
> Rerunning the same index updates with an older Solr (built from trunk in Feb 
> 2010) doesn't show this problem - the number of open files fluctuates up and 
> down as segments are created and merged, but stays basically constant.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: dev-h...@lucene.apache.org

Reply via email to