[ 
https://issues.apache.org/jira/browse/LUCENENET-257?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12779428#action_12779428
 ] 

Andrei Iliev commented on LUCENENET-257:
----------------------------------------

Thanks Michael, 
Something still bothering me. 
LogByteSizeMergePolicy has a public method SetMaxMergeMB:
public virtual void  SetMaxMergeMB(double mb)
                {
                        maxMergeSize = (long) (mb * 1024 * 1024); // <-- there 
is a chance to get negative maxMergeSize 
                }

If we started to fix Lucene java's  bug, would it be better to declare 
maxMergeSize as double?

> TestCheckIndex.TestDeletedDocs 
> -------------------------------
>
>                 Key: LUCENENET-257
>                 URL: https://issues.apache.org/jira/browse/LUCENENET-257
>             Project: Lucene.Net
>          Issue Type: Bug
>            Reporter: Andrei Iliev
>            Assignee: Digy
>         Attachments: LUCENENET-257.patch
>
>
> Setting writer.SetMaxBufferedDocs(2) couse flushing buffers after adding 
> every 2 docs. That  results in total 10 segments in the index and failing 
> TestDeletedDocs (test case assumes that there is only 1 segment file).  So 
> question arise:
> 1) does performing a flush has to start a new segment file?
> 2) if so, in order  to  run TestDeletedDocs smoothly either change  
> writer.SetMaxBufferedDocs(2) to, say,    writer.SetMaxBufferedDocs(20) or 
> call writer.optimize(1) before closing writer.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to