[ https://issues.apache.org/jira/browse/LUCENE-8043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=16269209#comment-16269209 ]
Adrien Grand commented on LUCENE-8043: -------------------------------------- I can reproduce this but I'm not familiar enough with IndexWriter to understand what causes this. At first I thought thay maybe this was due to the fact the we were giving back documents to early after merges, but actually we do that after updating the list of segment infos, so that looks ok to me. Yet this doesn't prevent the list of segment infos from reaching more that MAX_DOCS documents in {{IndexWriter.publishFlushedSegment}} during the test. [~simonwillnauer] or [~mikemccand] Do you know why this may occur? I wanted to look at the IW info stream to better understand what is happening but unfortunately this probably slows down things enough to prevent the issue from reproducing. It reproduces with assertions enabled ({{-ea}}), but no assertion breaks. > Attempting to add documents past limit can corrupt index > -------------------------------------------------------- > > Key: LUCENE-8043 > URL: https://issues.apache.org/jira/browse/LUCENE-8043 > Project: Lucene - Core > Issue Type: Bug > Components: core/index > Affects Versions: 4.10, 7.0, master (8.0) > Reporter: Yonik Seeley > Attachments: LUCENE-8043.patch > > > The IndexWriter check for too many documents does not always work, resulting > in going over the limit. Once this happens, Lucene refuses to open the index > and throws a CorruptIndexException: Too many documents. > This appears to affect all versions of Lucene/Solr (the check was first > implemented in LUCENE-5843 in v4.9.1/4.10 and we've seen this manifest in > 4.10) -- This message was sent by Atlassian JIRA (v6.4.14#64029) --------------------------------------------------------------------- To unsubscribe, e-mail: dev-unsubscr...@lucene.apache.org For additional commands, e-mail: dev-h...@lucene.apache.org