Split your large file into smaller fragments and index each fragment as a
document.

D.

On Fri, Feb 28, 2025 at 10:30 AM Daniel Cerqueira <dan.l...@lispclub.com>
wrote:

> Hi. I have apache-lucene version 10.1.0:
> ```
> $ pacman -Qs apache-lucene
> local/apache-lucene 10.1.0-1
>     Apache Lucene is a high-performance, full-featured text search engine
> library written entirely in Java.
> ```
>
> I am trying to build a lucene index for a large file.
> ```
> $ ll
> total 2,3G
> -rw------- 1 ** ** 2,3G 2022-12-03 00:35 n-gram5_utf8.txt
> ```
>
> Apache Lucene is blowing up with this large file. It does compute for a
> while, but then it reaches a point where this happens, before it is
> finished:
> ```
> $ java -cp
> /usr/share/java/apache-lucene/lucene-core-10.1.0.jar:/usr/share/java/apache-lucene/lucene-demo-10.1.0.jar:/usr/share/java/apache-lucene/lucene-analysis-common-10.1.0.jar
> org.apache.lucene.demo.IndexFiles -index . -docs n-gram5_utf8.txt
> Indexing to directory '.'...
> WARNING: A restricted method in java.lang.foreign.Linker has been called
> WARNING: java.lang.foreign.Linker::downcallHandle has been called by the
> unnamed module
> WARNING: Use --enable-native-access=ALL-UNNAMED to avoid a warning for
> this module
>
> fev. 27, 2025 3:38:14 DA TARDE
> org.apache.lucene.internal.vectorization.VectorizationProvider lookup
> WARNING: Java vector incubator module is not readable. For optimal vector
> performance, pass '--add-modules jdk.incubator.vector' to enable Vector API.
> adding n-gram5_utf8.txt
> Exception in thread "main" java.lang.IllegalArgumentException: startOffset
> must be non-negative, and endOffset must be >= startOffset; got
> startOffset=2147483645,endOffset=-2147483647
>         at
> org.apache.lucene.analysis.tokenattributes.PackedTokenAttributeImpl.setOffset(PackedTokenAttributeImpl.java:125)
>         at
> org.apache.lucene.analysis.standard.StandardTokenizer.incrementToken(StandardTokenizer.java:167)
>         at
> org.apache.lucene.analysis.LowerCaseFilter.incrementToken(LowerCaseFilter.java:37)
>         at
> org.apache.lucene.analysis.FilteringTokenFilter.incrementToken(FilteringTokenFilter.java:51)
>         at
> org.apache.lucene.index.IndexingChain$PerField.invertTokenStream(IndexingChain.java:1218)
>         at
> org.apache.lucene.index.IndexingChain$PerField.invert(IndexingChain.java:1196)
>         at
> org.apache.lucene.index.IndexingChain.processField(IndexingChain.java:741)
>         at
> org.apache.lucene.index.IndexingChain.processDocument(IndexingChain.java:618)
>         at
> org.apache.lucene.index.DocumentsWriterPerThread.updateDocuments(DocumentsWriterPerThread.java:274)
>         at
> org.apache.lucene.index.DocumentsWriter.updateDocuments(DocumentsWriter.java:425)
>         at
> org.apache.lucene.index.IndexWriter.updateDocuments(IndexWriter.java:1552)
>         at
> org.apache.lucene.index.IndexWriter.updateDocument(IndexWriter.java:1837)
>         at
> org.apache.lucene.index.IndexWriter.addDocument(IndexWriter.java:1477)
>         at org.apache.lucene.demo.IndexFiles.indexDoc(IndexFiles.java:274)
>         at org.apache.lucene.demo.IndexFiles.indexDocs(IndexFiles.java:225)
>         at org.apache.lucene.demo.IndexFiles.main(IndexFiles.java:158)
> ```
>
> How can this be fixed, and how can I build a lucene index for this
> large file?
>
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: java-user-unsubscr...@lucene.apache.org
> For additional commands, e-mail: java-user-h...@lucene.apache.org
>
>

Reply via email to