Actually what i figured out just now is that the problem is on the indexing
part. A document with a 15MB size is transformed in a 23MB index which is
not normal since on windows for the same document the index is 3MB. For the
indexing i use:
writer = new IndexWriter(index, new GreekAnalyzer(), !index.exists());
and to add documents:
doc.add(new
Field("contents",content,Field.Store.YES,Field.Index.TOKENIZED));

where "content" is a string with the content of the document. Should i
convert this string to UTF-8 using getBytes before i write it to the index??

-- 
View this message in context: 
http://www.nabble.com/Problem-using-Lucene-on-Ubuntu-tp15543843p15544612.html
Sent from the Lucene - Java Users mailing list archive at Nabble.com.


---------------------------------------------------------------------
To unsubscribe, e-mail: [EMAIL PROTECTED]
For additional commands, e-mail: [EMAIL PROTECTED]

Reply via email to