Re: Reducing heap space consumption for large dictionaries?

2011-12-13 Thread Maciej Lisiewski

W dniu 2011-12-13 05:48, Chris Male pisze:

Hi,

Its good to hear some feedback on using the Hunspell dictionaries.
  Lucene's support is pretty new so we're obviously looking to improve it.
  Could you open a JIRA issue so we can explore whether there is some ways
to reduce memory consumption?


Done:
https://issues.apache.org/jira/browse/SOLR-2968


--
Maciej Lisiewski


Re: Reducing heap space consumption for large dictionaries?

2011-12-12 Thread Maciej Lisiewski

Hi,

in my index schema I has defined a
DictionaryCompoundWordTokenFilterFactory and a
HunspellStemFilterFactory. Each FilterFactory has a dictionary with
about 100k entries.

To avoid an out of memory error I have to set the heap space to 128m
for 1 index.

Is there a way to reduce the memory consumption when parsing the dictionary?
I need to create several indexes and 128m for each index is too much.


Same problem here - even with an empty index (no data yet) and two 
fields using Hunspell (pl_PL) I had to increase heap size to over 2GB 
for solr to start at all..


Stempel using the very same dictionary works fine with 128M..

--
Maciej Lisiewski


Re: Reducing heap space consumption for large dictionaries?

2011-12-12 Thread Chris Male
Hi,

Its good to hear some feedback on using the Hunspell dictionaries.
 Lucene's support is pretty new so we're obviously looking to improve it.
 Could you open a JIRA issue so we can explore whether there is some ways
to reduce memory consumption?

On Tue, Dec 13, 2011 at 5:37 PM, Maciej Lisiewski c2h...@poczta.fm wrote:

 Hi,

 in my index schema I has defined a
 DictionaryCompoundWordTokenFil**terFactory and a
 HunspellStemFilterFactory. Each FilterFactory has a dictionary with
 about 100k entries.

 To avoid an out of memory error I have to set the heap space to 128m
 for 1 index.

 Is there a way to reduce the memory consumption when parsing the
 dictionary?
 I need to create several indexes and 128m for each index is too much.


 Same problem here - even with an empty index (no data yet) and two fields
 using Hunspell (pl_PL) I had to increase heap size to over 2GB for solr to
 start at all..

 Stempel using the very same dictionary works fine with 128M..

 --
 Maciej Lisiewski




-- 
Chris Male | Software Developer | DutchWorks | www.dutchworks.nl


Reducing heap space consumption for large dictionaries?

2011-12-07 Thread Mark Schoy
Hi,

in my index schema I has defined a
DictionaryCompoundWordTokenFilterFactory and a
HunspellStemFilterFactory. Each FilterFactory has a dictionary with
about 100k entries.

To avoid an out of memory error I have to set the heap space to 128m
for 1 index.

Is there a way to reduce the memory consumption when parsing the dictionary?
I need to create several indexes and 128m for each index is too much.

mark