Hi all,
        I seemed to have found the solution to this problem. Apparently, 
allocating enough virtual memory on the server seems to only solve on half of 
the problem. Even after allocating 4 gigs of Virtual memory on jboss server, I 
still did get the Out of memory on sorting. 
 
I didn't how ever notice that the LRU cache on my config was set to default 
which was still 512 megs of max memory. I had to increase that to a round 2 
gigs and the sorting did work perfectly ok.
 
Even though I am satisfied that I have found the solution to the problem, i am 
still not satisfied to know that Sort consumes so much memory. In no products 
have I seen sorting 10 fields take up 1 gig and half of virtual memory. I am 
not sure, if there could be a better implementation of this. But something 
doesn't seem right to me.
 
Thanks for all your support. It has truly been overwhelming.
 
Sundar



> From: [EMAIL PROTECTED]> To: solr-user@lucene.apache.org> Subject: RE: Out of 
> memory on Solr sorting> Date: Tue, 29 Jul 2008 10:43:05 -0700> > A sneaky 
> source of OutOfMemory errors is the permanent generation. If you> add this:> 
> -XX:PermSize=64m -XX:MaxPermSize=96m> You will increase the size of the 
> permanent generation. We found this> helped.> > Also note that when you 
> undeploy a war file, the old deployment has> permanent storage that is not 
> reclaimed, and so each undeploy/redeploy cycle> eats up the permanent 
> generation pool.> > -----Original Message-----> From: david w [mailto:[EMAIL 
> PROTECTED] > Sent: Tuesday, July 29, 2008 7:20 AM> To: 
> solr-user@lucene.apache.org> Subject: Re: Out of memory on Solr sorting> > 
> Hi, Daniel> > I got the same probem like Sundar. Is that possible to tell me 
> what> profiling tool you are using?> > thx a lot.> > /David> > On Tue, Jul 
> 29, 2008 at 8:19 PM, Daniel Alheiros> <[EMAIL PROTECTED]>wrote:> > > Hi 
> Sundar.> >> > Well it would be good if you could do some profiling on your 
> Solr app.> > I've done it during the indexing process so I could figure out 
> what > > was going on in the OutOfMemoryErrors I was getting.> >> > But you 
> won't definitelly need to have as much memory as your whole > > index size. I 
> have a 3.5 million documents (aprox. 10Gb) running on > > this 2Gb heap VM.> 
> >> > Cheers,> > Daniel> >> > -----Original Message-----> > From: sundar 
> shankar [mailto:[EMAIL PROTECTED]> > Sent: 23 July 2008 23:45> > To: 
> solr-user@lucene.apache.org> > Subject: RE: Out of memory on Solr sorting> >> 
> >> > Hi Daniel,> > I am afraid that didnt solve my problem. I was guessing my 
> > > problem was that I have too much of data and too little memory > > 
> allocated for that. I happened to read in couple of the posts which > > 
> mentioned that I need VM that is close to the size of my data(folder). > > I 
> have like 540 Megs now and a little more than a million and a half > > docs. 
> Ideally in that case 512 megs should be enough for me. In fact I > > am able 
> to perform all other operations now, commit, optmize, select, > > update, 
> nightly cron jobs to index data again. etc etc with no > > hassles. Even my 
> load tests perform very well. Just the sort and it > > doesnt seem to work. I 
> allocated> > 2 gigs of memory now. Still same results. Used the GC params u 
> gave me > > too. No change what so ever. Am not sure, whats going on. Is 
> there > > something that I can do to find out how much is needed in actuality 
> as > > my production server might need to be configured in accordance.> >> > 
> I dont store any documents. We basically fetch standard column data > > from 
> oracle database store them into Solr fields. Before I had > > EdgeNGram 
> configured and had Solr 1.2, My data size was less that half > > of what it 
> is right now. I guess if I remember right, it was of the > > order of 100 
> megs. The max size of a field right now might not cross a 100> chars too.> > 
> Quizzled even more now.> >> > -Sundar> >> > P.S: My configurations :> > Solr 
> 1.3> > Red hat> > 540 megs of data (1855013 docs)> > 2 gigs of memory 
> installed and allocated like this > > JAVA_OPTS=$JAVA_OPTS -Xms2048m 
> -Xmx2048m -XX:MinHeapFreeRatio=50 > > -XX:NewSize=1024m> > -XX:NewRatio=2 
> -Dsun.rmi.dgc.client.gcInterval=3600000> > 
> -Dsun.rmi.dgc.server.gcInterval=3600000> >> > Jboss 4.05> >> >> > > Subject: 
> RE: Out of memory on Solr sorting> > > Date: Wed, 23 Jul 2008 10:49:06 +0100> 
> > > From: [EMAIL PROTECTED]> > > To: solr-user@lucene.apache.org> > >> > > 
> Hi> > >> > > I haven't read the whole thread so I will take my chances here.> 
> > >> > > I've been fighting recently to keep my Solr instances stable because 
> > > > they were frequently crashing with OutOfMemoryErrors. I'm using Solr> > 
> > 1.2 and when it happens there is a bug that makes the index locked > > > 
> unless you restart Solr... So in my cenario it was extremelly> > damaging.> > 
> >> > > After some profiling I realized that my major problem was caused by > 
> > > the way the JVM heap was being used as I haven't configured it to > > > 
> run using any advanced configuration (I had just made it bigger - > > > Xmx 
> and Xms 1.5 Gb), it's running on Sun JVM 1.5 (the most recent > > > 1.5> > > 
> available) and it's deployed on a Jboss 4.2 on a RHEL.> > >> > > So my 
> findings were too many objects were being allocated on the old > > > 
> generation area of the heap, which makes them harder to be disposed, > > > 
> and also the default behaviour was letting the heap get too filled > > > up 
> before kicking a GC and according to the JVM specs the default is > > > if 
> after a short period when a full gc is executed if a certain > > > percentage 
> of the heap is not freed an OutOfMemoryError should be> > thrown.> > >> > > 
> I've changed my JVM startup params and it's working extremelly > > > stable 
> since then:> > >> > > -Xmx2048m -Xms2048m -XX:MinHeapFreeRatio=50 
> -XX:NewSize=1024m> > > -XX:NewRatio=2 
> -Dsun.rmi.dgc.client.gcInterval=3600000> > > 
> -Dsun.rmi.dgc.server.gcInterval=3600000> > >> > > I hope it helps.> > >> > > 
> Regards,> > > Daniel Alheiros> > >> > > -----Original Message-----> > > From: 
> Fuad Efendi [mailto:[EMAIL PROTECTED]> > > Sent: 22 July 2008 23:23> > > To: 
> solr-user@lucene.apache.org> > > Subject: RE: Out of memory on Solr sorting> 
> > >> > > Yes, it is a cache, it stores "sorted" by "sorted field" array of > 
> > > Document IDs together with sorted fields; query results can > > > 
> intersect with it and reorder accordingly.> > >> > > But memory requirements 
> should be well documented.> > >> > > It uses internally WeakHashMap which is 
> not good(!!!) - a lot of > > > "underground" warming ups of caches which SOLR 
> is not aware of...> > > Could be.> > >> > > I think Lucene-SOLR developers 
> should join this discussion:> > >> > >> > > /**> > > * Expert: The default 
> cache implementation, storing all values in > > > memory.> > > * A 
> WeakHashMap is used for storage.> > > *> > > ..............> > >> > > // 
> inherit javadocs> > > public StringIndex getStringIndex(IndexReader reader, 
> String field)> > > throws IOException {> > > return (StringIndex) 
> stringsIndexCache.get(reader, field);> > > }> > >> > > Cache 
> stringsIndexCache = new Cache() {> > >> > > protected Object 
> createValue(IndexReader reader, Object fieldKey)> > > throws IOException {> > 
> > String field = ((String) fieldKey).intern();> > > final int[] retArray = 
> new int[reader.maxDoc()];> > > String[] mterms = new 
> String[reader.maxDoc()+1];> > > TermDocs termDocs = reader.termDocs();> > > 
> TermEnum termEnum = reader.terms (new Term (field, "")); > > > 
> ....................> > >> > >> > >> > >> > >> > > Quoting Fuad Efendi 
> <[EMAIL PROTECTED]>:> > >> > > > I am hoping [new StringIndex (retArray, 
> mterms)] is called only > > > > once> >> > > > per-sort-field and cached 
> somewhere at Lucene;> > > >> > > > theoretically you need multiply number of 
> documents on size of > > > > field> >> > > > (supposing that field contains 
> unique text); you need not tokenize > > > > this field; you need not store 
> TermVector.> > > >> > > > for 2 000 000 documents with simple untokenized 
> text field such as > > > > title of book (256 bytes) you need probably 512 
> 000 000 bytes per > > > > Searcher, and as Mark mentioned you should limit 
> number of > > > > searchers> >> > > > in SOLR.> > > >> > > > So that Xmx512M 
> is definitely not enough even for simple cases...> > > >> > > >> > > > 
> Quoting sundar shankar <[EMAIL PROTECTED]>:> > > >> > > >> I haven't seen the 
> source code before, But I don't know why the > > > >> sorting isn't done 
> after the fetch is done. Wouldn't that make it> >> > > >> more faster. at 
> least in case of field level sorting? I could be> >> > > >> wrong too and the 
> implementation might probably be better. But > > > >> don't know why all of 
> the fields have had to be loaded.> > > >>> > > >>> > > >>> > > >>> > > >>> > 
> > >>> Date: Tue, 22 Jul 2008 14:26:26 -0700> From: [EMAIL PROTECTED]> To:> >> 
> > > >>> solr-user@lucene.apache.org> Subject: Re: Out of memory on Solr> >> > 
> > >>> sorting> > > Ok, after some analysis of FieldCacheImpl:> > - it > > > 
> >>> sorting> > > is> > > >>> supposed that (sorted) Enumeration of "terms" is 
> less than > > > > >>> total number of documents> (so that SOLR uses specific 
> field > > > >>> type> >> > > >>> for sorted searches: > solr.StrField with 
> omitNorms="true")> >> > > >>> It creates int[reader.maxDoc()] array, checks 
> (sorted)> > > >>> Enumeration of > "terms" (untokenized solr.StrField), and > 
> > > >>> populates array with document > Ids.> > > - it also creates > > > >>> 
> array of String> String[] mterms = new > > > >>> String[reader.maxDoc()+1];> 
> > Why> > >> > > >>> do we need that? For 1G> > > >>> document with average 
> term/StrField size > of 100 bytes (which > > > >>> could be unique text!!!) 
> it will create kind of > huge 100Gb > > > >>> cache> > >> > > >>> which is 
> not really needed...> StringIndex value = new > > > >>> StringIndex> > >> > > 
> >>> (retArray, mterms);> > If I understand correctly...> > > >>> StringIndex 
> _must_ be a file in a > filesystem for such a> > case...> > > >>> We create 
> StringIndex, and retrieve top > 10 documents, huge > > > >>> overhead.> > > 
> >> > > >>>> > Quoting Fuad Efendi <[EMAIL PROTECTED]>:> > >> > Ok, what is> > 
> > >>> confusing me is implicit guess that FieldCache contains> > "field"> >> 
> > > >>> and Lucene uses in-memory sort instead of using file-system> >> >> > 
> > >>> "index".......> >> > Array syze: 100Mb (25M x 4 bytes), and it > > > 
> >>> is> >> > > >>> just pointers (4-byte> > integers) to documents in index.> 
> >> >> >> > > >>> org.apache.lucene.search.FieldCacheImpl$10.createValue> > 
> ...> >> >> > > >>> 357: protected Object createValue(IndexReader reader, 
> Object > > > >>> fieldKey)> > 358: throws IOException {> > 359: String field 
> => > > >>> ((String) fieldKey).intern();> > 360: final int[] retArray = new> 
> >> > > >>> int[reader.maxDoc()]; // > > OutOfMemoryError!!!> > ...> > 408:> 
> >> > > >>> StringIndex value = new StringIndex (retArray, mterms);> > 409:> 
> >> > > >>> return value;> > 410: }> > ...> >> > It's very confusing, I > > > 
> >>> don't> >> > > >>> know such internals...> >> >> >>>>> <field name="XXX"> 
> > > >>> type="string" indexed="true" stored="true" > >>>>> > > > >>> 
> termVectors="true"/>> >>>>> The sorting is done based on string> >> > > >>> 
> field.> >> >> > I think Sundar should not use > > > >>> 
> [termVectors="true"]...> >> >> >> > Quoting Mark Miller > > > >>> <[EMAIL 
> PROTECTED]>:> >> >> Hmmm...I think its 32bits an > > > >>> integer with an 
> index entry for each doc, so> >>> >>> >> **25 > > > >>> 000> >> > > >>> 000 x 
> 32 bits = 95.3674316 megabytes**> >>> >> Then you have > > > >>> the> >> > > 
> >>> string array that contains each unique term from your> >> > > > >>> 
> index...you can guess that based on the number of terms in your> >> > > >>> 
> index> >> and an avg length guess.> >>> >> There is some other> > > >>> 
> overhead beyond the sort cache as well, but thats> >> the bulk > > > >>> of> 
> >> > > >>> what it will add. I think my memory may be bad with my> >> > > > 
> >>> original estimate :)> >>> >> Fuad Efendi wrote:> >>> Thank you > > > >>> 
> very much Mark,> >>>> >>> it explains me a lot;> >>>> >>> I am> > > >>> 
> guessing: for 1,000,000 documents with a [string] field of > > > > >>> >>>> 
> >> > > >>> average size 1024 bytes I need 1Gb for single IndexSearcher > > > 
> > >>> >>>> >> > > >>> instance; field-level cache it is used internally by 
> Lucene > > > >>> (can > >>> Lucene manage size if it?); we can't have 1G of > 
> > > >>> such documents > >>> without having 1Tb RAM...> >>>> >>>> >>>> > > > 
> >>> >>> Quoting Mark Miller <[EMAIL PROTECTED]>:> >>>> >>>> > > > >>> Fuad 
> Efendi wrote:> >>>>>> SEVERE: java.lang.OutOfMemoryError:> > > >>> 
> allocLargeObjectOrArray - Object> >>>>>> size: 100767936, Num> > > >>> 
> elements: 25191979> >>>>>>> > > >>>>>>>>> >>>>> I just noticed, this is an 
> exact number of documents > > > >>>>>>>>> >>>>> in> > > >>> index: 25191979> 
> >>>>>> >>>>> (http://www.tokenizer.org/, you > > > >>> can> >> > > >>> sort - 
> click headers Id, > >>>>> [COuntry, Site, Price] in a > > > >>> table; 
> experimental)> >>>>>> >>>>>> >>>>> If array is allocated> >> > > >>> ONLY on 
> new searcher warming up I am > >>>>> _extremely_ happy...> >> > > >>> I had 
> constant OOMs during past month (SUN > >>>>> Java 5).> > > > >>> >>>>> >> > > 
> >>> It is only on warmup - I believe its lazy loaded, so the first> >> > > 
> >>> time a> >>>> search is done (solr does the search as part of > > > >>> 
> warmup I believe) the> >>>> fieldcache is loaded. The > > > >>> underlying> 
> >> > > >>> IndexReader is the key to the>> > > >>>>>>> fieldcache, so until 
> you get a new IndexReader > > > >>>>>>> (SolrSearcher in> > > >>> solr> >>>> 
> world?) the field cache will be good. Keep in mind> > > >>> that as a 
> searcher> >>>> is warming, the other search is still > > > >>> serving, so 
> that will up the ram> >>>> requirements...and since > > > >>> I> >> > > >>> 
> think you can have >1 searchers on> >>>> deck...you get the > > > >>> idea.> 
> >>>>> >>>> As far as the number I gave, thats from a > > > >>> memory made 
> months and months> >>>> ago, so go with what you > > > >>> see.> >>>>>> 
> >>>>>> >>>>>>> > > >>>>>>>> Quoting Fuad Efendi <[EMAIL PROTECTED]>:> >>>>>> 
> >>>>>> I've > > > >>>>>>>> even> > > >>> seen exceptions (posted here) when 
> "sort"-type queries caused>> > > >>>>>>>>> Lucene to allocate 100Mb arrays, 
> here is what happened to> > > >>> me:> >>>>>>> >>>>>> SEVERE: 
> java.lang.OutOfMemoryError:> > > >>> allocLargeObjectOrArray - Object> >>>>>> 
> size: 100767936, Num> > > >>> elements: 25191979> >>>>>> at> >>>>>>> > > >>>> 
> > > org.apache.lucene.search.FieldCacheImpl$10.createValue(FieldCacheImpl.> > 
> > ja> > > va:360) > >>>>>> at> >>>>>>> > > 
> org.apache.lucene.search.FieldCacheImpl$Cache.get(FieldCacheImpl.java:> > > 
> 72> > > ) - it does not happen after I increased from 4096M to 8192M > > > > 
> >>>>>> (JRockit> >>>>>> R27; more intelligent stacktrace, isn't > > > it?)> 
> >>>>>>>> > > >>>>>> Thanks Mark; I didn't know that it happens only once (on 
> > > > >>>>>> warming> > > up a> >>>>>> searcher).> >>>>>>> >>>>>>> >>>>>>> 
> >>>>>> Quoting Mark > > > Miller <[EMAIL PROTECTED]>:> >>>>>>> >>>>>>> 
> Because to sort > > > efficiently, Solr loads the term to sort on for each> 
> >>>>>>> doc in > > > the index into an array. For ints,longs, etc its just an 
> array>> > > >>>>>>> the size of the number of docs in your index (i believe> 
> > > deleted or> >>>>>>> not). For a String its an array to hold each > > > 
> unique string and an array>> > > >>>>>>> of ints indexing into the String 
> array.> >>>>>>>> >>>>>>> So > > > >>>>>>> if> > > you do a sort, and search 
> for something that only gets 1 doc as a>> > > >>>>>>> hit...your still 
> loading up that field cache for every > > > >>>>>>> single> > > doc in> 
> >>>>>>> your index on the first search. With solr, this > > > happens in the> 
> >>>>>>> background as it warms up the searcher. The > > > end story is, you 
> need more> >>>>>>> RAM to accommodate the sort > > > most likely...have you 
> upped your xmx> >>>>>>> setting? I think you > > > can roughly say a 2 
> million doc index would need> >>>>>>> 40-50 MB > > > (depending and rough, 
> but to give an idea) per field your> >>>>>>> > > > sorting on.> >>>>>>>> 
> >>>>>>> - Mark> >>>>>>>> >>>>>>> sundar > > > shankar wrote:> >>>>>>>> Thanks 
> Fuad.> >>>>>>>> But why does just > > > sorting provide an OOM. I > >>>>>>>> 
> executed the query without > > > adding the sort clause it executed > 
> >>>>>>>> perfectly. In fact I > > > even tried remove the maxrows=10 and > 
> >>>>>>>> executed. it came out> fine.> > > Queries with bigger results > 
> >>>>>>>> seems to come out fine too. > > > But> >> > > why just sort of that 
> too > >>>>>>>> just 10 rows??> >>>>>>>> > > > -Sundar>> >> > > >>>>>>>>>> > > 
> >>>>>>>>> >>>>>>>>> >>>>>>>>> >>>>>>>>> Date: Tue, 22 Jul 2008> > > >>>>>>>>> 
> >>>>>>>>> >>>>>>>>> >>>>>>>>> 12:24:35> > > -0700> From: [EMAIL PROTECTED]> > 
> >>>>>>>>> To:> > > solr-user@lucene.apache.org> Subject: RE: Out of > 
> >>>>>>>>> memory > > > on> >> > > Solr sorting> > > >>>>>>>>>> > > 
> org.apache.lucene.search.FieldCacheImpl$10.createValue(FieldCacheImpl.> > > 
> ja va:403)> > - this piece of code do not request Array[100M] (as I > > > 
> seen with > Lucene), it asks only few bytes / Kb for a field...> > > > > > 
> Probably> > > 128 - 512 is not enough; it is also advisable to use equal 
> sizes> > > > -Xms1024M -Xmx1024M> (it minimizes GC frequency, and itensures 
> that > > > 1024M is available at startup)> > OOM happens also with fragmented 
> > > > memory, when application requests big > contigues fragment and GC is > 
> > > unable to optimize; looks like your > application requests a little > > > 
> and memory is not available...> > > Quoting sundar shankar > > > <[EMAIL 
> PROTECTED]>:> > >> >> >> >> From:> > > [EMAIL PROTECTED]> >> To: 
> solr-user@lucene.apache.org> >>> > > Subject: Out of memory on Solr sorting> 
> >> Date: Tue, 22 Jul 2008> > > 19:11:02 +0000> >>> >>> >> Hi,> >> Sorry again 
> fellos. I am not sure > > > whats happening. The day with > >> solr is bad 
> for me I guess. EZMLM > > > didnt let me send any mails this > >> morning. 
> Asked me to confirm > > > subscription and when I did, it said I > >> was 
> already a member. > > > Now my mails are all coming out bad. Sorry > >> for 
> troubling y'all > > > this bad. I hope this mail comes out right.> >> >> > 
> Hi,> > We are > > > developing a product in a agile manner and the current > 
> > > > > implementation has a data of size just about a 800 megs in dev.> > > 
> > > The> >> > > memory allocated to solr on dev (Dual core Linux box) is 
> 128-512.> > > > >>> > > > My config> > =========> >> >> > > <!-- autocommit 
> pending docs if certain criteria are met> > > > > <autoCommit>> > 
> <maxDocs>10000</maxDocs>> > <maxTime>1000</maxTime>> > > > >> >> > > 
> </autoCommit>> > -->> >> > <filterCache> > class="solr.LRUCache"> > > > > 
> size="512"> > initialSize="512"> > autowarmCount="256"/>> >> > > > > 
> <queryResultCache> > class="solr.LRUCache"> > size="512"> > > > > 
> initialSize="512"> > autowarmCount="256"/>> >> > <documentCache> > > > > 
> class="solr.LRUCache"> > size="512"> > initialSize="512"> > > > > 
> autowarmCount="0"/>> >> > > > > 
> <enableLazyFieldLoading>true</enableLazyFieldLoading>> >> >> > My> > > 
> Field>> > > > =======> >> > <fieldType name="autocomplete"> > > > 
> class="solr.TextField">> <analyzer type="index">> > <tokenizer> > > 
> class="solr.KeywordTokenizerFactory"/>> > <filter > > > 
> class="solr.LowerCaseFilterFactory" />> > <filter > > > 
> class="solr.PatternReplaceFilterFactory" > > pattern="([^a-z0-9])"> > > 
> replacement="" replace="all" />> > <filter > > > 
> class="solr.EdgeNGramFilterFactory" > > maxGramSize="100"> > > 
> minGramSize="1" />> > </analyzer>> > <analyzer type="query">> > > > > 
> <tokenizer class="solr.KeywordTokenizerFactory"/>> > <filter > > > 
> class="solr.LowerCaseFilterFactory" />> > <filter > > > 
> class="solr.PatternReplaceFilterFactory" > > pattern="([^a-z0-9])"> > > 
> replacement="" replace="all" />> > <filter > > > 
> class="solr.PatternReplaceFilterFactory" > > pattern="^(.{20})(.*)?"> > > 
> replacement="$1" replace="all" />> > </analyzer>> > </fieldType>> >>> > > >>> 
> > > > Problem> > ======> >> > I execute a query that returns 24 rows of> > > 
> result. I pick 10 out of > > it. I have no problem when I execute > > > 
> this.>> > > > But When I do sort it by a String field that is fetched from 
> this > > > > >> > > > >> > > result. I get an OOM. I am able to execute 
> several> > other queries > > > with no problem. Just having a sort asc clause 
> added > > to the > > > query throws an OOM. Why is that.> > What should I 
> have ideally > > > done. My config on QA is pretty similar > > to the dev box 
> and > > > probably has more data than on dev.> > It didnt throw any OOM 
> during > > > the integration test. The Autocomplete > > is a new field we 
> added > > > recently.> >> > Another point is that the indexing is done with a 
> > > > field of type string> > <field name="XXX" type="string" indexed="true"> 
> >> > > stored="true" > > termVectors="true"/>> >> > and the autocomplete > > 
> > field is a copy field.>> > > >> > The sorting is done based on string 
> field.> >> > Please do > > > >> > lemme> > > know what mistake am I doing?> 
> >> > Regards> > Sundar> >> > P.S: The > > > stack trace of the exception is> 
> >> >> > Caused by:> > > org.apache.solr.client.solrj.SolrServerException: 
> Error > > > > > executing> > > query> > at > >> > > 
> org.apache.solr.client.solrj.request.QueryRequest.process(QueryReque> > > st> 
> > > .j> > > ava:86)> > at > >> > > 
> org.apache.solr.client.solrj.impl.BaseSolrServer.query(BaseSolrServer.> > > 
> ja> > > va:101)> > at > >> > > 
> com.apollo.sisaw.solr.service.AbstractSolrSearchService.makeSolrQuer> > > y( 
> Ab stractSolrSearchService.java:193)> > ... 105 more> > Caused > > > by:> > > 
> org.apache.solr.common.SolrException: Java heap space > >> > > 
> java.lang.OutOfMemoryError: Java heap space> > at > > > > > 
> org.apache.lucene.search.FieldCacheImpl$10.createValue(FieldCacheImpl.> > > 
> ja> > > va:403)> > at> > > 
> org.apache.lucene.search.FieldCacheImpl$Cache.get(FieldCacheImpl.java:> > > 
> 72> > > )> > at > >> > > 
> org.apache.lucene.search.FieldCacheImpl.getStringIndex(FieldCacheImpl.> > > 
> ja> > > va:352)> > at > >> > > 
> org.apache.lucene.search.FieldSortedHitQueue.comparatorString(FieldS> > > or> 
> > > te> > > dHitQueue.java:416)> > at > >> > > 
> org.apache.lucene.search.FieldSortedHitQueue$1.createValue(FieldSort> > > ed> 
> > > Hi> > > tQueue.java:207)> > at> > > 
> org.apache.lucene.search.FieldCacheImpl$Cache.get(FieldCacheImpl.java:> > > 
> 72> > > )> > at > >> > > 
> org.apache.lucene.search.FieldSortedHitQueue.getCachedComparator(Fie> > > ld> 
> > > So> > > rtedHitQueue.java:168)> > at > >> > >> > 
> org.apache.lucene.search.FieldSortedHitQueue.<init>(FieldSortedHitQueue.> > > 
> java:56)> > at > >> > >> > 
> org.apache.solr.search.SolrIndexSearcher.getDocListNC(SolrIndexSearcher.> > > 
> java:907)> > at > >> > > 
> org.apache.solr.search.SolrIndexSearcher.getDocListC(SolrIndexSearch> > > er> 
> > > .j> > > ava:838)> > at > >> > > 
> org.apache.solr.search.SolrIndexSearcher.search(SolrIndexSearcher.ja> > > va> 
> > > :2> > > 69)> > at > >> > >> > 
> org.apache.solr.handler.component.QueryComponent.process(QueryComponent.> > > 
> java:160)> > at > >> > > 
> org.apache.solr.handler.component.SearchHandler.handleRequestBody(Se> > > ar> 
> > > ch> > > Handler.java:156)> > at > >> > > 
> org.apache.solr.handler.RequestHandlerBase.handleRequest(RequestHand> > > le> 
> > > rB> > > ase.java:128)> > at> > > 
> org.apache.solr.core.SolrCore.execute(SolrCore.java:1025)> > at > > > > > 
> org.apache.solr.servlet.SolrDispatchFilter.execute(SolrDispatchFilter.> > > 
> ja> > > va:338)> > at > >> > > 
> org.apache.solr.servlet.SolrDispatchFilter.doFilter(SolrDispatchFilt> > > er> 
> > > .j> > > ava:272)> > at > >> > > 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(App> > > li> 
> > > ca> > > tionFilterChain.java:202)> > at > >> > > 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(Application> > > Fi> 
> > > lt> > > erChain.java:173)> > at > >> > > 
> org.jboss.web.tomcat.filters.ReplyHeaderFilter.doFilter(ReplyHeaderF> > > il> 
> > > te> > > r.java:96)> > at > >> > > 
> org.apache.catalina.core.ApplicationFilterChain.internalDoFilter(App> > > li> 
> > > ca> > > tionFilterChain.java:202)> > at > >> > > 
> org.apache.catalina.core.ApplicationFilterChain.doFilter(Application> > > Fi> 
> > > lt> > > erChain.java:173)> > at > >> > > 
> org.apache.catalina.core.StandardWrapperValve.invoke(StandardWrapper> > > Va> 
> > > lv> > > e.java:213)> > at > >> > > 
> org.apache.catalina.core.StandardContextValve.invoke(StandardContext> > > Va> 
> > > lv> > > e.java:178)> > at > >> > > 
> org.jboss.web.tomcat.security.SecurityAssociationValve.invoke(Securi> > > ty> 
> > > As> > > sociationValve.java:175)> > at > >> > > 
> org.jboss.web.tomcat.security.JaccContextValve.invoke(JaccContextVal> > > ve> 
> > > .j> > > ava:74)> > at > >> > > 
> org.apache.catalina.core.StandardHostValve.invoke(StandardHostValve.> > > ja> 
> > > va> > > :126)> > at > >> > > 
> org.apache.catalina.valves.ErrorReportValve.invoke(ErrorReportValve.> > > ja> 
> > > va> > > :105)> > at > >> > > 
> org.jboss.web.tomcat.tc5.jca.CachedConnectionValve.invoke(CachedConn> > > ec> 
> > > ti> > > onValve.java:156)> > at > >> > >> > 
> org.apache.catalina.core.StandardEngineValve.invoke(StandardEngineValve.> > > 
> java:107)> > at > >> > > 
> org.apache.catalina.connector.CoyoteAdapter.service(CoyoteAdapter.ja> > > va> 
> > > :1> > > 48)> > at> > > 
> org.apache.coyote.http11.Http11Processor.process(Http11Processor.java:> > > 
> 86> > > 9)> >> >> > > 
> _________________________________________________________________> > > > > 
> Wish to Marry Now? Click Here to Register FREE> > > >>>>>>>>> > > > 
> http://www.shaadi.com/registration/user/index.php?ptnr=mhottag>>> > > 
> >>>>>>>>>>> >>>>>>>>>>> >>>>>>>>> > > 
> _________________________________________________________________>> > > 
> >>>>>>>> Missed your favourite programme? Stop surfing TV channels > > > 
> >>>>>>>> and> >> > > >>>>>>>> > start planning your weekend TV viewing with 
> our > > > > >>>>>>>> > >>>>>>>>> > > comprehensive TV Listing> >>>>>>>>> > > 
> http://entertainment.in.msn.com/TV/TVListing.aspx> >>>>>>>>> >>>>>>> > > 
> >>>>>> >>>>>> >>>> >>>> >>>>> > > >>> >> > > >>>>> > > >> 
> _________________________________________________________________> > > >> 
> Wish to Marry Now? Join Shaadi.com FREE!> > > >> 
> http://www.shaadi.com/registration/user/index.php?ptnr=mhottag> > >> > >> > 
> >> > >> > > http://www.bbc.co.uk/> > > This e-mail (and any attachments) is 
> confidential and may contain> > personal views which are not the views of the 
> BBC unless specifically > > stated.> > > If you have received it in error, 
> please delete it from your system.> > > Do not use, copy or disclose the 
> information in any way nor act in> > reliance on it and notify the sender 
> immediately.> > > Please note that the BBC monitors e-mails sent or 
> received.> > > Further communication will signify your consent to this.> > >> 
> >> > _________________________________________________________________> > 
> Missed your favourite programme? Stop surfing TV channels and start > > 
> planning your weekend TV viewing with our comprehensive TV Listing > > 
> http://entertainment.in.msn.com/TV/TVListing.aspx> >> > 
> http://www.bbc.co.uk/> > This e-mail (and any attachments) is confidential 
> and may contain > > personal views which are not the views of the BBC unless 
> specifically> stated.> > If you have received it in error, please delete it 
> from your system.> > Do not use, copy or disclose the information in any way 
> nor act in > > reliance on it and notify the sender immediately.> > Please 
> note that the BBC monitors e-mails sent or received.> > Further communication 
> will signify your consent to this.> >> >> 
_________________________________________________________________
Searching for the best deals on travel? Visit MSN Travel.
http://msn.coxandkings.co.in/cnk/cnk.do

Reply via email to