You need to give more heap memory to the process. 800 meg is tiny for that
amount of data.

Regards,
Mark Walkom

Infrastructure Engineer
Campaign Monitor
email: [email protected]
web: www.campaignmonitor.com


On 25 May 2014 16:26, Umut Yerci <[email protected]> wrote:

>    Hi,
>
>    I am working on centos5 and I run elasticsearch with version 1.0.0
> with *-Xms808m -Xmx808m -Xss256k*parameters. There are 17 index and total
> 30200583 docs. Each index's docs count between 1000000 and 2000000. I
> create request query like ( each index have date field );
>
>
> {
>   "query": {
>     "bool": {
>       "must": [
>         {
>           "range": {
>             "date": {
>               "to": "2014-06-01 14:14:00",
>               "from": "2014-04-01 00:00:00"
>             }
>           }
>         }
>       ],
>       "should": [],
>       "must_not": [],
>       "minimum_number_should_match": 1
>     }
>   },
>   "from": 0,
>   "size": "50"
> }
>
>
> It give response;
>
> {
>    took: 5903
>    timed_out: false
>    _shards: {
>       total: 17
>       successful: 17
>       failed: 0
>    },
>    hits: {
>    total: 30200583
> ...
> ...
> ...}
>
>
>
> However when I send query on elasticsearch-head tool for last 50 rows like;
>
> {
>   ...
>   ...
>   ...
>   "from": 30200533,
>   "size": "50"
> }
>
> It does not give a response and throw exception like;
>
> ava.lang.OutOfMemoryError: Java heap space
>         at org.apache.lucene.store.DataOutput.copyBytes(DataOutput.java:247)
>         at org.apache.lucene.store.Directory.copy(Directory.java:186)
>         at 
> org.elasticsearch.index.store.Store$StoreDirectory.copy(Store.java:348)
>         at 
> org.apache.lucene.store.TrackingDirectoryWrapper.copy(TrackingDirectoryWrapper.java:50)
>         at 
> org.apache.lucene.index.IndexWriter.createCompoundFile(IndexWriter.java:4596)
>         at 
> org.apache.lucene.index.DocumentsWriterPerThread.sealFlushedSegment(DocumentsWriterPerThread.java:535)
>         at 
> org.apache.lucene.index.DocumentsWriterPerThread.flush(DocumentsWriterPerThread.java:502)
>         at 
> org.apache.lucene.index.DocumentsWriter.doFlush(DocumentsWriter.java:506)
>         at 
> org.apache.lucene.index.DocumentsWriter.flushAllThreads(DocumentsWriter.java:616)
>         at org.apache.lucene.index.IndexWriter.getReader(IndexWriter.java:370)
>         at 
> org.apache.lucene.index.StandardDirectoryReader.doOpenFromWriter(StandardDirectoryReader.java:285)
>         at 
> org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:260)
>         at 
> org.apache.lucene.index.StandardDirectoryReader.doOpenIfChanged(StandardDirectoryReader.java:250)
>         at 
> org.apache.lucene.index.DirectoryReader.openIfChanged(DirectoryReader.java:170)
>         at 
> org.apache.lucene.search.XSearcherManager.refreshIfNeeded(XSearcherManager.java:123)
>         at 
> org.apache.lucene.search.XSearcherManager.refreshIfNeeded(XSearcherManager.java:59)
>         at 
> org.apache.lucene.search.XReferenceManager.doMaybeRefresh(XReferenceManager.java:180)
>         at 
> org.apache.lucene.search.XReferenceManager.maybeRefresh(XReferenceManager.java:229)
>         at 
> org.elasticsearch.index.engine.internal.InternalEngine.refresh(InternalEngine.java:730)
>         at 
> org.elasticsearch.index.shard.service.InternalIndexShard.refresh(InternalIndexShard.java:477)
>         at 
> org.elasticsearch.index.shard.service.InternalIndexShard$EngineRefresher$1.run(InternalIndexShard.java:924)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.runTask(ThreadPoolExecutor.java:886)
>         at 
> java.util.concurrent.ThreadPoolExecutor$Worker.run(ThreadPoolExecutor.java:908)
>         at java.lang.Thread.run(Thread.java:619)
>
> What is the problem? Is it not enough java heap space or does my query
> cause this heap space error?
>
> I asked same question in stackoverflow. Soutions ,which recommended in
> stackoverflow, are not applicable for me. Anyone can give another solutions
> for this problem?
>
>  --
> You received this message because you are subscribed to the Google Groups
> "elasticsearch" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/elasticsearch/08a3e138-495b-4150-8520-06cfab04c31d%40googlegroups.com<https://groups.google.com/d/msgid/elasticsearch/08a3e138-495b-4150-8520-06cfab04c31d%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
> For more options, visit https://groups.google.com/d/optout.
>

-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/CAEM624YYoYQsbKwZLjZAR7J3QB1w9Od_0EeQTn2_wJxvJjzXSw%40mail.gmail.com.
For more options, visit https://groups.google.com/d/optout.

Reply via email to