Hi, I would like to ask for your help with the following :

I have two servers on which the elasticsearch is already installed and they 
are configured with the following :
cluster.name: essearchlocal

thrift.port: 9500

index.cache.field.type: soft
index.cache.field.max_size: 20000
index.cache.field.expire: 25m   
index.refresh_interval: 30s                  
indices.fielddata.cache.size: 15%
indices.fielddata.cache.expire: 6h
indices.cache.filter.size: 15%
indices.cache.filter.expire: 6h
indices.memory.index_buffer_size: 70%  

indices.store.throttle.type: merge
indices.store.throttle.max_bytes_per_sec: 25mb  

indices.recovery.concurrent_streams: 3
indices.recovery.max_bytes_per_sec: 20mb
index.translog.flush_threshold_ops: 25000   

index.merge.policy.max_merge_size: 500mb
index.store.compress.stored: true

threadpool.search.type: fixed
threadpool.search.size: 20 
threadpool.search.queue_size: 1000  
 
threadpool.bulk.type: fixed
threadpool.bulk.size: 20
threadpool.bulk.queue_size: 1000
 
threadpool.index.type: fixed
threadpool.index.size: 20
threadpool.index.queue_size: 1000
index.store.type: mmapfs


# cluster settings
# index.merge.policy.max_merged_segment: 1gb   # it hasn't been tested yet
# indices.fielddata.cache.size: 10%   # it hasn't been tested yet
# multicast.enabled: false # it hasn't been tested yet
# discovery.zen.ping_timeout: 5s # it hasn't been tested yet
# discovery.zen.minimum_master_nodes: 2 # it hasn't been tested yet
# discovery.zen.ping.unicast.hosts: [""] # it hasn't been tested yet
# cluster.routing.allocation.cluster_concurrent_rebalance: 2

-----------------------------------------------------------------------

Each server has 24GB of physical memory available, but we are running 
elasticsearch in a standalone mode. 
There are two applications that insert logs into Elasticsearch at the same 
time, and 16Gb of memory is added to Elasticsearch for handling requests. 
There are 8 GB of memory available for the system.

So our problem is, when a search is started the load avarage of the system 
will be 'too high', and the system 
will be unusable, and sometimes we will get SearchRequest Exception 
(java.lang.outofmemoryerror java heap space) .

Could you suggest me a better configuration for solve these problems ?

We can't provide more server for Elasticsearch and we can run only it on 
one node, so we can't run it in cluster mode. 

Are the problems solved, if we put more memory into the server on where the 
Elasticsearch is running ?

I also wanted to ask how can I make the elasticsearch to recover itself 
quickly  ?
We have 150 index ( 750 shard ) , and each index size is about 20 GB.

I tried to give high values to recovery options, but it hasn't effect.


indices.recovery.concurrent_streams: 20
indices.recovery.max_bytes_per_sec: 1500mb



Our servers configuration :
 - Intel(R) Xeon(R) CPU E5345 @ 2.33Ghz (8 core)
 - 24 GB memory DDR2 
 - 4 TB SSHD ( Hybrid ) 

(Elasticsearch version : 0.90.10)


-- 
You received this message because you are subscribed to the Google Groups 
"elasticsearch" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/elasticsearch/d286b4ca-0d84-4d1f-a123-d2628649ac10%40googlegroups.com.
For more options, visit https://groups.google.com/groups/opt_out.

Reply via email to