Alan,

If your d-hosts are paging heavily, they may be at or near their capacity limits. If so, you should think about adding more d-hosts. You should measure the current Bytes/fragment utilization, and project growth for the next 12-24 months.

You can calculate Bytes/fragment from the Forest status page or the xdmp:forest-status() API call. As a rule of thumb, this "forest memory" shouldn't be more than 1/3 of main memory (ie, 10.67-GB for 32-GB). Of the rest, another 1/3 will probably be group-level caches, and the final 1/3 will be used by in-memory stands, queries, merges, and the OS. That's just a rule of thumb, but it works fairly well for me.

In general, you should cap forest size at the lesser of about 200-GB on-disk, or when total in-memory size approaches 1/3 of main memory. The number of forests per host is generally dictated by this, plus a recommendation of 1 Forest per 2 CPU-cores. If you decide that a given forest is close to its limit, you can set it to be "delete-only" in the forest configuration. If a forest is too large, you may want to physically move it to another host, or perhaps use Corb or XQSync to migrate documents from one forest to another.

All of the above is somewhat simplified: running a TB-scale database calls for careful planning and constant monitoring. There is also some useful leverage to be gained from system tuning, but all that gets a bit involved for an email thread.

Some of these topics are discussed in the docs:

http://developer.marklogic.com/pubs/4.1/books/cluster.pdf

http://developer.marklogic.com/pubs/4.1/books/performance.pdf

http://developer.marklogic.com/pubs/4.1/books/admin.pdf

-- Mike

On 2009-06-29 18:19, Alan Darnell wrote:
We have a cluster of 6 servers — two e-nodes and 4 d-nodes.  All six machines 
are configured with 32GB of RAM.  The 2 e-nodes behave reasonably, eating up 
about 16GB of RAM.  All four d-nodes eat up 32GB of RAM quickly after MarkLogic 
is started. Two of the machines start eating into swap after about a day of use 
and this starts to slow the performance of the whole cluster.  What causes 
these machines to eat up swap and how can we rebalance forests so that memory 
is consumed where it is free?

Alan



_______________________________________________
General mailing list
[email protected]
http://xqzone.com/mailman/listinfo/general

Reply via email to