[ 
https://issues.apache.org/jira/browse/ACCUMULO-2613?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=13957915#comment-13957915
 ] 

Eric Newton commented on ACCUMULO-2613:
---------------------------------------

I don't want to seem argumentative, because I really don't know if using this 
cache for the WAL is a good idea, or not.  But I can think of some issues:

 * hopefully, in your clusters, recovery is an unusual operation
 * WAL has to write to disk to survive power loss, making it a bad candidate 
for RAM-only storage
 * Others have purposefully turned off caching of WAL data to make memory 
available for other things, since reading them at all is unusual

We already know we can improve recovery time by reducing the largest WAL size, 
parallelizing read/sort, and computing a more optimal leaseRecovery timeout.  I 
would strongly suggest a more in-depth look into recovery before even 
experimenting with HDFS caching.

> Take advantage of HDFS caching to improve MTTR
> ----------------------------------------------
>
>                 Key: ACCUMULO-2613
>                 URL: https://issues.apache.org/jira/browse/ACCUMULO-2613
>             Project: Accumulo
>          Issue Type: Improvement
>            Reporter: Sean Busbey
>            Priority: Critical
>              Labels: recovery
>             Fix For: 1.7.0
>
>
> Hadoop 2.3.0 added [HDFS 
> caching|http://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-hdfs/CentralizedCacheManagement.html].
> We should use this for small internal use tables (like !METADATA) and we 
> should probably have a configurable option to use it for tables, with a stern 
> warning that it should only be enabled on small tables that will be 
> frequently used.



--
This message was sent by Atlassian JIRA
(v6.2#6252)

Reply via email to