Hi Eric,
Please check my recent experience with the dataloss issue.

http://lucene.472066.n3.nabble.com/Blocks-are-getting-corrupted-under-very-high-load-tt3527403.html#a3532671

http://lucene.472066.n3.nabble.com/Update-RE-Blocks-are-getting-corrupted-under-very-high-load-tt3537420.html#none

 Performance degradtion is 9% with 10 Cleinet Threads , 8 node cluster,  24TB 
each machine. It is always safe to set this flag if your machine will get good 
amount of load. 

Regards,
Uma
________________________________________
From: Eric Hwang [ehw...@fb.com]
Sent: Friday, December 02, 2011 9:56 AM
To: Hairong Kuang
Cc: Zheng Shao; hdfs-dev@hadoop.apache.org
Subject: Re: another HDFS configuration for scribeH

Hi Hairong,

What is the risk for this change? How much more testing do you think will be 
needed?

-Eric

From: Hairong Kuang <hair...@fb.com<mailto:hair...@fb.com>>
Date: Thu, 1 Dec 2011 20:22:10 -0800
To: Internal Use <ehw...@fb.com<mailto:ehw...@fb.com>>
Cc: Zheng Shao <zs...@fb.com<mailto:zs...@fb.com>>, 
"hdfs-dev@hadoop.apache.org<mailto:hdfs-dev@hadoop.apache.org>" 
<hdfs-dev@hadoop.apache.org<mailto:hdfs-dev@hadoop.apache.org>>
Subject: another HDFS configuration for scribeH

Hi Eric,

I was debugging a bizar data corruption case in the silver cluster today and 
realized that there is a very important configuration that scribeH cluster 
should set. Could you please set dfs.datanode.synconclose to be true in ScribeH 
for next week's push? This is will guarantee that block data get persisted to 
disk on close, so preventing data loss when datanodes get rebooted.

Thanks,
Hairong

Reply via email to