Thank you very much for the answer, Harsh J. Your suggestion totally
makes sense, and I can do what I wanted :)

Best,

On Sun, May 20, 2012 at 10:53 PM, Harsh J <ha...@cloudera.com> wrote:
> Kyungyong Lee,
>
> One way: This may be possible to do if you inflate the
> "dfs.datanode.du.reserved" property on the specific DataNode to a very
> large bytes value (> maximum volume size). This way your NN will still
> consider the DN as a valid one that carries readable blocks, but when
> writing files this DN will never be selected due to its
> false-lack-of-space report.
>
> On Mon, May 21, 2012 at 12:37 AM, Kyungyong Lee <iamkyungy...@gmail.com> 
> wrote:
>> Hello,
>>
>> I would like to ask if I can do the following. Assuming that I have a
>> datanode, i.e., dn1, which already contains some useful blocks. Here,
>> I do not want to save new data blocks to the datanode, but I still
>> want to use the blocks that already exist in the datanode (dn1).
>> I considered to use exclude file (dfs.hosts.exclude). However, if I
>> add "dn1" to the exclude file list, I cannot use blocks that are
>> already contained in dn1. If it is right, can you please give me some
>> guidances to do what I'm thinking using HDFS?
>>
>> Thanks,
>
>
>
> --
> Harsh J

Reply via email to