I was under the impression that the only way to decomission nodes in version 0.12.3 is to specify the nodes in a file and then point
dfs.hosts.exclude to that file.

/Johan

Timothy Chklovski wrote:
which commands are you issuing to decommission the nodes?

On 4/29/07, Johan Oskarsson <[EMAIL PROTECTED]> wrote:

Hi.

I'm trying to decommission 10 datanodes of 35 in our cluster. The
process have been running for a couple of days
but only one node have finished. Perhaps I should have tried to
decommission one at the time?
I was afraid it would lead to unnecessary  transfers as the node being
decommissioned would probably have copied data to other nodes
that I was going to decommission later.

There's no way of seeing how far the process have come?

The logs contain a lot of these:

2007-04-29 16:56:56,411 WARN org.apache.hadoop.fs.FSNamesystem: Not able
to place enough replicas, still in need of 1

Is that related to the decommission process?

/Johan



Reply via email to