Maybe grep for
2011-02-25 18:47:05,564 INFO
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Decommission complete for
node 102.1.1.1:50010
In the namenode log to see if decommission is completed?
I remember a similar problem was reported just a few days ago ( in attachment)
by James Litton. According to James, no block was missing after the node was
removed, however, it was unclear when/if the decommission process was finished.
From: Rita [mailto:rmorgan...@gmail.com]
Sent: Thursday, February 24, 2011 5:59 AM
To: hdfs-user@hadoop.apache.org
Cc: Harsh J
Subject: Re: datanode down alert
Thanks for the response.
I am asking because of the following issue,
https://issues.apache.org/jira/browse/HDFS-694
When I decommission a data node it shows up in the "Dead list" on the webgui
coincidentally it also shows up in the "Live" nodes.
I want to make sure this node is fully decommissioned before I remove it from
the cluster.
On Tue, Feb 15, 2011 at 9:13 AM, Harsh J
<qwertyman...@gmail.com<mailto:qwertyman...@gmail.com>> wrote:
I know of a way but I do not know for sure if that is what you're looking for:
DFSClient.datanodeReport(DataNodeReportType.DEAD) should give you a
list of all DEAD data nodes as per the NameNode.
Although I believe the reports cost a lot, so do not do it often (rpcs the NN).
On Tue, Feb 15, 2011 at 6:51 PM, Rita
<rmorgan...@gmail.com<mailto:rmorgan...@gmail.com>> wrote:
> Is there a programmatic way to determine if a datanode is down?
>
>
>
> --
> --- Get your facts first, then you can distort them as you please.--
>
--
Harsh J
www.harshj.com<http://www.harshj.com>
--
--- Get your facts first, then you can distort them as you please.--
--- Begin Message ---
Tanping,
Thank you for the reply. The nodes were marked as “decommissioning in
progress.” My concern was that they never reached a decommissioned state. I
have since begun taking the nodes down and have not have any data blocks
missing, so I suspect the process worked. It was just unclear when the process
was complete.
James
On 2/17/11 12:59 PM, "Tanping Wang" <tanp...@yahoo-inc.com> wrote:
James,
After issuing a command to decommission a node, you should at least be able to
see the following log messages in the namenode logs
Setting the excludes files to some_file_contains_decommissioing_hostname
Refreshing hosts (include/exclude) list
If you do not see these log messages, maybe you want to check
1) Weather you have set
<property>
<name>dfs.hosts.exclude</name>
<value>some_file_contains_decommissioing_hostname</value>
</property>
In hdfs-site.xml
2) If you have your this decommissioning hostname file in place.
Regards,
Tanping
From: James Litton [mailto:james.lit...@chacha.com]
Sent: Friday, February 11, 2011 1:10 PM
To: hdfs-user@hadoop.apache.org
Subject: Decommissioning Nodes
While decommissioning nodes I am seeing the following in my namenode logs:
2011-02-11 21:05:16,290 WARN
org.apache.hadoop.hdfs.server.namenode.FSNamesystem: Not able to place enough
replicas, still in need of 5
I haven’t seen any progress of decommissioning nodes in several days. I have 12
total nodes with 6 being decommissioned and a replication factor of 3. How long
should I expect this to take? Is there a way to force this to move forward?
Thank you.
--- End Message ---