[ 
https://issues.apache.org/jira/browse/HADOOP-442?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wendy Chien updated HADOOP-442:
-------------------------------

    Attachment: hadoop-442-11.patch

Thanks for looking over the patch, Dhruba!  I updated it to incorporate 
Dhruba's comments.

1. TestDecommission.waitNodeState waits for 1 second now. 
2. I do mean to check for DECOMMISSION_INRPOGRESS to make sure the decommission 
began, but I want to stop it before it finishes so I can test commissioning a 
node works too.  
3. refreshNodes now returns void.
4. UnregisteredDatanodeException was already there, but I also added 
DisallowedDatanodeException to that clause.  I'm inclined to leave them 
together since they are similar.
6. added synchronized to verifyNodeRegistration, and removed it from 
start/stopDecommission.
7. removed the new code from pendingTransfers
8. I moved verifyNodeShutdown to FSNamesystem.  


> slaves file should include an 'exclude' section, to prevent "bad" datanodes 
> and tasktrackers from disrupting  a cluster
> -----------------------------------------------------------------------------------------------------------------------
>
>                 Key: HADOOP-442
>                 URL: https://issues.apache.org/jira/browse/HADOOP-442
>             Project: Hadoop
>          Issue Type: Bug
>          Components: conf
>            Reporter: Yoram Arnon
>         Assigned To: Wendy Chien
>         Attachments: hadoop-442-10.patch, hadoop-442-11.patch, 
> hadoop-442-8.patch
>
>
> I recently had a few nodes go bad, such that they were inaccessible to ssh, 
> but were still running their java processes.
> tasks that executed on them were failing, causing jobs to fail.
> I couldn't stop the java processes, because of the ssh issue, so I was 
> helpless until I could actually power down these nodes.
> restarting the cluster doesn't help, even when removing the bad nodes from 
> the slaves file - they just reconnect and are accepted.
> while we plan to avoid tasks from launching on the same nodes over and over, 
> what I'd like is to be able to prevent rogue processes from connecting to the 
> masters.
> Ideally, the slaves file will contain an 'exclude' section, which will list 
> nodes that shouldn't be accessed, and should be ignored if they try to 
> connect. That would also help in configuring the slaves file for a large 
> cluster - I'd list the full range of machines in the cluster, then list the 
> ones that are down in the 'exclude' section

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.

Reply via email to