which distribution are you using?

Regards,
*Stanley Shi,*



On Thu, Jul 24, 2014 at 4:38 AM, andrew touchet <adt...@latech.edu> wrote:

> I should have added this in my first email but I do get an error in the
> data node's log file
>
> '2014-07-12 19:39:58,027 INFO
> org.apache.hadoop.hdfs.server.datanode.DataNode: BlockReport of 0 blocks
> got processed in 1 msecs'
>
>
>
> On Wed, Jul 23, 2014 at 3:18 PM, andrew touchet <adt...@latech.edu> wrote:
>
>> Hello,
>>
>> I am Decommissioning data nodes for an OS upgrade on a HPC cluster .
>> Currently, users can run jobs that use data stored on /hdfs. They are able
>> to access all datanodes/compute nodes except the one being decommissioned.
>>
>> Is this safe to do? Will edited files affect the decommissioning node?
>>
>> I've been adding the nodes to /usr/lib/hadoop-0.20/conf/hosts_exclude and
>> running   'hadoop dfsadmin -refreshNodes' on the name name node.  Then I
>> simply wait for log files to report completion. After upgrade, I simply
>> remove the node from hosts_exlude and start hadoop again on the datanode.
>>
>> Also: Under the namenode web interface I just noticed that the node I
>> have decommissioned previously now has 0 Configured capacity, Used,
>> Remaining memory and is now 100% Used.
>>
>> I used the same /etc/sysconfig/hadoop file from before the upgrade,
>> removed the node from hosts_exclude, and ran '-refreshNodes' afterwards.
>>
>> What steps have I missed in the decommissioning process or while bringing
>> the data node back online?
>>
>>
>>
>>
>

Reply via email to