namenode recovery

2015-03-25 Thread Brian Jeltema
I have a question about a recovery scenario for Hadoop 2.4. I have a small development cluster, no HA configured, that was taken down cleanly, that is, all services were stopped (via Ambari) and all the nodes were then rebooted. However, the reboot of the namenode system failed; that system is

Re: namenode recovery

2015-03-25 Thread Harsh J
Not automatically. You'd need to copy over the VERSION and fsimage* files from your SecondaryNN's checkpoint directory over to the new NameNode's configured name directory to start it back up with the checkpointed data. On Wed, Mar 25, 2015 at 5:40 PM, Brian Jeltema

Re: hadoop namenode recovery

2013-01-17 Thread Michel Segel
MapR was the first vendor to remove the NN as a SPOF. They did this w their 1.0 release when it first came out. The downside is that their release is proprietary and very different in terms of the underlying architecture from Apace based releases. Horton works relies on VMware as a key piece of

Re: hadoop namenode recovery

2013-01-17 Thread Harsh J
Randy, A very slow NFS is certainly troublesome, and slows down the write performance for every edit waiting to get logged to disk (there's a logSync_avg_time metric that could be monitored for this), and therefore a dedicated NFS mount is required if you are unwilling to use the proper HA-HDFS

Re: hadoop namenode recovery

2013-01-15 Thread randy
What happens to the NN and/or performance if there's a problem with the NFS server? Or the network? Thanks, randy On 01/14/2013 11:36 PM, Harsh J wrote: Its very rare to observe an NN crash due to a software bug in production. Most of the times its a hardware fault you should worry about. On

Re: hadoop namenode recovery

2013-01-15 Thread Harsh J
The NFS mount is to be soft-mounted; so if the NFS goes down, the NN ejects it out and continues with the local disk. If auto-restore is configured, it will re-add the NFS if its detected good again later. On Wed, Jan 16, 2013 at 7:04 AM, randy randy...@comcast.net wrote: What happens to the

RE: hadoop namenode recovery

2013-01-15 Thread Rakesh R
http://zookeeper.apache.org/doc/trunk/bookkeeperOverview.html Thanks, Rakesh From: Harsh J [ha...@cloudera.com] Sent: Wednesday, January 16, 2013 9:44 AM To: user@hadoop.apache.org Subject: Re: hadoop namenode recovery The NFS mount is to be soft-mounted; so

hadoop namenode recovery

2013-01-14 Thread Panshul Whisper
Hello, Is there a standard way to prevent the failure of Namenode crash in a Hadoop cluster? or what is the standard or best practice for overcoming the Single point failure problem of Hadoop. I am not ready to take chances on a production server with Hadoop 2.0 Alpha release, which claims to

Re: hadoop namenode recovery

2013-01-14 Thread bejoy . hadoop
the NN back again. Regards Bejoy KS Sent from remote device, Please excuse typos -Original Message- From: Panshul Whisper ouchwhis...@gmail.com Date: Mon, 14 Jan 2013 17:25:08 To: user@hadoop.apache.org Reply-To: user@hadoop.apache.org Subject: hadoop namenode recovery Hello

Re: hadoop namenode recovery

2013-01-14 Thread Panshul Whisper
, Please excuse typos -- *From: * Panshul Whisper ouchwhis...@gmail.com *Date: *Mon, 14 Jan 2013 17:25:08 -0800 *To: *user@hadoop.apache.org *ReplyTo: * user@hadoop.apache.org *Subject: *hadoop namenode recovery Hello, Is there a standard way to prevent

Re: hadoop namenode recovery

2013-01-14 Thread bejoy . hadoop
To: user@hadoop.apache.org; bejoy.had...@gmail.com Subject: Re: hadoop namenode recovery thank you for the reply. Is there a way with which I can configure my cluster to switch to the Secondary Name Node automatically in case of the Primary Name Node failure? When I run my current Hadoop, I see

Re: hadoop namenode recovery

2013-01-14 Thread Panshul Whisper
Bejoy KS Sent from remote device, Please excuse typos -- *From: * Panshul Whisper ouchwhis...@gmail.com *Date: *Mon, 14 Jan 2013 19:04:24 -0800 *To: *user@hadoop.apache.org; bejoy.had...@gmail.com *Subject: *Re: hadoop namenode recovery thank you for the reply

Re: hadoop namenode recovery

2013-01-14 Thread Panshul Whisper
-- *From: * Panshul Whisper ouchwhis...@gmail.com *Date: *Mon, 14 Jan 2013 19:04:24 -0800 *To: *user@hadoop.apache.org; bejoy.had...@gmail.com *Subject: *Re: hadoop namenode recovery thank you for the reply. Is there a way with which I can configure my cluster to switch to the Secondary Name Node

Re: hadoop namenode recovery

2013-01-14 Thread nagarjuna kanamarlapudi
namenode recovery thank you for the reply. Is there a way with which I can configure my cluster to switch to the Secondary Name Node automatically in case of the Primary Name Node failure? When I run my current Hadoop, I see the primary and secondary both Name nodes running. I was wondering

Re: hadoop namenode recovery

2013-01-14 Thread anil gupta
remote device, Please excuse typos -- *From: * Panshul Whisper ouchwhis...@gmail.com *Date: *Mon, 14 Jan 2013 19:04:24 -0800 *To: *user@hadoop.apache.org; bejoy.had...@gmail.com *Subject: *Re: hadoop namenode recovery thank you for the reply. Is there a way

Re: hadoop namenode recovery

2013-01-14 Thread Harsh J
Its very rare to observe an NN crash due to a software bug in production. Most of the times its a hardware fault you should worry about. On 1.x, or any non-HA-carrying release, the best you can get to safeguard against a total loss is to have redundant disk volumes configured, one preferably over

updating datanode config files on namenode recovery

2012-05-01 Thread sumadhur
When a name node goes down and I bring up another machine as the new name node using the back up on a shared folder, do I have to update config files in each data node to point to the new name node and job tracker manually?   Or is there some other way of doing it automatically?   Thanks,

Re: updating datanode config files on namenode recovery

2012-05-01 Thread Bejoy KS
: Tue, 1 May 2012 16:16:14 To: common-user@hadoop.apache.orgcommon-user@hadoop.apache.org Reply-To: common-user@hadoop.apache.org Subject: updating datanode config files on namenode recovery When a name node goes down and I bring up another machine as the new name node using the back up

namenode recovery from checkpoint node

2012-03-23 Thread Juan Pino
Hello, I am trying to recover a namenode that failed, maybe using the checkpoint node. When I start dfs, I get this in the logs (see end of email). I think my metadata is corrupt. I also think this is because hadoop was checkpointing and the machine shut down at the same time. Note that this is a

namenode recovery and hbase

2011-04-21 Thread labtrax
Hello, in Pro Hadoop (Jason Venner, 2009) is written that assuming my namenode crashes all not written blocks are lost. But that there is (might be) a sync feature in 0.20 (p. 120) to avoid this lost. Whats about this feature? Do you know if is existent? And might it be the feature which is