I have a question about a recovery scenario for Hadoop 2.4.
I have a small development cluster, no HA configured, that was taken down
cleanly,
that is, all services were stopped (via Ambari) and all the nodes were then
rebooted.
However, the reboot of the namenode system failed; that system is
Not automatically. You'd need to copy over the VERSION and fsimage* files
from your SecondaryNN's checkpoint directory over to the new NameNode's
configured name directory to start it back up with the checkpointed data.
On Wed, Mar 25, 2015 at 5:40 PM, Brian Jeltema
MapR was the first vendor to remove the NN as a SPOF.
They did this w their 1.0 release when it first came out. The downside is that
their release is proprietary and very different in terms of the underlying
architecture from Apace based releases.
Horton works relies on VMware as a key piece of
Randy,
A very slow NFS is certainly troublesome, and slows down the write
performance for every edit waiting to get logged to disk (there's a
logSync_avg_time metric that could be monitored for this), and therefore a
dedicated NFS mount is required if you are unwilling to use the proper
HA-HDFS
What happens to the NN and/or performance if there's a problem with the
NFS server? Or the network?
Thanks,
randy
On 01/14/2013 11:36 PM, Harsh J wrote:
Its very rare to observe an NN crash due to a software bug in
production. Most of the times its a hardware fault you should worry about.
On
The NFS mount is to be soft-mounted; so if the NFS goes down, the NN ejects
it out and continues with the local disk. If auto-restore is configured, it
will re-add the NFS if its detected good again later.
On Wed, Jan 16, 2013 at 7:04 AM, randy randy...@comcast.net wrote:
What happens to the
http://zookeeper.apache.org/doc/trunk/bookkeeperOverview.html
Thanks,
Rakesh
From: Harsh J [ha...@cloudera.com]
Sent: Wednesday, January 16, 2013 9:44 AM
To: user@hadoop.apache.org
Subject: Re: hadoop namenode recovery
The NFS mount is to be soft-mounted; so
Hello,
Is there a standard way to prevent the failure of Namenode crash in a
Hadoop cluster?
or what is the standard or best practice for overcoming the Single point
failure problem of Hadoop.
I am not ready to take chances on a production server with Hadoop 2.0 Alpha
release, which claims to
the NN back
again.
Regards
Bejoy KS
Sent from remote device, Please excuse typos
-Original Message-
From: Panshul Whisper ouchwhis...@gmail.com
Date: Mon, 14 Jan 2013 17:25:08
To: user@hadoop.apache.org
Reply-To: user@hadoop.apache.org
Subject: hadoop namenode recovery
Hello
, Please excuse typos
--
*From: * Panshul Whisper ouchwhis...@gmail.com
*Date: *Mon, 14 Jan 2013 17:25:08 -0800
*To: *user@hadoop.apache.org
*ReplyTo: * user@hadoop.apache.org
*Subject: *hadoop namenode recovery
Hello,
Is there a standard way to prevent
To: user@hadoop.apache.org; bejoy.had...@gmail.com
Subject: Re: hadoop namenode recovery
thank you for the reply.
Is there a way with which I can configure my cluster to switch to the
Secondary Name Node automatically in case of the Primary Name Node failure?
When I run my current Hadoop, I see
Bejoy KS
Sent from remote device, Please excuse typos
--
*From: * Panshul Whisper ouchwhis...@gmail.com
*Date: *Mon, 14 Jan 2013 19:04:24 -0800
*To: *user@hadoop.apache.org; bejoy.had...@gmail.com
*Subject: *Re: hadoop namenode recovery
thank you for the reply
--
*From: * Panshul Whisper ouchwhis...@gmail.com
*Date: *Mon, 14 Jan 2013 19:04:24 -0800
*To: *user@hadoop.apache.org; bejoy.had...@gmail.com
*Subject: *Re: hadoop namenode recovery
thank you for the reply.
Is there a way with which I can configure my cluster to switch to the
Secondary Name Node
namenode recovery
thank you for the reply.
Is there a way with which I can configure my cluster to switch to the
Secondary Name Node automatically in case of the Primary Name Node failure?
When I run my current Hadoop, I see the primary and secondary both Name
nodes running. I was wondering
remote device, Please excuse typos
--
*From: * Panshul Whisper ouchwhis...@gmail.com
*Date: *Mon, 14 Jan 2013 19:04:24 -0800
*To: *user@hadoop.apache.org; bejoy.had...@gmail.com
*Subject: *Re: hadoop namenode recovery
thank you for the reply.
Is there a way
Its very rare to observe an NN crash due to a software bug in production.
Most of the times its a hardware fault you should worry about.
On 1.x, or any non-HA-carrying release, the best you can get to safeguard
against a total loss is to have redundant disk volumes configured, one
preferably over
When a name node goes down and I bring up another machine as the new name node
using the back up on a shared folder, do I have to update config files in each
data node to point to the new name node and job tracker manually?
Or is there some other way of doing it automatically?
Thanks,
: Tue, 1 May 2012 16:16:14
To: common-user@hadoop.apache.orgcommon-user@hadoop.apache.org
Reply-To: common-user@hadoop.apache.org
Subject: updating datanode config files on namenode recovery
When a name node goes down and I bring up another machine as the new name node
using the back up
Hello,
I am trying to recover a namenode that failed, maybe using the checkpoint
node.
When I start dfs, I get this in the logs (see end of email).
I think my metadata is corrupt. I also think this is because hadoop was
checkpointing and the machine shut down at the same time.
Note that this is a
Hello,
in Pro Hadoop (Jason Venner, 2009) is written that assuming my namenode
crashes all not written blocks are lost. But that there is (might be) a sync
feature in 0.20 (p. 120) to avoid this lost.
Whats about this feature? Do you know if is existent? And might it be the
feature which is
20 matches
Mail list logo