Kool, i'm able to turn it off now.
Many thanks to you, Konstantin
Konstantin Shvachko wrote:
You can always manually leave safe mode.
bin/hadoop dfsadmin -safemode leave
And then set replication to whatever or even remove unwanted files
that have bad blocks.
Nguyen Kien Trung wrote:
Thanks Konstantin. I understand now
" At startup the name node accepts data node reports collecting
information about block locations. In order to leave safe mode it
needs to collect a configurable percentage called threshold of
blocks, which satisfy the minimal replication condition. The minimal
replication condition is that each block must have at least
dfs.replication.min replicas. When the threshold is reached the name
node extends safe mode for a configurable amount of time to let the
remaining data nodes to check in before it will start replicating
missing blocks. Then the name node leaves safe mode."
I realized that lots of blocks are missing their replicas. And that
turns into the situation of SAFE mode.
Does it make sense to allow the command ./hadoop dfs -setrep to work
even the name node is in safe mode? Otherwise the name node is idle
forever
Konstantin Shvachko wrote:
You can run "hadoop fsck / " to see how many blocks are missing on
your cluster.
See definition of safe mode here.
http://lucene.apache.org/hadoop/api/org/apache/hadoop/dfs/NameNode.html#setSafeMode(org.apache.hadoop.dfs.FSConstants.SafeModeAction)
--Konstantin
erolagnab wrote:
Hi all,
Just wondering what is the reason causing NameNode is on SafeMode
forever?
I've left my machine running for 2 days and it's still on Safe Mode.
Trung