RE: How to set Rack Id of DataNodes?

2013-04-15 Thread Vijay Thakorlal
Hi Mohammad, Yes that's correct your rack awareness script takes the IP address of a node and returns the rack name/id. You then just have to ensure the script is executable and referenced (using an absolute path) in the parameter topology.script.file.name in core-site.xml. Regards,

RE: NameNode failure and recovery!

2013-04-03 Thread Vijay Thakorlal
Hi Rahul, The SNN does not act as a backup / standby NameNode in the event of failure. The sole purpose of the Secondary NameNode (or as it’s otherwise / more correctly known as the Checkpoint Node) is to perform checkpointing of the current state of HDFS: The SNN retrieves the

RE: In Compatible clusterIDs

2013-02-20 Thread Vijay Thakorlal
Hi Nagarjuna, What's is in your /etc/hosts file? I think the line in logs where it says DataNodeRegistration(0.0.0.0 [..], should be the hostname or IP of the datanode (124.123.215.187 since you said it's a pseudo-distributed setup) and not 0.0.0.0. By the way are you using the dfs.hosts

RE: getimage failed in Name Node Log

2013-02-15 Thread Vijay Thakorlal
Hi Janesh, I think your SNN may be starting up with the wrong IP, I'm sure the machine parameter should say 192.168.0.101? http://namenode:50070/getimage?putimage=1 http://namenode:50070/getimage?putimage=1port=50090machine=0.0.0.0token= -32:1989419481:0:136084943:1360849122845

RE: Error for Pseudo-distributed Mode

2013-02-12 Thread Vijay Thakorlal
Hi, Could you first try running the example: $ /usr/bin/hadoop jar /usr/lib/hadoop-0.20-mapreduce/hadoop-examples.jar grep input output 'dfs[a-z.]+' Do you receive the same error? Not sure if it's related to a lack of RAM, but as the stack trace shows errors with network timeout (I

RE: Reg Too many fetch-failures Error

2013-02-01 Thread Vijay Thakorlal
Hi Manoj, As you may be aware this means the reduces are unable to fetch intermediate data from TaskTrackers that ran map tasks - you can try: * increasing tasktracker.http.threads so there are more threads to handle fetch requests from reduces. * decreasing mapreduce.reduce.parallel.copies

RE: Maximum Storage size in a Single datanode

2013-01-30 Thread Vijay Thakorlal
Hi Jeba, There are other considerations too, for example, if a single node holds 1 PB of data and if it were to die this would cause a significant amount of traffic as NameNode arranges for new replicas to be created. Vijay From: Bertrand Dechoux [mailto:decho...@gmail.com] Sent: 30