Error in script.
On Wed, May 8, 2013 at 7:11 AM, Chris Embree <[email protected]> wrote: > Your script has an error in it. Please test your script using both IP > Addresses and Names, outside of hadoop. > > > On Wed, May 8, 2013 at 10:01 AM, Mohammad Mustaqeem < > [email protected]> wrote: > >> I have done this and found following error in log - >> >> 2013-05-08 18:53:45,221 WARN org.apache.hadoop.net.ScriptBasedMapping: >> Exception running >> /home/mustaqeem/development/hadoop-2.0.3-alpha/etc/hadoop/rack.sh 127.0.0.1 >> org.apache.hadoop.util.Shell$ExitCodeException: >> /home/mustaqeem/development/hadoop-2.0.3-alpha/etc/hadoop/rack.sh: 8: >> /home/mustaqeem/development/hadoop-2.0.3-alpha/etc/hadoop/rack.sh: Syntax >> error: "(" unexpected (expecting "done") >> >> at org.apache.hadoop.util.Shell.runCommand(Shell.java:202) >> at org.apache.hadoop.util.Shell.run(Shell.java:129) >> at >> org.apache.hadoop.util.Shell$ShellCommandExecutor.execute(Shell.java:322) >> at >> org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.runResolveCommand(ScriptBasedMapping.java:241) >> at >> org.apache.hadoop.net.ScriptBasedMapping$RawScriptBasedMapping.resolve(ScriptBasedMapping.java:179) >> at >> org.apache.hadoop.net.CachedDNSToSwitchMapping.resolve(CachedDNSToSwitchMapping.java:119) >> at >> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager.resolveNetworkLocation(DatanodeManager.java:454) >> at >> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager.registerDatanode(DatanodeManager.java:713) >> at >> org.apache.hadoop.hdfs.server.namenode.FSNamesystem.registerDatanode(FSNamesystem.java:3459) >> at >> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.registerDatanode(NameNodeRpcServer.java:881) >> at >> org.apache.hadoop.hdfs.protocolPB.DatanodeProtocolServerSideTranslatorPB.registerDatanode(DatanodeProtocolServerSideTranslatorPB.java:90) >> at >> org.apache.hadoop.hdfs.protocol.proto.DatanodeProtocolProtos$DatanodeProtocolService$2.callBlockingMethod(DatanodeProtocolProtos.java:18295) >> at >> org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:454) >> at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1014) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1735) >> at org.apache.hadoop.ipc.Server$Handler$1.run(Server.java:1731) >> at java.security.AccessController.doPrivileged(Native Method) >> at javax.security.auth.Subject.doAs(Subject.java:415) >> at >> org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1441) >> at org.apache.hadoop.ipc.Server$Handler.run(Server.java:1729) >> 2013-05-08 18:53:45,223 ERROR >> org.apache.hadoop.hdfs.server.blockmanagement.DatanodeManager: The resolve >> call returned null! Using /default-rack for host [127.0.0.1] >> >> >> >> On Wed, May 8, 2013 at 7:18 PM, Leonid Fedotov >> <[email protected]>wrote: >> >>> You can put this parameter to core-site.xml or hdfs-site.xml >>> It both parsed during the HDFS startup. >>> >>> Leonid >>> >>> >>> On Wed, May 8, 2013 at 6:43 AM, Mohammad Mustaqeem < >>> [email protected]> wrote: >>> >>>> Hello everyone, >>>> I was searching for how to make the hadoop cluster rack-aware and I >>>> find out from here >>>> http://hadoop.apache.org/docs/r2.0.4-alpha/hadoop-project-dist/hadoop-common/ClusterSetup.html#Hadoop_Rack_Awareness >>>> that >>>> we can do this by giving property of "topology.script.file.name". But >>>> here it is not written where to put this >>>> <property> >>>> <name>topology.script.file.name</name> >>>> >>>> <value>/home/mustaqeem/development/hadoop-2.0.3-alpha/etc/hadoop/rack.sh</value> >>>> </property> >>>> >>>> Means in which configuration file. >>>> I am using hadoop-2.0.3-alpha. >>>> >>>> >>>> -- >>>> *With regards ---* >>>> *Mohammad Mustaqeem*, >>>> M.Tech (CSE) >>>> MNNIT Allahabad >>>> 9026604270 >>>> >>>> >>>> >>> >> >> >> -- >> *With regards ---* >> *Mohammad Mustaqeem*, >> M.Tech (CSE) >> MNNIT Allahabad >> 9026604270 >> >> >> >
