Can you check if the ports are opened by running telnet command.
Run below command from source machine to destination machine and check if
this help

$telnet <IP address> <port number>
Ex: $telnet 192.168.1.60 9000


Let's Hadooping....!

Bests
Sidharth
Mob: +91 8197555599
LinkedIn: www.linkedin.com/in/sidharthkumar2792

On 28-Apr-2017 10:32 AM, "Bhushan Pathak" <bhushan.patha...@gmail.com>
wrote:

> Hello All,
>
> 1. The slave & master can ping each other as well as use passwordless SSH
> 2. The actual IP starts with 10.x.x.x, I have put in the config file as I
> cannot share  the actual IP
> 3. The namenode is formatted. I executed the 'hdfs namenode -format' again
> just to rule out the possibility
> 4. I did not configure anything in the master file. I don;t think Hadoop
> 2.7.3 has a master file to be configured
> 5. The netstat command [sudo netstat -tulpn | grep '51150' ] does not
> give any output.
>
> Even if I change  the port number to a different one, say 52220, 50000, I
> still get the same error.
>
> Thanks
> Bhushan Pathak
>
> Thanks
> Bhushan Pathak
>
> On Fri, Apr 28, 2017 at 7:52 AM, Lei Cao <charlie.c...@hotmail.com> wrote:
>
>> Hi Mr. Bhushan,
>>
>> Have you tried to format namenode?
>> Here's the command:
>> hdfs namenode -format
>>
>> I've encountered such problem as namenode cannot be started. This command
>> line easily fixed my problem.
>>
>> Hope this can help you.
>>
>> Sincerely,
>> Lei Cao
>>
>>
>> On Apr 27, 2017, at 12:09, Brahma Reddy Battula <
>> brahmareddy.batt...@huawei.com> wrote:
>>
>> *Please check “hostname –i” .*
>>
>>
>>
>>
>>
>> *1)      **What’s configured in the “master” file.(you shared only slave
>> file).?*
>>
>>
>>
>> *2)      **Can you able to “ping master”?*
>>
>>
>>
>> *3)      **Can you configure like this check once..?*
>>
>> *                1.1.1.1 master*
>>
>>
>>
>>
>>
>> Regards
>>
>> Brahma Reddy Battula
>>
>>
>>
>> *From:* Bhushan Pathak [mailto:bhushan.patha...@gmail.com
>> <bhushan.patha...@gmail.com>]
>> *Sent:* 27 April 2017 18:16
>> *To:* Brahma Reddy Battula
>> *Cc:* user@hadoop.apache.org
>> *Subject:* Re: Hadoop 2.7.3 cluster namenode not starting
>>
>>
>>
>> Some additional info -
>>
>> OS: CentOS 7
>>
>> RAM: 8GB
>>
>>
>>
>> Thanks
>>
>> Bhushan Pathak
>>
>>
>> Thanks
>>
>> Bhushan Pathak
>>
>>
>>
>> On Thu, Apr 27, 2017 at 3:34 PM, Bhushan Pathak <
>> bhushan.patha...@gmail.com> wrote:
>>
>> Yes, I'm running the command on the master node.
>>
>>
>>
>> Attached are the config files & the hosts file. I have updated the IP
>> address only as per company policy, so that original IP addresses are not
>> shared.
>>
>>
>>
>> The same config files & hosts file exist on all 3 nodes.
>>
>>
>>
>> Thanks
>>
>> Bhushan Pathak
>>
>>
>> Thanks
>>
>> Bhushan Pathak
>>
>>
>>
>> On Thu, Apr 27, 2017 at 3:02 PM, Brahma Reddy Battula <
>> brahmareddy.batt...@huawei.com> wrote:
>>
>> Are you sure that you are starting in same machine (master)..?
>>
>>
>>
>> Please share “/etc/hosts” and configuration files..
>>
>>
>>
>>
>>
>> Regards
>>
>> Brahma Reddy Battula
>>
>>
>>
>> *From:* Bhushan Pathak [mailto:bhushan.patha...@gmail.com]
>> *Sent:* 27 April 2017 17:18
>> *To:* user@hadoop.apache.org
>> *Subject:* Fwd: Hadoop 2.7.3 cluster namenode not starting
>>
>>
>>
>> Hello
>>
>>
>>
>> I have a 3-node cluster where I have installed hadoop 2.7.3. I have
>> updated core-site.xml, mapred-site.xml, slaves, hdfs-site.xml,
>> yarn-site.xml, hadoop-env.sh files with basic settings on all 3 nodes.
>>
>>
>>
>> When I execute start-dfs.sh on the master node, the namenode does not
>> start. The logs contain the following error -
>>
>> 2017-04-27 14:17:57,166 ERROR 
>> org.apache.hadoop.hdfs.server.namenode.NameNode:
>> Failed to start namenode.
>>
>> java.net.BindException: Problem binding to [master:51150]
>> java.net.BindException: Cannot assign requested address; For more details
>> see:  http://wiki.apache.org/hadoop/BindException
>>
>>         at sun.reflect.NativeConstructorAccessorImpl.newInstance0(Native
>> Method)
>>
>>         at sun.reflect.NativeConstructorAccessorImpl.newInstance(Native
>> ConstructorAccessorImpl.java:62)
>>
>>         at sun.reflect.DelegatingConstructorAccessorImpl.newInstance(De
>> legatingConstructorAccessorImpl.java:45)
>>
>>         at java.lang.reflect.Constructor.newInstance(Constructor.java:4
>> 23)
>>
>>         at org.apache.hadoop.net.NetUtils.wrapWithMessage(NetUtils.
>> java:792)
>>
>>         at org.apache.hadoop.net.NetUtils.wrapException(NetUtils.java:
>> 721)
>>
>>         at org.apache.hadoop.ipc.Server.bind(Server.java:425)
>>
>>         at org.apache.hadoop.ipc.Server$Listener.<init>(Server.java:574)
>>
>>         at org.apache.hadoop.ipc.Server.<init>(Server.java:2215)
>>
>>         at org.apache.hadoop.ipc.RPC$Server.<init>(RPC.java:951)
>>
>>         at org.apache.hadoop.ipc.ProtobufRpcEngine$Server.<init>(
>> ProtobufRpcEngine.java:534)
>>
>>         at org.apache.hadoop.ipc.ProtobufRpcEngine.getServer(ProtobufRp
>> cEngine.java:509)
>>
>>         at org.apache.hadoop.ipc.RPC$Builder.build(RPC.java:796)
>>
>>         at org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.<in
>> it>(NameNodeRpcServer.java:345)
>>
>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcSer
>> ver(NameNode.java:674)
>>
>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(N
>> ameNode.java:647)
>>
>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameN
>> ode.java:812)
>>
>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.<init>(NameN
>> ode.java:796)
>>
>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNo
>> de(NameNode.java:1493)
>>
>>         at org.apache.hadoop.hdfs.server.namenode.NameNode.main(NameNod
>> e.java:1559)
>>
>> Caused by: java.net.BindException: Cannot assign requested address
>>
>>         at sun.nio.ch.Net.bind0(Native Method)
>>
>>         at sun.nio.ch.Net.bind(Net.java:433)
>>
>>         at sun.nio.ch.Net.bind(Net.java:425)
>>
>>         at sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelI
>> mpl.java:223)
>>
>>         at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.
>> java:74)
>>
>>         at org.apache.hadoop.ipc.Server.bind(Server.java:408)
>>
>>         ... 13 more
>>
>> 2017-04-27 14:17:57,171 INFO org.apache.hadoop.util.ExitUtil: Exiting
>> with status 1
>>
>> 2017-04-27 14:17:57,176 INFO org.apache.hadoop.hdfs.server.namenode.NameNode:
>> SHUTDOWN_MSG:
>>
>> /************************************************************
>>
>> SHUTDOWN_MSG: Shutting down NameNode at master/1.1.1.1
>>
>> ************************************************************/
>>
>>
>>
>>
>>
>>
>>
>> I have changed the port number multiple times, every time I get the same
>> error. How do I get past this?
>>
>>
>>
>>
>>
>>
>>
>> Thanks
>>
>> Bhushan Pathak
>>
>>
>>
>>
>>
>>
>>
>>
>

Reply via email to