Thats cool.
On Sep 28, 2015 8:44 PM, "Daniel Watrous" <dwmaill...@gmail.com> wrote:

> After talking with the Vagrant community I decided I was being too clever
> trying to run the datanodes on a separate subnet from the master node. I
> changed my configuration to have all three hosts on the same subnet and
> everything works just as expected.
>
> Thanks for all your help and input.
>
> On Mon, Sep 28, 2015 at 9:07 AM, Daniel Watrous <dwmaill...@gmail.com>
> wrote:
>
>> Vinay,
>>
>> There is no gateway to 51.*. These are IP addresses that I set in my
>> Vagrantfile for virtualbox as part of a private network:
>> master.vm.network "private_network", ip: 192.168.51.4
>>
>> This allows me to spin up all the hosts for my cluster automatically and
>> know that they always have the same IP addresses.
>>
>> From hadoop-data1 (192.168.52.4) I have unrestricted access to
>> hadoop-master (192.168.51.4)
>>
>> hadoop@hadoop-data1:~$ ifconfig
>> eth1      Link encap:Ethernet  HWaddr 08:00:27:b9:55:25
>>           inet addr:192.168.52.4  Bcast:192.168.52.255  Mask:255.255.255.0
>> hadoop@hadoop-data1:~$ ping hadoop-master
>> PING hadoop-master (192.168.51.4) 56(84) bytes of data.
>> 64 bytes from hadoop-master (192.168.51.4): icmp_seq=1 ttl=63 time=3.13 ms
>> 64 bytes from hadoop-master (192.168.51.4): icmp_seq=2 ttl=63 time=2.72 ms
>>
>> I'm not sure I understand exactly what you're asking for, but from the
>> master I can run this
>>
>> vagrant@hadoop-master:~$ sudo netstat -tnulp | grep 54310
>> tcp        0      0 0.0.0.0:54310           0.0.0.0:*
>> LISTEN      22944/java
>>
>> I understand what you're saying about a gateway often existing at that
>> address for a subnet. I'm not familiar enough with Vagrant to answer this
>> right now, but I will put in a question there.
>>
>> I can also change the other two IP addresses to be on the same 51.
>> subnet. I may try that next.
>>
>>
>>
>> On Mon, Sep 28, 2015 at 8:33 AM, Vinayakumar B <vinayakum...@apache.org>
>> wrote:
>>
>>> 192.168.51.1 might be gateway to 51.* subnet right?
>>>
>>> Can you verify whether connections from outside 51 subnet, to 51.4
>>> machine using other subnet IP as remote IP. ?
>>>
>>> You can create any connection, may not be namenode-datanode.
>>>
>>> for ex: Connection from 192.168.52.4 dn to 192.168.51.4 namenode should
>>> result in following, when checked using netstat command in namenode
>>> machine. "netstat -tnulp | grep <NN_RPC_PORT>"
>>>
>>> Output should be something like below
>>>
>>> tcp        0      0   192.168.51.4:54310        192.168.52.4:32567
>>>         LISTEN      -
>>>
>>>
>>> If the Foreign Ip is listing as 192.168.51.1 instead of 192.168.52.4,
>>> then the gateway, is not passing original client IP forward, its
>>> re-creating connections with its own IP. in such case problem will be with
>>> the gateway.
>>>
>>> Its just a guess, reality could be different.
>>>
>>> please check and let me know.
>>>
>>> -Vinay
>>>
>>> On Mon, Sep 28, 2015 at 6:45 PM, Daniel Watrous <dwmaill...@gmail.com>
>>> wrote:
>>>
>>>> Thanks to Namikaze pointing out that I should have sent the namenode
>>>> log as a pastbin
>>>>
>>>> http://pastebin.com/u33bBbgu
>>>>
>>>>
>>>> On Mon, Sep 28, 2015 at 8:02 AM, Daniel Watrous <dwmaill...@gmail.com>
>>>> wrote:
>>>>
>>>>> I have posted the namenode logs here:
>>>>> https://gist.github.com/dwatrous/dafaa7695698f36a5d93
>>>>>
>>>>> Thanks for all the help.
>>>>>
>>>>> On Sun, Sep 27, 2015 at 10:28 AM, Brahma Reddy Battula <
>>>>> brahmareddy.batt...@hotmail.com> wrote:
>>>>>
>>>>>> Thanks for sharing the logs.
>>>>>>
>>>>>> Problem is interesting..can you please post namenode logs and dual IP
>>>>>> configurations(thinking problem with gateway while sending requests from
>>>>>> 52.1 segment to 51.1 segment..)
>>>>>>
>>>>>> Thanks And Regards
>>>>>> Brahma Reddy Battula
>>>>>>
>>>>>>
>>>>>> ------------------------------
>>>>>> Date: Fri, 25 Sep 2015 12:19:00 -0500
>>>>>>
>>>>>> Subject: Re: Problem running example (wrong IP address)
>>>>>> From: dwmaill...@gmail.com
>>>>>> To: user@hadoop.apache.org
>>>>>>
>>>>>> hadoop-master http://pastebin.com/yVF8vCYS
>>>>>> hadoop-data1 http://pastebin.com/xMEdf01e
>>>>>> hadoop-data2 http://pastebin.com/prqd02eZ
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Fri, Sep 25, 2015 at 11:53 AM, Brahma Reddy Battula <
>>>>>> brahmareddy.batt...@hotmail.com> wrote:
>>>>>>
>>>>>> sorry,I am not able to access the logs, could please post in paste
>>>>>> bin or attach the 192.168.51.6( as your query is why different IP)
>>>>>> DN logs and namenode logs here..?
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>> Thanks And Regards
>>>>>> Brahma Reddy Battula
>>>>>>
>>>>>>
>>>>>> ------------------------------
>>>>>> Date: Fri, 25 Sep 2015 11:16:55 -0500
>>>>>> Subject: Re: Problem running example (wrong IP address)
>>>>>> From: dwmaill...@gmail.com
>>>>>> To: user@hadoop.apache.org
>>>>>>
>>>>>>
>>>>>> Brahma,
>>>>>>
>>>>>> Thanks for the reply. I'll keep this conversation here in the user
>>>>>> list. The /etc/hosts file is identical on all three nodes
>>>>>>
>>>>>> hadoop@hadoop-data1:~$ cat /etc/hosts
>>>>>> 127.0.0.1 localhost
>>>>>> 192.168.51.4 hadoop-master
>>>>>> 192.168.52.4 hadoop-data1
>>>>>> 192.168.52.6 hadoop-data2
>>>>>>
>>>>>> hadoop@hadoop-data2:~$ cat /etc/hosts
>>>>>> 127.0.0.1 localhost
>>>>>> 192.168.51.4 hadoop-master
>>>>>> 192.168.52.4 hadoop-data1
>>>>>> 192.168.52.6 hadoop-data2
>>>>>>
>>>>>> hadoop@hadoop-master:~$ cat /etc/hosts
>>>>>> 127.0.0.1 localhost
>>>>>> 192.168.51.4 hadoop-master
>>>>>> 192.168.52.4 hadoop-data1
>>>>>> 192.168.52.6 hadoop-data2
>>>>>>
>>>>>> Here are the startup logs for all three nodes:
>>>>>> https://gist.github.com/dwatrous/7241bb804a9be8f9303f
>>>>>> https://gist.github.com/dwatrous/bcd85cda23d6eca3a68b
>>>>>> https://gist.github.com/dwatrous/922c4f773aded0137fa3
>>>>>>
>>>>>> Thanks for your help.
>>>>>>
>>>>>>
>>>>>> On Fri, Sep 25, 2015 at 10:33 AM, Brahma Reddy Battula <
>>>>>> brahmareddy.batt...@huawei.com> wrote:
>>>>>>
>>>>>> Seems DN started in three machines and failed in
>>>>>> hadoop-data1(192.168.52.4)..
>>>>>>
>>>>>>
>>>>>> 192.168.51.6 : giving IP as 192.168.51.1 
>>>>>> <http://192.168.51.1:50010>...can
>>>>>> you please check /etc/hosts file of 192.168.51.6 (might be
>>>>>> 192.168.51.1 <http://192.168.51.1:50010> is configured in /etc/hosts)
>>>>>>
>>>>>> 192.168.52.4 : datanode startup might be failed ( you can check this
>>>>>> node logs)
>>>>>>
>>>>>> 192.168.51.4 :  <http://192.168.51.4:50010> Datanode starup is
>>>>>> success..which is in master node..
>>>>>>
>>>>>>
>>>>>>
>>>>>> Thanks & Regards
>>>>>>  Brahma Reddy Battula
>>>>>>
>>>>>>
>>>>>>
>>>>>> ------------------------------
>>>>>> *From:* Daniel Watrous [dwmaill...@gmail.com]
>>>>>> *Sent:* Friday, September 25, 2015 8:41 PM
>>>>>> *To:* user@hadoop.apache.org
>>>>>> *Subject:* Re: Problem running example (wrong IP address)
>>>>>>
>>>>>> I'm still stuck on this and posted it to stackoverflow:
>>>>>>
>>>>>> http://stackoverflow.com/questions/32785256/hadoop-datanode-binds-wrong-ip-address
>>>>>>
>>>>>> Thanks,
>>>>>> Daniel
>>>>>>
>>>>>> On Fri, Sep 25, 2015 at 8:28 AM, Daniel Watrous <dwmaill...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>> I could really use some help here. As you can see from the output
>>>>>> below, the two attached datanodes are identified with a non-existent IP
>>>>>> address. Can someone tell me how that gets selected or how to explicitly
>>>>>> set it. Also, why are both datanodes shown under the same name/IP?
>>>>>>
>>>>>> hadoop@hadoop-master:~$ hdfs dfsadmin -report
>>>>>> Configured Capacity: 84482326528 (78.68 GB)
>>>>>> Present Capacity: 75745546240 (70.54 GB)
>>>>>> DFS Remaining: 75744862208 (70.54 GB)
>>>>>> DFS Used: 684032 (668 KB)
>>>>>> DFS Used%: 0.00%
>>>>>> Under replicated blocks: 0
>>>>>> Blocks with corrupt replicas: 0
>>>>>> Missing blocks: 0
>>>>>> Missing blocks (with replication factor 1): 0
>>>>>>
>>>>>> -------------------------------------------------
>>>>>> Live datanodes (2):
>>>>>>
>>>>>> Name: 192.168.51.1:50010 (192.168.51.1)
>>>>>> Hostname: hadoop-data1
>>>>>> Decommission Status : Normal
>>>>>> Configured Capacity: 42241163264 (39.34 GB)
>>>>>> DFS Used: 303104 (296 KB)
>>>>>> Non DFS Used: 4302479360 (4.01 GB)
>>>>>> DFS Remaining: 37938380800 (35.33 GB)
>>>>>> DFS Used%: 0.00%
>>>>>> DFS Remaining%: 89.81%
>>>>>> Configured Cache Capacity: 0 (0 B)
>>>>>> Cache Used: 0 (0 B)
>>>>>> Cache Remaining: 0 (0 B)
>>>>>> Cache Used%: 100.00%
>>>>>> Cache Remaining%: 0.00%
>>>>>> Xceivers: 1
>>>>>> Last contact: Fri Sep 25 13:25:37 UTC 2015
>>>>>>
>>>>>>
>>>>>> Name: 192.168.51.4:50010 (hadoop-master)
>>>>>> Hostname: hadoop-master
>>>>>> Decommission Status : Normal
>>>>>> Configured Capacity: 42241163264 (39.34 GB)
>>>>>> DFS Used: 380928 (372 KB)
>>>>>> Non DFS Used: 4434300928 (4.13 GB)
>>>>>> DFS Remaining: 37806481408 (35.21 GB)
>>>>>> DFS Used%: 0.00%
>>>>>> DFS Remaining%: 89.50%
>>>>>> Configured Cache Capacity: 0 (0 B)
>>>>>> Cache Used: 0 (0 B)
>>>>>> Cache Remaining: 0 (0 B)
>>>>>> Cache Used%: 100.00%
>>>>>> Cache Remaining%: 0.00%
>>>>>> Xceivers: 1
>>>>>> Last contact: Fri Sep 25 13:25:38 UTC 2015
>>>>>>
>>>>>>
>>>>>>
>>>>>> On Thu, Sep 24, 2015 at 5:05 PM, Daniel Watrous <dwmaill...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>> The IP address is clearly wrong, but I'm not sure how it gets set.
>>>>>> Can someone tell me how to configure it to choose a valid IP address?
>>>>>>
>>>>>> On Thu, Sep 24, 2015 at 3:26 PM, Daniel Watrous <dwmaill...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>> I just noticed that both datanodes appear to have chosen that IP
>>>>>> address and bound that port for HDFS communication.
>>>>>>
>>>>>> http://screencast.com/t/OQNbrWFF
>>>>>>
>>>>>> Any idea why this would be? Is there some way to specify which
>>>>>> IP/hostname should be used for that?
>>>>>>
>>>>>> On Thu, Sep 24, 2015 at 3:11 PM, Daniel Watrous <dwmaill...@gmail.com
>>>>>> > wrote:
>>>>>>
>>>>>> When I try to run a map reduce example, I get the following error:
>>>>>>
>>>>>> hadoop@hadoop-master:~$ hadoop jar
>>>>>> /usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.1.jar
>>>>>> pi 10 30
>>>>>> Number of Maps  = 10
>>>>>> Samples per Map = 30
>>>>>> 15/09/24 20:04:28 INFO hdfs.DFSClient: Exception in
>>>>>> createBlockOutputStream
>>>>>> java.io.IOException: Got error, status message , ack with
>>>>>> firstBadLink as 192.168.51.1:50010
>>>>>>         at
>>>>>> org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:140)
>>>>>>         at
>>>>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1334)
>>>>>>         at
>>>>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1237)
>>>>>>         at
>>>>>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:449)
>>>>>> 15/09/24 20:04:28 INFO hdfs.DFSClient: Abandoning
>>>>>> BP-852923283-127.0.1.1-1443119668806:blk_1073741825_1001
>>>>>> 15/09/24 20:04:28 INFO hdfs.DFSClient: Excluding datanode
>>>>>> DatanodeInfoWithStorage[192.168.51.1:50010
>>>>>> ,DS-45f6e06d-752e-41e8-ac25-ca88bce80d00,DISK]
>>>>>> 15/09/24 20:04:28 WARN hdfs.DFSClient: Slow waitForAckedSeqno took
>>>>>> 65357ms (threshold=30000ms)
>>>>>> Wrote input for Map #0
>>>>>>
>>>>>> I'm not sure why it's trying to access 192.168.51.1:50010, which
>>>>>> isn't even a valid IP address in my setup.
>>>>>>
>>>>>> Daniel
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>>
>>>>>
>>>>
>>>
>>
>

Reply via email to