192.168.51.1 might be gateway to 51.* subnet right? Can you verify whether connections from outside 51 subnet, to 51.4 machine using other subnet IP as remote IP. ?
You can create any connection, may not be namenode-datanode. for ex: Connection from 192.168.52.4 dn to 192.168.51.4 namenode should result in following, when checked using netstat command in namenode machine. "netstat -tnulp | grep <NN_RPC_PORT>" Output should be something like below tcp 0 0 192.168.51.4:54310 192.168.52.4:32567 LISTEN - If the Foreign Ip is listing as 192.168.51.1 instead of 192.168.52.4, then the gateway, is not passing original client IP forward, its re-creating connections with its own IP. in such case problem will be with the gateway. Its just a guess, reality could be different. please check and let me know. -Vinay On Mon, Sep 28, 2015 at 6:45 PM, Daniel Watrous <[email protected]> wrote: > Thanks to Namikaze pointing out that I should have sent the namenode log > as a pastbin > > http://pastebin.com/u33bBbgu > > > On Mon, Sep 28, 2015 at 8:02 AM, Daniel Watrous <[email protected]> > wrote: > >> I have posted the namenode logs here: >> https://gist.github.com/dwatrous/dafaa7695698f36a5d93 >> >> Thanks for all the help. >> >> On Sun, Sep 27, 2015 at 10:28 AM, Brahma Reddy Battula < >> [email protected]> wrote: >> >>> Thanks for sharing the logs. >>> >>> Problem is interesting..can you please post namenode logs and dual IP >>> configurations(thinking problem with gateway while sending requests from >>> 52.1 segment to 51.1 segment..) >>> >>> Thanks And Regards >>> Brahma Reddy Battula >>> >>> >>> ------------------------------ >>> Date: Fri, 25 Sep 2015 12:19:00 -0500 >>> >>> Subject: Re: Problem running example (wrong IP address) >>> From: [email protected] >>> To: [email protected] >>> >>> hadoop-master http://pastebin.com/yVF8vCYS >>> hadoop-data1 http://pastebin.com/xMEdf01e >>> hadoop-data2 http://pastebin.com/prqd02eZ >>> >>> >>> >>> On Fri, Sep 25, 2015 at 11:53 AM, Brahma Reddy Battula < >>> [email protected]> wrote: >>> >>> sorry,I am not able to access the logs, could please post in paste bin >>> or attach the 192.168.51.6( as your query is why different IP) DN logs >>> and namenode logs here..? >>> >>> >>> >>> >>> Thanks And Regards >>> Brahma Reddy Battula >>> >>> >>> ------------------------------ >>> Date: Fri, 25 Sep 2015 11:16:55 -0500 >>> Subject: Re: Problem running example (wrong IP address) >>> From: [email protected] >>> To: [email protected] >>> >>> >>> Brahma, >>> >>> Thanks for the reply. I'll keep this conversation here in the user list. >>> The /etc/hosts file is identical on all three nodes >>> >>> hadoop@hadoop-data1:~$ cat /etc/hosts >>> 127.0.0.1 localhost >>> 192.168.51.4 hadoop-master >>> 192.168.52.4 hadoop-data1 >>> 192.168.52.6 hadoop-data2 >>> >>> hadoop@hadoop-data2:~$ cat /etc/hosts >>> 127.0.0.1 localhost >>> 192.168.51.4 hadoop-master >>> 192.168.52.4 hadoop-data1 >>> 192.168.52.6 hadoop-data2 >>> >>> hadoop@hadoop-master:~$ cat /etc/hosts >>> 127.0.0.1 localhost >>> 192.168.51.4 hadoop-master >>> 192.168.52.4 hadoop-data1 >>> 192.168.52.6 hadoop-data2 >>> >>> Here are the startup logs for all three nodes: >>> https://gist.github.com/dwatrous/7241bb804a9be8f9303f >>> https://gist.github.com/dwatrous/bcd85cda23d6eca3a68b >>> https://gist.github.com/dwatrous/922c4f773aded0137fa3 >>> >>> Thanks for your help. >>> >>> >>> On Fri, Sep 25, 2015 at 10:33 AM, Brahma Reddy Battula < >>> [email protected]> wrote: >>> >>> Seems DN started in three machines and failed in >>> hadoop-data1(192.168.52.4).. >>> >>> >>> 192.168.51.6 : giving IP as 192.168.51.1 <http://192.168.51.1:50010>...can >>> you please check /etc/hosts file of 192.168.51.6 (might be 192.168.51.1 >>> <http://192.168.51.1:50010> is configured in /etc/hosts) >>> >>> 192.168.52.4 : datanode startup might be failed ( you can check this >>> node logs) >>> >>> 192.168.51.4 : <http://192.168.51.4:50010> Datanode starup is >>> success..which is in master node.. >>> >>> >>> >>> Thanks & Regards >>> Brahma Reddy Battula >>> >>> >>> >>> ------------------------------ >>> *From:* Daniel Watrous [[email protected]] >>> *Sent:* Friday, September 25, 2015 8:41 PM >>> *To:* [email protected] >>> *Subject:* Re: Problem running example (wrong IP address) >>> >>> I'm still stuck on this and posted it to stackoverflow: >>> >>> http://stackoverflow.com/questions/32785256/hadoop-datanode-binds-wrong-ip-address >>> >>> Thanks, >>> Daniel >>> >>> On Fri, Sep 25, 2015 at 8:28 AM, Daniel Watrous <[email protected]> >>> wrote: >>> >>> I could really use some help here. As you can see from the output below, >>> the two attached datanodes are identified with a non-existent IP address. >>> Can someone tell me how that gets selected or how to explicitly set it. >>> Also, why are both datanodes shown under the same name/IP? >>> >>> hadoop@hadoop-master:~$ hdfs dfsadmin -report >>> Configured Capacity: 84482326528 (78.68 GB) >>> Present Capacity: 75745546240 (70.54 GB) >>> DFS Remaining: 75744862208 (70.54 GB) >>> DFS Used: 684032 (668 KB) >>> DFS Used%: 0.00% >>> Under replicated blocks: 0 >>> Blocks with corrupt replicas: 0 >>> Missing blocks: 0 >>> Missing blocks (with replication factor 1): 0 >>> >>> ------------------------------------------------- >>> Live datanodes (2): >>> >>> Name: 192.168.51.1:50010 (192.168.51.1) >>> Hostname: hadoop-data1 >>> Decommission Status : Normal >>> Configured Capacity: 42241163264 (39.34 GB) >>> DFS Used: 303104 (296 KB) >>> Non DFS Used: 4302479360 (4.01 GB) >>> DFS Remaining: 37938380800 (35.33 GB) >>> DFS Used%: 0.00% >>> DFS Remaining%: 89.81% >>> Configured Cache Capacity: 0 (0 B) >>> Cache Used: 0 (0 B) >>> Cache Remaining: 0 (0 B) >>> Cache Used%: 100.00% >>> Cache Remaining%: 0.00% >>> Xceivers: 1 >>> Last contact: Fri Sep 25 13:25:37 UTC 2015 >>> >>> >>> Name: 192.168.51.4:50010 (hadoop-master) >>> Hostname: hadoop-master >>> Decommission Status : Normal >>> Configured Capacity: 42241163264 (39.34 GB) >>> DFS Used: 380928 (372 KB) >>> Non DFS Used: 4434300928 (4.13 GB) >>> DFS Remaining: 37806481408 (35.21 GB) >>> DFS Used%: 0.00% >>> DFS Remaining%: 89.50% >>> Configured Cache Capacity: 0 (0 B) >>> Cache Used: 0 (0 B) >>> Cache Remaining: 0 (0 B) >>> Cache Used%: 100.00% >>> Cache Remaining%: 0.00% >>> Xceivers: 1 >>> Last contact: Fri Sep 25 13:25:38 UTC 2015 >>> >>> >>> >>> On Thu, Sep 24, 2015 at 5:05 PM, Daniel Watrous <[email protected]> >>> wrote: >>> >>> The IP address is clearly wrong, but I'm not sure how it gets set. Can >>> someone tell me how to configure it to choose a valid IP address? >>> >>> On Thu, Sep 24, 2015 at 3:26 PM, Daniel Watrous <[email protected]> >>> wrote: >>> >>> I just noticed that both datanodes appear to have chosen that IP address >>> and bound that port for HDFS communication. >>> >>> http://screencast.com/t/OQNbrWFF >>> >>> Any idea why this would be? Is there some way to specify which >>> IP/hostname should be used for that? >>> >>> On Thu, Sep 24, 2015 at 3:11 PM, Daniel Watrous <[email protected]> >>> wrote: >>> >>> When I try to run a map reduce example, I get the following error: >>> >>> hadoop@hadoop-master:~$ hadoop jar >>> /usr/local/hadoop/share/hadoop/mapreduce/hadoop-mapreduce-examples-2.7.1.jar >>> pi 10 30 >>> Number of Maps = 10 >>> Samples per Map = 30 >>> 15/09/24 20:04:28 INFO hdfs.DFSClient: Exception in >>> createBlockOutputStream >>> java.io.IOException: Got error, status message , ack with firstBadLink >>> as 192.168.51.1:50010 >>> at >>> org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:140) >>> at >>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1334) >>> at >>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1237) >>> at >>> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:449) >>> 15/09/24 20:04:28 INFO hdfs.DFSClient: Abandoning >>> BP-852923283-127.0.1.1-1443119668806:blk_1073741825_1001 >>> 15/09/24 20:04:28 INFO hdfs.DFSClient: Excluding datanode >>> DatanodeInfoWithStorage[192.168.51.1:50010 >>> ,DS-45f6e06d-752e-41e8-ac25-ca88bce80d00,DISK] >>> 15/09/24 20:04:28 WARN hdfs.DFSClient: Slow waitForAckedSeqno took >>> 65357ms (threshold=30000ms) >>> Wrote input for Map #0 >>> >>> I'm not sure why it's trying to access 192.168.51.1:50010, which isn't >>> even a valid IP address in my setup. >>> >>> Daniel >>> >>> >>> >>> >>> >>> >>> >>> >> >
