Root cause: Network related issue? can you tell us more detailedly? Thank you
2015-04-09 13:51 GMT+08:00 sandeep vura <[email protected]>: > Our issue has been resolved. > > Root cause: Network related issue. > > Thanks for each and everyone spent sometime and replied to my questions. > > Regards, > Sandeep.v > > On Thu, Apr 9, 2015 at 10:45 AM, sandeep vura <[email protected]> > wrote: > >> Can anyone give solution for my issue? >> >> On Thu, Apr 9, 2015 at 12:48 AM, sandeep vura <[email protected]> >> wrote: >> >>> Exactly but every time it picks randomly. Our datanodes are >>> 192.168.2.81,192.168.2.82,192.168.2.83,192.168.2.84,192.168.2.85 >>> >>> Namenode : 192.168.2.80 >>> >>> If i restarts the cluster next time it will show 192.168.2.81:50010 >>> connection closed >>> >>> On Thu, Apr 9, 2015 at 12:28 AM, Liaw, Huat (MTO) <[email protected]> >>> wrote: >>> >>>> You can not start 192.168.2.84:50010…. closed by ((192.168.2.x >>>> -datanode)) >>>> >>>> >>>> >>>> *From:* sandeep vura [mailto:[email protected]] >>>> *Sent:* April 8, 2015 2:39 PM >>>> >>>> *To:* [email protected] >>>> *Subject:* Re: Unable to load file from local to HDFS cluster >>>> >>>> >>>> >>>> We are using this setup from a very long time.We are able to run all >>>> the jobs successfully but suddenly went wrong with namenode. >>>> >>>> >>>> >>>> On Thu, Apr 9, 2015 at 12:06 AM, sandeep vura <[email protected]> >>>> wrote: >>>> >>>> I have also noticed another issue when starting hadoop cluster >>>> start-all.sh command >>>> >>>> >>>> >>>> namenode and datanode daemons are starting.But sometimes one of the >>>> datanode would drop the connection and it shows the message connection >>>> closed by ((192.168.2.x -datanode)) everytime when it restart the hadoop >>>> cluster datanode will keeps changing . >>>> >>>> >>>> >>>> for example 1st time when i starts hadoop cluster - 192.168.2.1 - >>>> connection closed >>>> >>>> 2nd time when i starts hadoop cluster - 192.168.2.2-connection closed >>>> .This point again 192.168.2.1 will starts successfuly without any errors. >>>> >>>> >>>> >>>> I couldn't able to figure out the issue exactly.Is issue relates to >>>> network or Hadoop configuration. >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>>> On Wed, Apr 8, 2015 at 11:54 PM, Liaw, Huat (MTO) <[email protected]> >>>> wrote: >>>> >>>> hadoop fs -put <source> <destination> Copy from remote location to HDFS >>>> >>>> >>>> >>>> *From:* sandeep vura [mailto:[email protected]] >>>> *Sent:* April 8, 2015 2:24 PM >>>> *To:* [email protected] >>>> *Subject:* Re: Unable to load file from local to HDFS cluster >>>> >>>> >>>> >>>> Sorry Liaw,I tried same command but its didn't resolve. >>>> >>>> >>>> >>>> Regards, >>>> >>>> Sandeep.V >>>> >>>> >>>> >>>> On Wed, Apr 8, 2015 at 11:37 PM, Liaw, Huat (MTO) <[email protected]> >>>> wrote: >>>> >>>> Should be hadoop dfs -put >>>> >>>> >>>> >>>> *From:* sandeep vura [mailto:[email protected]] >>>> *Sent:* April 8, 2015 1:53 PM >>>> *To:* [email protected] >>>> *Subject:* Unable to load file from local to HDFS cluster >>>> >>>> >>>> >>>> Hi, >>>> >>>> >>>> >>>> When loading a file from local to HDFS cluster using the below command >>>> >>>> >>>> >>>> hadoop fs -put sales.txt /sales_dept. >>>> >>>> >>>> >>>> Getting the following exception.Please let me know how to resolve this >>>> issue asap.Please find the attached is the logs that is displaying on >>>> namenode. >>>> >>>> >>>> >>>> Regards, >>>> >>>> Sandeep.v >>>> >>>> >>>> >>>> >>>> >>>> >>>> >>> >>> >> >
