Oh, I see. Is that you have configured a conflicted port before? 2015-04-09 18:36 GMT+08:00 sandeep vura <[email protected]>:
> Hi Yanghaogn, > > Sure, We couldn't able to load the file from local to HDFS. Its getting > exception DFSOutputStream connection refused,which means packets are not > receiving properly from namenode to datanodes .However,if we start clusters > our datanodes are not starting properly and getting connection closed > exception. > > Our Hadoop WebUI also opening very slow ,ssh connection also very > slow.Then finally we have changed our network ports and checked the > performance of the cluster it works good. > > Issue was fixed in Namenode server network port. > > Regards, > Sandeep.v > > > On Thu, Apr 9, 2015 at 12:30 PM, 杨浩 <[email protected]> wrote: > >> Root cause: Network related issue? >> can you tell us more detailedly? Thank you >> >> 2015-04-09 13:51 GMT+08:00 sandeep vura <[email protected]>: >> >>> Our issue has been resolved. >>> >>> Root cause: Network related issue. >>> >>> Thanks for each and everyone spent sometime and replied to my questions. >>> >>> Regards, >>> Sandeep.v >>> >>> On Thu, Apr 9, 2015 at 10:45 AM, sandeep vura <[email protected]> >>> wrote: >>> >>>> Can anyone give solution for my issue? >>>> >>>> On Thu, Apr 9, 2015 at 12:48 AM, sandeep vura <[email protected]> >>>> wrote: >>>> >>>>> Exactly but every time it picks randomly. Our datanodes are >>>>> 192.168.2.81,192.168.2.82,192.168.2.83,192.168.2.84,192.168.2.85 >>>>> >>>>> Namenode : 192.168.2.80 >>>>> >>>>> If i restarts the cluster next time it will show 192.168.2.81:50010 >>>>> connection closed >>>>> >>>>> On Thu, Apr 9, 2015 at 12:28 AM, Liaw, Huat (MTO) < >>>>> [email protected]> wrote: >>>>> >>>>>> You can not start 192.168.2.84:50010…. closed by ((192.168.2.x >>>>>> -datanode)) >>>>>> >>>>>> >>>>>> >>>>>> *From:* sandeep vura [mailto:[email protected]] >>>>>> *Sent:* April 8, 2015 2:39 PM >>>>>> >>>>>> *To:* [email protected] >>>>>> *Subject:* Re: Unable to load file from local to HDFS cluster >>>>>> >>>>>> >>>>>> >>>>>> We are using this setup from a very long time.We are able to run all >>>>>> the jobs successfully but suddenly went wrong with namenode. >>>>>> >>>>>> >>>>>> >>>>>> On Thu, Apr 9, 2015 at 12:06 AM, sandeep vura <[email protected]> >>>>>> wrote: >>>>>> >>>>>> I have also noticed another issue when starting hadoop cluster >>>>>> start-all.sh command >>>>>> >>>>>> >>>>>> >>>>>> namenode and datanode daemons are starting.But sometimes one of the >>>>>> datanode would drop the connection and it shows the message connection >>>>>> closed by ((192.168.2.x -datanode)) everytime when it restart the hadoop >>>>>> cluster datanode will keeps changing . >>>>>> >>>>>> >>>>>> >>>>>> for example 1st time when i starts hadoop cluster - 192.168.2.1 - >>>>>> connection closed >>>>>> >>>>>> 2nd time when i starts hadoop cluster - 192.168.2.2-connection closed >>>>>> .This point again 192.168.2.1 will starts successfuly without any errors. >>>>>> >>>>>> >>>>>> >>>>>> I couldn't able to figure out the issue exactly.Is issue relates to >>>>>> network or Hadoop configuration. >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> On Wed, Apr 8, 2015 at 11:54 PM, Liaw, Huat (MTO) < >>>>>> [email protected]> wrote: >>>>>> >>>>>> hadoop fs -put <source> <destination> Copy from remote location to >>>>>> HDFS >>>>>> >>>>>> >>>>>> >>>>>> *From:* sandeep vura [mailto:[email protected]] >>>>>> *Sent:* April 8, 2015 2:24 PM >>>>>> *To:* [email protected] >>>>>> *Subject:* Re: Unable to load file from local to HDFS cluster >>>>>> >>>>>> >>>>>> >>>>>> Sorry Liaw,I tried same command but its didn't resolve. >>>>>> >>>>>> >>>>>> >>>>>> Regards, >>>>>> >>>>>> Sandeep.V >>>>>> >>>>>> >>>>>> >>>>>> On Wed, Apr 8, 2015 at 11:37 PM, Liaw, Huat (MTO) < >>>>>> [email protected]> wrote: >>>>>> >>>>>> Should be hadoop dfs -put >>>>>> >>>>>> >>>>>> >>>>>> *From:* sandeep vura [mailto:[email protected]] >>>>>> *Sent:* April 8, 2015 1:53 PM >>>>>> *To:* [email protected] >>>>>> *Subject:* Unable to load file from local to HDFS cluster >>>>>> >>>>>> >>>>>> >>>>>> Hi, >>>>>> >>>>>> >>>>>> >>>>>> When loading a file from local to HDFS cluster using the below >>>>>> command >>>>>> >>>>>> >>>>>> >>>>>> hadoop fs -put sales.txt /sales_dept. >>>>>> >>>>>> >>>>>> >>>>>> Getting the following exception.Please let me know how to resolve >>>>>> this issue asap.Please find the attached is the logs that is displaying >>>>>> on >>>>>> namenode. >>>>>> >>>>>> >>>>>> >>>>>> Regards, >>>>>> >>>>>> Sandeep.v >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>>> >>>>> >>>>> >>>> >>> >> >
