Yes, "hadoop fs -ls hdfs://987.65.43.21:8020/" this is not working.
I check ping 987.65.43.21 which is working fine but telnet 987.65.43.21 8020(also tried port 80, 50070) is throwing error *telnet: Unable to connect to remote host: Connection timed out* I think this is where the problem is. Also checked the conf folder. There isn't any property related to rpc-address in hdfs-site.xml . And the cluster is non-HA. On Fri, May 24, 2019 at 1:18 PM yangtao.yt <yangtao...@alibaba-inc.com> wrote: > Hi, akshay > > Can you successfully execute command such as "hadoop fs -ls > hdfs://987.65.43.21:8020/" on the machine you ran distcp? I think the > answer is no. > If both servers are reachable, I think you should check conf > dfs.namenode.rpc-address or dfs.namenode.rpc-address.<nameservice>.nn1 (if > HA enabled) from hdfs-site.xml to fetch the correct rpc port. > > Best, > Tao Yang > > 在 2019年5月24日,下午3:20,akshay naidu <akshaynaid...@gmail.com> 写道: > > Hello Joey, > Just to understand distcp I am trying to copy one file. Otherwise the data > is to be copied is > 1.5TB . > Anyways I tried running -cp but looks like the issue is in connectivity. > See logs :- > hdfs dfs -cp hdfs:// > 123.45.67.89:54310/data-analytics/spike/beNginxLogs/today/123.45.67.89_2019-05-22.access.log.gz > hdfs://987.65.43.21:50070/distCp/ > 19/05/24 07:15:22 INFO ipc.Client: Retrying connect to server: > li868-219.members.linode.com/ 987.65.43.21:50070. Already tried 0 > time(s); maxRetries=45 > 19/05/24 07:15:42 INFO ipc.Client: Retrying connect to server: > li868-219.members.linode.com/ 987.65.43.21:50070. Already tried 1 > time(s); maxRetries=45 > 19/05/24 07:16:02 INFO ipc.Client: Retrying connect to server: > li868-219.members.linode.com/ 987.65.43.21:50070. Already tried 2 > time(s); maxRetries=45 > . > . > facing same issue. > Any Idea? > Thanks. Regards > > On Fri, May 24, 2019 at 8:10 AM Joey Krabacher <jkrabac...@gmail.com> > wrote: > >> It looks like you're just trying to copy 1 file? >> Why not use 'hdfs dfs -cp ...' instead? >> >> On Thu, May 23, 2019, 21:22 yangtao.yt <yangtao...@alibaba-inc.com> >> wrote: >> >>> Hi, akshay >>> >>> Seems it’s not distcp’s business, SocketTimeout exceptions may be caused >>> by network unreachable or unavailable remote server, you can communicate >>> with the target hdfs cluster directly on the machine you executed distcp >>> command to have a test. >>> Fully causes and suggestions given by community can be fetched from >>> here: https://wiki.apache.org/hadoop/SocketTimeout >>> >>> There is a doubt in your distcp command, why using port 50070 (http >>> port) instead of 8020 (rpc port) for the target hdfs cluster? I’m confusing >>> that it still can connect with 8020 according to your logs. >>> >>> Best, >>> Tao Yang >>> >>> 在 2019年5月23日,下午8:54,akshay naidu <akshaynaid...@gmail.com> 写道: >>> >>> sun.reflect.NativeConstructorAccessorImpl.newInstance0 >>> >>> >>> >