Hello Joey,
Just to understand distcp I am trying to copy one file. Otherwise the data
is to be copied is > 1.5TB .
Anyways I tried running -cp but looks like the issue is in connectivity.
See logs :-
hdfs dfs -cp hdfs://
123.45.67.89:54310/data-analytics/spike/beNginxLogs/today/123.45.67.89_2019-05-22.access.log.gz
hdfs://987.65.43.21:50070/distCp/
19/05/24 07:15:22 INFO ipc.Client: Retrying connect to server:
li868-219.members.linode.com/ 987.65.43.21:50070. Already tried 0 time(s);
maxRetries=45
19/05/24 07:15:42 INFO ipc.Client: Retrying connect to server:
li868-219.members.linode.com/ 987.65.43.21:50070. Already tried 1 time(s);
maxRetries=45
19/05/24 07:16:02 INFO ipc.Client: Retrying connect to server:
li868-219.members.linode.com/ 987.65.43.21:50070. Already tried 2 time(s);
maxRetries=45
.
.
facing same issue.
Any Idea?
Thanks. Regards

On Fri, May 24, 2019 at 8:10 AM Joey Krabacher <jkrabac...@gmail.com> wrote:

> It looks like you're just trying to copy 1 file?
> Why not use 'hdfs dfs -cp ...' instead?
>
> On Thu, May 23, 2019, 21:22 yangtao.yt <yangtao...@alibaba-inc.com> wrote:
>
>> Hi, akshay
>>
>> Seems it’s not distcp’s business, SocketTimeout exceptions may be caused
>> by network unreachable or unavailable remote server, you can communicate
>> with the target hdfs cluster directly on the machine you executed distcp
>> command to have a test.
>> Fully causes and suggestions given by community can be fetched from here:
>> https://wiki.apache.org/hadoop/SocketTimeout
>>
>> There is a doubt in your distcp command, why using port 50070 (http port)
>> instead of 8020 (rpc port) for the target hdfs cluster? I’m confusing that
>> it still can connect with 8020 according to your logs.
>>
>> Best,
>> Tao Yang
>>
>> 在 2019年5月23日,下午8:54,akshay naidu <akshaynaid...@gmail.com> 写道:
>>
>> sun.reflect.NativeConstructorAccessorImpl.newInstance0
>>
>>
>>

Reply via email to