Can anyone please help me here.

On Tue, May 28, 2019 at 2:45 PM Behroz Sikander <behro...@gmail.com> wrote:

> Hello,
> I have a client application which uses HDFS API to push data. My HDFS is
> setup in HA mode.
> My client application has certain timeout requirements e.g. an upload
> shouldn't take longer than 10 seconds otherwise the operation is considered
> a failure.
>
> Sometimes, I notice the following exceptions in my code
>
> 2019-05-23 14:13:31,356 INFO Thread-11 org.apache.hadoop.hdfs.DFSClient []: 
> Exception in createBlockOutputStream
> java.io.IOException: Got error, status message , ack with firstBadLink as 
> xx.xx.xx.xx:50010
>         at 
> org.apache.hadoop.hdfs.protocol.datatransfer.DataTransferProtoUtil.checkBlockOpStatus(DataTransferProtoUtil.java:142)
>         at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.createBlockOutputStream(DFSOutputStream.java:1359)
>         at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.nextBlockOutputStream(DFSOutputStream.java:1262)
>         at 
> org.apache.hadoop.hdfs.DFSOutputStream$DataStreamer.run(DFSOutputStream.java:448)2019-05-23
>  14:13:31,357 INFO Thread-11 org.apache.hadoop.hdfs.DFSClient []: Abandoning 
> BP-1672040070-127.0.0.1-1527078068582:blk_1073742599_17752019-05-23 
> 14:13:31,384 INFO Thread-11 org.apache.hadoop.hdfs.DFSClient []: Excluding 
> datanode 
> DatanodeInfoWithStorage[xx.xx.xx.xx:50010,DS-4be6740a-bad8-438b-99d6-fbc50d7760dd,DISK]
>
>
> Even though the upload succeeds but takes around *20 seconds*.
>
> I am interested in the configuration property which controls this timeout, so 
> I can adjust it
>
> according to my client's requirements. Or is there any configuration which 
> says
>
> if the operation is not completed within X seconds, then HDFS API throws an 
> exception?
>
> Can it be ipc.client.connect.timeout?
>
>
> Thanks.
>
> Behroz
>
>

Reply via email to