Thanks, I will investigate.

> On Feb 10, 2019, at 5:01 AM, HuoRuilong <[email protected]> wrote:
> 
> Hi Sudhir,
> 
> Sorry for the late reply as I was on vocation and had limited access to 
> internet.
> 
> Usually the error message means that the connection between hdfs client and 
> server is broken.
> 1. How did you write data to hdfs (i.e., libhdfs3, other libraries or 
> applications) before you accessing it using libhdfs3?
> 2. What's the format of the data (i.e, parquet, orc)?
> 3. To dig deeper, you might need to go through how hawq use libhdfs3 to 
> access data on hdfs, and then check if there is anything difference with your 
> program.
> 
> Best regards,
> Ruilong Huo
> 
> At 2019-02-08 09:36:37, "Sudhir Babu Pothineni" <[email protected]> wrote:
> Hi Ruilong,
> 
> It’s through my own program.
> 
> Thanks
> Sudhir
> 
>> On Feb 7, 2019, at 7:11 PM, Ruilong Huo <[email protected]> wrote:
>> 
>> Hi Sudhir,
>> 
>> Please let us know if you are accessing HDFS through HAWQ or your program? 
>> 
>> Best regards,
>> Ruilong Huo
>> 
>> At 2019-02-08 07:37:47, "Yi JIN" <[email protected]> wrote:
>> Sudhir, you are welcome ;) I guess some guys are on holiday due to lunar new 
>> year holiday, I am pinging them to get response to you asap. 
>> 
>> Yi
>> 
>>> On Fri, Feb 8, 2019 at 12:52 AM Sudhir Babu Pothineni 
>>> <[email protected]> wrote:
>>> Thanks Yi!
>>> 
>>> We are using hadoop-2.6.0-cdh5.14.0
>>> 
>>> 
>>>> On Feb 7, 2019, at 5:21 AM, Yi JIN <[email protected]> wrote:
>>>> 
>>>> Hi Sudhir,
>>>> 
>>>> I think the topics about libhdfs3 should be placed here. Can you provide 
>>>> full version number of HDFS you are using? Thanks.
>>>> 
>>>> Best
>>>> Yi (yjin)
>>>> 
>>>>> On Thu, Feb 7, 2019 at 7:56 AM Sudhir Babu Pothineni 
>>>>> <[email protected]> wrote:
>>>>> Don’t no if this is right place to post this, I am using libhdfs3 from 
>>>>> hawq bundle, I am getting following error very frequently,
>>>>> 
>>>>> Any suggestions?
>>>>> 
>>>>> 20190206 09:05:46:683402 DETAIL - RemoteBlockReader.cpp: 332: 
>>>>> HdfsIOException: RemoteBlockReader: failed to read Block: [block pool ID: 
>>>>> BP-408958698-192.168.93.200-1519838163456 block ID 1074022852_282029] 
>>>>> from Datanode: hostXXX.XXX.com(192.168.93.204).
>>>>> 20190206 09:05:46:683411 DETAIL -         @       
>>>>> Hdfs::Internal::RemoteBlockReader::read(char*, int)
>>>>> 20190206 09:05:46:683420 DETAIL -         @       
>>>>> Hdfs::Internal::InputStreamImpl::readOneBlock(char*, int, bool)
>>>>> 20190206 09:05:46:683427 DETAIL -         @       
>>>>> Hdfs::Internal::InputStreamImpl::readInternal(char*, int)
>>>>> 20190206 09:05:46:683435 DETAIL -         @       
>>>>> Hdfs::Internal::InputStreamImpl::read(char*, int)
>>>>> 20190206 09:05:46:683441 DETAIL -         @       hdfsRead
>>>>> 
>>>>> .....
>>>>> 20190206 09:05:46:683878 DETAIL - Caused by
>>>>> 20190206 09:05:46:683886 DETAIL - TcpSocket.cpp: 69: 
>>>>> HdfsNetworkException: Read 3040 bytes failed from "192.168.93.204:50010": 
>>>>> (errno: 104) Connection reset by peer
>>>>> 20190206 09:05:46:683894 DETAIL -         @       
>>>>> Hdfs::Internal::TcpSocketImpl::read(char*, int)
>>>>> 20190206 09:05:46:683901 DETAIL -         @       
>>>>> Hdfs::Internal::TcpSocketImpl::readFully(char*, int, int)
>>>>> 20190206 09:05:46:683908 DETAIL -         @       
>>>>> Hdfs::Internal::RemoteBlockReader::readNextPacket()
>>>>> 20190206 09:05:46:683916 DETAIL -         @       
>>>>> Hdfs::Internal::RemoteBlockReader::read(char*, int)
>>>>> 20190206 09:05:46:683923 DETAIL -         @       
>>>>> Hdfs::Internal::InputStreamImpl::readOneBlock(char*, int, bool)
>>>>> 20190206 09:05:46:683931 DETAIL -         @       
>>>>> Hdfs::Internal::InputStreamImpl::readInternal(char*, int)
>>>>> 20190206 09:05:46:683938 DETAIL -         @       
>>>>> Hdfs::Internal::InputStreamImpl::read(char*, int)
>>>>> 20190206 09:05:46:683945 DETAIL -         @       hdfsRead
>>>>> .....
>>>>> , retry read again from another Datanode.

Reply via email to