Sudhir,
Cool. Please let us know if you have any progress on this issue. Best regards, Ruilong Huo At 2019-02-11 03:55:13, "Sudhir Babu Pothineni" <[email protected]> wrote: Thanks, I will investigate. On Feb 10, 2019, at 5:01 AM, HuoRuilong <[email protected]> wrote: Hi Sudhir, Sorry for the late reply as I was on vocation and had limited access to internet. Usually the error message means that the connection between hdfs client and server is broken. 1. How did you write data to hdfs (i.e., libhdfs3, other libraries or applications) before you accessing it using libhdfs3? 2. What's the format of the data (i.e, parquet, orc)? 3. To dig deeper, you might need to go through how hawq use libhdfs3 to access data on hdfs, and then check if there is anything difference with your program. Best regards, Ruilong Huo At 2019-02-08 09:36:37, "Sudhir Babu Pothineni" <[email protected]> wrote: Hi Ruilong, It’s through my own program. Thanks Sudhir On Feb 7, 2019, at 7:11 PM, Ruilong Huo <[email protected]> wrote: Hi Sudhir, Please let us know if you are accessing HDFS through HAWQ or your program? Best regards, Ruilong Huo At 2019-02-08 07:37:47, "Yi JIN" <[email protected]> wrote: Sudhir, you are welcome ;) I guess some guys are on holiday due to lunar new year holiday, I am pinging them to get response to you asap. Yi On Fri, Feb 8, 2019 at 12:52 AM Sudhir Babu Pothineni <[email protected]> wrote: Thanks Yi! We are using hadoop-2.6.0-cdh5.14.0 On Feb 7, 2019, at 5:21 AM, Yi JIN <[email protected]> wrote: Hi Sudhir, I think the topics about libhdfs3 should be placed here. Can you provide full version number of HDFS you are using? Thanks. Best Yi (yjin) On Thu, Feb 7, 2019 at 7:56 AM Sudhir Babu Pothineni <[email protected]> wrote: Don’t no if this is right place to post this, I am using libhdfs3 from hawq bundle, I am getting following error very frequently, Any suggestions? 20190206 09:05:46:683402 DETAIL - RemoteBlockReader.cpp: 332: HdfsIOException: RemoteBlockReader: failed to read Block: [block pool ID: BP-408958698-192.168.93.200-1519838163456 block ID 1074022852_282029] from Datanode: hostXXX.XXX.com(192.168.93.204). 20190206 09:05:46:683411 DETAIL - @Hdfs::Internal::RemoteBlockReader::read(char*, int) 20190206 09:05:46:683420 DETAIL - @Hdfs::Internal::InputStreamImpl::readOneBlock(char*, int, bool) 20190206 09:05:46:683427 DETAIL - @Hdfs::Internal::InputStreamImpl::readInternal(char*, int) 20190206 09:05:46:683435 DETAIL - @Hdfs::Internal::InputStreamImpl::read(char*, int) 20190206 09:05:46:683441 DETAIL - @hdfsRead ..... 20190206 09:05:46:683878 DETAIL - Caused by 20190206 09:05:46:683886 DETAIL - TcpSocket.cpp: 69: HdfsNetworkException: Read 3040 bytes failed from "192.168.93.204:50010": (errno: 104) Connection reset by peer 20190206 09:05:46:683894 DETAIL - @Hdfs::Internal::TcpSocketImpl::read(char*, int) 20190206 09:05:46:683901 DETAIL - @Hdfs::Internal::TcpSocketImpl::readFully(char*, int, int) 20190206 09:05:46:683908 DETAIL - @Hdfs::Internal::RemoteBlockReader::readNextPacket() 20190206 09:05:46:683916 DETAIL - @Hdfs::Internal::RemoteBlockReader::read(char*, int) 20190206 09:05:46:683923 DETAIL - @Hdfs::Internal::InputStreamImpl::readOneBlock(char*, int, bool) 20190206 09:05:46:683931 DETAIL - @Hdfs::Internal::InputStreamImpl::readInternal(char*, int) 20190206 09:05:46:683938 DETAIL - @Hdfs::Internal::InputStreamImpl::read(char*, int) 20190206 09:05:46:683945 DETAIL - @hdfsRead ..... , retry read again from another Datanode.
