Thank you for Tim Armstrong, the version of impala I am using is based on 3.2, 
it contains the fixed IMPALA-7738. When the error "hdfsOpenFile() for 
<filename> at backend <hostname:port> failed to finish
before the <hdfs_operation_timeout_sec> second timeout " appears, it may also 
appears “ Failed to submit hdfsOpenFile() for <filename> at backend 
<hostname:port> to thread pool after waiting <hdfs_operation_timeout_sec> 
seconds"

> 2020年11月3日 上午2:59,Tim Armstrong <tarmstr...@cloudera.com> 写道:
> 
> https://issues.apache.org/jira/browse/IMPALA-7738 can mitigate this to a
> large degree (it will cancel the query and log the error if the timeout is
> exceeded). It is not a 100% solution since the thread pool is fixed in size
> and can in theory get exhausted.
> 
> On Sun, Nov 1, 2020 at 9:18 PM hexianqing <hexianqing...@126.com> wrote:
> 
>> Hi all,
>> I hit the issue that queries Stuck on Failed HDFS Calls and not Timing out
>> several times when the Namenode is heavily loaded。
>> In Impala Known Issues, it is described as follows:
>> "In Impala 3.2 and higher, if the following error appears multiple times
>> in a short duration while running a query, it would mean that the
>> connection between the impalad and the HDFS NameNode is in a bad state and
>> hence the impalad would have to be restarted:
>> "hdfsOpenFile() for <filename> at backend <hostname:port> failed to finish
>> before the <hdfs_operation_timeout_sec> second timeout "
>> In Impala 3.1 and lower, the same issue would cause Impala to wait for a
>> long time or hang without showing the above error message.
>> Apache Issue: HADOOP-15720(
>> https://issues.apache.org/jira/browse/HADOOP-15720)
>> Affected Versions: All versions of Impala
>> Workaround: Restart the impalad in the bad state."
>> I wonder if there is a way to avoid this or is there a plan to fix it,
>> thank you!
>> 

Reply via email to