[ 
https://issues.apache.org/jira/browse/HDFS-15112?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17015568#comment-17015568
 ] 

Íñigo Goiri commented on HDFS-15112:
------------------------------------

I found the issue.
The problem is that now when we try to create a file, we also get a 
ConnectException now.
The creation in the Router triggers getBlockLocations() to see if the file 
exists.
So this change breaks the assumption that we can create files when a subcluster 
is down.
The issue is in getCreateLocation() which now cannot handle this.
I added a new check to ignore those in  [^HDFS-15112.008.patch].

> RBF: Do not return FileNotFoundException when a subcluster is unavailable 
> --------------------------------------------------------------------------
>
>                 Key: HDFS-15112
>                 URL: https://issues.apache.org/jira/browse/HDFS-15112
>             Project: Hadoop HDFS
>          Issue Type: Improvement
>            Reporter: Íñigo Goiri
>            Assignee: Íñigo Goiri
>            Priority: Major
>         Attachments: HDFS-15112.000.patch, HDFS-15112.001.patch, 
> HDFS-15112.002.patch, HDFS-15112.004.patch, HDFS-15112.005.patch, 
> HDFS-15112.006.patch, HDFS-15112.007.patch, HDFS-15112.patch
>
>
> If we have a mount point using HASH_ALL across two subclusters and one of 
> them is down, we may return FileNotFoundException while the file is just in 
> the unavailable subcluster.
> We should not return FileNotFoundException but something that shows that the 
> subcluster is unavailable.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to