[
https://issues.apache.org/jira/browse/HDFS-8332?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=14546107#comment-14546107
]
Sean Busbey commented on HDFS-8332:
-----------------------------------
+1 for trunk/branc-3 only. A "Known Issue" note on the next set of branch-2
release notes would be a nice-to-have as well.
{quote}
Also, I'd like to suggest that we change pre-commit to trigger
hadoop-hdfs-httpfs tests automatically for all hadoop-hdfs patches. We've seen
problems like this in the past. hadoop-hdfs-httpfs gets patched so infrequently
that it's easy to miss it when a hadoop-hdfs change introduces a test failure.
As a practical matter, we might not be able to add those tests until the
current HDFS test runs get optimized.
{quote}
Leave a note on HADOOP-11929, [~aw] is already specifying that hadoop-hdfs
needs to have hadoop-common built with native bits. Not sure if expanding to
"under tests always do this other module if this module changes" will be in
scope or a new ticket.
> DFS client API calls should check filesystem closed
> ---------------------------------------------------
>
> Key: HDFS-8332
> URL: https://issues.apache.org/jira/browse/HDFS-8332
> Project: Hadoop HDFS
> Issue Type: Bug
> Reporter: Rakesh R
> Assignee: Rakesh R
> Fix For: 2.8.0
>
> Attachments: HDFS-8332-000.patch, HDFS-8332-001.patch,
> HDFS-8332-002-Branch-2.patch, HDFS-8332-002.patch,
> HDFS-8332.001.branch-2.patch
>
>
> I could see {{listCacheDirectives()}} and {{listCachePools()}} APIs can be
> called even after the filesystem close. Instead these calls should do
> {{checkOpen}} and throws:
> {code}
> java.io.IOException: Filesystem closed
> at org.apache.hadoop.hdfs.DFSClient.checkOpen(DFSClient.java:464)
> {code}
--
This message was sent by Atlassian JIRA
(v6.3.4#6332)