[jira] [Created] (HDFS-15121) Update overview of libhdfs
YiSheng Lien created HDFS-15121: --- Summary: Update overview of libhdfs Key: HDFS-15121 URL: https://issues.apache.org/jira/browse/HDFS-15121 Project: Hadoop HDFS Issue Type: Improvement Reporter: YiSheng Lien In [doc of libhdfs|https://hadoop.apache.org/docs/current/hadoop-project-dist/hadoop-hdfs/LibHdfs.html], the description *It provides C APIs to a subset of the HDFS APIs to manipulate HDFS files and the filesystem.* could be more friendly when we update it to *It provides C APIs to a subset of the HDFS APIs to manipulate HDFS compatible files and the compatible filesystem. (Such as o3fs in Ozone)* **Surely, we should include the classpath to use it. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-15120) Refresh BlockPlacementPolicy at runtime.
Jinglun created HDFS-15120: -- Summary: Refresh BlockPlacementPolicy at runtime. Key: HDFS-15120 URL: https://issues.apache.org/jira/browse/HDFS-15120 Project: Hadoop HDFS Issue Type: Improvement Reporter: Jinglun Assignee: Jinglun Now if we want to switch BlockPlacementPolicies we need to restart the NameNode. It would be convenient if we can switch it at runtime. For example we can switch between AvailableSpaceBlockPlacementPolicy and BlockPlacementPolicyDefault as needed. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-15119) Allow expiration of cached locations in DFSInputStream
Ahmed Hussein created HDFS-15119: Summary: Allow expiration of cached locations in DFSInputStream Key: HDFS-15119 URL: https://issues.apache.org/jira/browse/HDFS-15119 Project: Hadoop HDFS Issue Type: Improvement Components: dfsclient Reporter: Ahmed Hussein Assignee: Ahmed Hussein Staleness and other transient conditions can affect reads for a long time since the block locations may not be re-fetched. It makes sense to make cached locations to expire. For example, we may not take advantage of local-reads since the nodes are blacklisted and have not been updated. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-15118) [SBN Read] Slow clients when Observer reads are enabled but there are no Observers on the cluster.
Konstantin Shvachko created HDFS-15118: -- Summary: [SBN Read] Slow clients when Observer reads are enabled but there are no Observers on the cluster. Key: HDFS-15118 URL: https://issues.apache.org/jira/browse/HDFS-15118 Project: Hadoop HDFS Issue Type: Bug Components: hdfs-client Affects Versions: 2.10.0 Reporter: Konstantin Shvachko We see substantial degradation in performance of HDFS clients, when Observer reads are enabled via {{ObserverReadProxyProvider}}, but there are no ObserverNodes on the cluster. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
Re: [DISCUSS] Guidelines for Code cleanup JIRAs
+1 Can we also make sure to add a label for the code cleanup Jiras? At least, this will make it easy to search and filter jiras. On Mon, Jan 13, 2020 at 7:24 AM Wei-Chiu Chuang wrote: > +1 > > On Thu, Jan 9, 2020 at 9:33 AM epa...@apache.org > wrote: > > > There was some discussion on > > https://issues.apache.org/jira/browse/YARN-9052 > > about concerns surrounding the costs/benefits of code cleanup JIRAs. This > > email > > is to get the discussion going within a wider audience. > > > > The positive points for code cleanup JIRAs: > > - Clean up tech debt > > - Make code more readable > > - Make code more maintainable > > - Make code more performant > > > > The concerns regarding code cleanup JIRAs are as follows: > > - If the changes only go into trunk, then contributors and committers > > trying to > > backport to prior releases will have to create and test multiple patch > > versions. > > - Some have voiced concerns that code cleanup JIRAs may not be tested as > > thoroughly as features and bug fixes because functionality is not > > supposed to > > change. > > - Any patches awaiting review that are touching the same code will have > to > > be > > redone, re-tested, and re-reviewed. > > - JIRAs that are opened for code cleanup and not worked on right away > tend > > to > > clutter up the JIRA space. > > > > Here are my opinions: > > - Code changes of any kind force a non-trivial amount of overhead for > other > > developers. For code cleanup JIRAs, sometimes the usability, > > maintainability, > > and performance is worth the overhead (as in the case of YARN-9052). > > - Before opening any JIRA, please always consider whether or not the > added > > usability will outweigh the added pain you are causing other > developers. > > - If you believe the benefits outweigh the costs, please backport the > > changes > > yourself to all active lines. My preference is to port all the way back > > to 2.10. > > - Please don't run code analysis tools and then open many JIRAs that > > document > > those findings. That activity does not put any thought into this > > cost-benefit > > analysis. > > > > Thanks everyone. I'm looking forward to your thoughts. I appreciate all > > you do > > for the open source community and it is always a pleasure to work with > you. > > -Eric Payne > > > > - > > To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org > > For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org > > > > >
Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86
For more details, see https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1380/ [Jan 12, 2020 12:48:39 PM] (snemeth) YARN-10067. Add dry-run feature to FS-CS converter tool. Contributed by [Jan 12, 2020 1:04:15 PM] (snemeth) YARN-9866. u:user2:%primary_group is not working as expected. -1 overall The following subsystems voted -1: asflicense findbugs pathlen unit xml The following subsystems voted -1 but were configured to be filtered/ignored: cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: XML : Parsing Error(s): hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-excerpt.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-output-missing-tags2.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-nodemanager/src/test/resources/nvidia-smi-sample-output.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/fair-scheduler-invalid.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-resourcemanager/src/test/resources/yarn-site-with-invalid-allocation-file-ref.xml FindBugs : module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-applications-mawo/hadoop-yarn-applications-mawo-core Class org.apache.hadoop.applications.mawo.server.common.TaskStatus implements Cloneable but does not define or use clone method At TaskStatus.java:does not define or use clone method At TaskStatus.java:[lines 39-346] Equals method for org.apache.hadoop.applications.mawo.server.worker.WorkerId assumes the argument is of type WorkerId At WorkerId.java:the argument is of type WorkerId At WorkerId.java:[line 114] org.apache.hadoop.applications.mawo.server.worker.WorkerId.equals(Object) does not check for null argument At WorkerId.java:null argument At WorkerId.java:[lines 114-115] FindBugs : module:hadoop-cloud-storage-project/hadoop-cos Redundant nullcheck of dir, which is known to be non-null in org.apache.hadoop.fs.cosn.BufferPool.createDir(String) Redundant null check at BufferPool.java:is known to be non-null in org.apache.hadoop.fs.cosn.BufferPool.createDir(String) Redundant null check at BufferPool.java:[line 66] org.apache.hadoop.fs.cosn.CosNInputStream$ReadBuffer.getBuffer() may expose internal representation by returning CosNInputStream$ReadBuffer.buffer At CosNInputStream.java:by returning CosNInputStream$ReadBuffer.buffer At CosNInputStream.java:[line 87] Found reliance on default encoding in org.apache.hadoop.fs.cosn.CosNativeFileSystemStore.storeFile(String, File, byte[]):in org.apache.hadoop.fs.cosn.CosNativeFileSystemStore.storeFile(String, File, byte[]): new String(byte[]) At CosNativeFileSystemStore.java:[line 199] Found reliance on default encoding in org.apache.hadoop.fs.cosn.CosNativeFileSystemStore.storeFileWithRetry(String, InputStream, byte[], long):in org.apache.hadoop.fs.cosn.CosNativeFileSystemStore.storeFileWithRetry(String, InputStream, byte[], long): new String(byte[]) At CosNativeFileSystemStore.java:[line 178] org.apache.hadoop.fs.cosn.CosNativeFileSystemStore.uploadPart(File, String, String, int) may fail to clean up java.io.InputStream Obligation to clean up resource created at CosNativeFileSystemStore.java:fail to clean up java.io.InputStream Obligation to clean up resource created at CosNativeFileSystemStore.java:[line 252] is not discharged Failed junit tests : hadoop.fs.viewfs.TestViewFileSystemWithAuthorityLocalFileSystem hadoop.fs.viewfs.TestViewFsTrash hadoop.hdfs.server.balancer.TestBalancer hadoop.hdfs.server.namenode.TestRedudantBlocks hadoop.hdfs.server.blockmanagement.TestUnderReplicatedBlocks hadoop.hdfs.TestDeadNodeDetection hadoop.fs.http.client.TestHttpFSFWithSWebhdfsFileSystem hadoop.yarn.server.resourcemanager.scheduler.fair.TestFairSchedulerPreemption hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageDomain hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowRun hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowActivity hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageApps hadoop.yarn.server.timelineservice.storage.TestTimelineWriterHBaseDown hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowRunCompaction
Apache Hadoop qbt Report: branch2.10+JDK7 on Linux/x86
For more details, see https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/ No changes -1 overall The following subsystems voted -1: asflicense findbugs hadolint pathlen unit xml The following subsystems voted -1 but were configured to be filtered/ignored: cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace The following subsystems are considered long running: (runtime bigger than 1h 0m 0s) unit Specific tests: XML : Parsing Error(s): hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/empty-configuration.xml hadoop-tools/hadoop-azure/src/config/checkstyle-suppressions.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/public/crossdomain.xml hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/public/crossdomain.xml FindBugs : module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client Boxed value is unboxed and then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:then immediately reboxed in org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result, byte[], byte[], KeyConverter, ValueConverter, boolean) At ColumnRWHelper.java:[line 335] Failed junit tests : hadoop.ipc.TestProtoBufRpc hadoop.contrib.bkjournal.TestBookKeeperHACheckpoints hadoop.hdfs.server.balancer.TestBalancerRPCDelay hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys hadoop.contrib.bkjournal.TestBookKeeperHACheckpoints hadoop.registry.secure.TestSecureLogins hadoop.yarn.server.timelineservice.security.TestTimelineAuthFilterForV2 hadoop.yarn.client.api.impl.TestAMRMClient hadoop.mapreduce.v2.TestSpeculativeExecutionWithMRApp cc: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-compile-cc-root-jdk1.7.0_95.txt [4.0K] javac: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-compile-javac-root-jdk1.7.0_95.txt [328K] cc: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-compile-cc-root-jdk1.8.0_232.txt [4.0K] javac: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-compile-javac-root-jdk1.8.0_232.txt [308K] checkstyle: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-checkstyle-root.txt [16M] hadolint: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-patch-hadolint.txt [4.0K] pathlen: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/pathlen.txt [12K] pylint: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-patch-pylint.txt [24K] shellcheck: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-patch-shellcheck.txt [56K] shelldocs: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-patch-shelldocs.txt [8.0K] whitespace: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/whitespace-eol.txt [12M] https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/whitespace-tabs.txt [1.3M] xml: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/xml.txt [12K] findbugs: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timelineservice-hbase_hadoop-yarn-server-timelineservice-hbase-client-warnings.html [8.0K] javadoc: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-javadoc-javadoc-root-jdk1.7.0_95.txt [16K] https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/diff-javadoc-javadoc-root-jdk1.8.0_232.txt [1.1M] unit: https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt [160K] https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt [324K] https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/566/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs_src_contrib_bkjournal.txt [12K]
Re: [DISCUSS] Guidelines for Code cleanup JIRAs
+1 On Thu, Jan 9, 2020 at 9:33 AM epa...@apache.org wrote: > There was some discussion on > https://issues.apache.org/jira/browse/YARN-9052 > about concerns surrounding the costs/benefits of code cleanup JIRAs. This > email > is to get the discussion going within a wider audience. > > The positive points for code cleanup JIRAs: > - Clean up tech debt > - Make code more readable > - Make code more maintainable > - Make code more performant > > The concerns regarding code cleanup JIRAs are as follows: > - If the changes only go into trunk, then contributors and committers > trying to > backport to prior releases will have to create and test multiple patch > versions. > - Some have voiced concerns that code cleanup JIRAs may not be tested as > thoroughly as features and bug fixes because functionality is not > supposed to > change. > - Any patches awaiting review that are touching the same code will have to > be > redone, re-tested, and re-reviewed. > - JIRAs that are opened for code cleanup and not worked on right away tend > to > clutter up the JIRA space. > > Here are my opinions: > - Code changes of any kind force a non-trivial amount of overhead for other > developers. For code cleanup JIRAs, sometimes the usability, > maintainability, > and performance is worth the overhead (as in the case of YARN-9052). > - Before opening any JIRA, please always consider whether or not the added > usability will outweigh the added pain you are causing other developers. > - If you believe the benefits outweigh the costs, please backport the > changes > yourself to all active lines. My preference is to port all the way back > to 2.10. > - Please don't run code analysis tools and then open many JIRAs that > document > those findings. That activity does not put any thought into this > cost-benefit > analysis. > > Thanks everyone. I'm looking forward to your thoughts. I appreciate all > you do > for the open source community and it is always a pleasure to work with you. > -Eric Payne > > - > To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org > For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org > >
[jira] [Created] (HDFS-15117) EC: Add getECTopologyResultForPolicies to DistributedFileSystem
Ayush Saxena created HDFS-15117: --- Summary: EC: Add getECTopologyResultForPolicies to DistributedFileSystem Key: HDFS-15117 URL: https://issues.apache.org/jira/browse/HDFS-15117 Project: Hadoop HDFS Issue Type: Improvement Reporter: Ayush Saxena Assignee: Ayush Saxena Add getECTopologyResultForPolicies API to distributed filesystem. It is as of now only present as part of ECAdmin. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-15116) Correct spelling of comments for NNStorage.setRestoreFailedStorage
Xudong Cao created HDFS-15116: - Summary: Correct spelling of comments for NNStorage.setRestoreFailedStorage Key: HDFS-15116 URL: https://issues.apache.org/jira/browse/HDFS-15116 Project: Hadoop HDFS Issue Type: Bug Affects Versions: 3.2.1 Reporter: Xudong Cao Assignee: Xudong Cao Correct spelling of comments for NNStorage.setRestoreFailedStorage. -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
[jira] [Created] (HDFS-15115) Namenode crash caused by NPE in BlockPlacementPolicyDefault when dynamically change logger to debug
wangzhixiang created HDFS-15115: --- Summary: Namenode crash caused by NPE in BlockPlacementPolicyDefault when dynamically change logger to debug Key: HDFS-15115 URL: https://issues.apache.org/jira/browse/HDFS-15115 Project: Hadoop HDFS Issue Type: Bug Reporter: wangzhixiang Attachments: 0001-fix-up-the-Namenode-crash-bug-caused-by-dynamically-.patch To get debug info, we dynamically change the logger of BlockPlacementPolicyDefault to debug when namenode is running. However, the Namenode crashs. From the log, we find some NPE in BlockPlacementPolicyDefault.chooseRandom. Because *StringBuilder builder* will be used 4 times in BlockPlacementPolicyDefault.chooseRandom method. While the *builder* only initializes in the first time of this method. If we change the logger of BlockPlacementPolicyDefault to debug after the part, the *builder* in remaining part is *NULL* and cause *NPE* -- This message was sent by Atlassian Jira (v8.3.4#803005) - To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org