[
https://issues.apache.org/jira/browse/HADOOP-19541?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=17946943#comment-17946943
]
ASF GitHub Bot commented on HADOOP-19541:
-----------------------------------------
fuchaohong commented on code in PR #7611:
URL: https://github.com/apache/hadoop/pull/7611#discussion_r2057602437
##########
hadoop-tools/hadoop-archives/src/test/java/org/apache/hadoop/tools/TestHadoopArchives.java:
##########
@@ -804,5 +804,31 @@ public void testCopyToLocal() throws Exception {
localFs.delete(tmpPath, true);
}
}
-
+
+ @Test
+ public void testBlockSize() throws Exception {
+ conf.set(HadoopArchives.HAR_BLOCKSIZE_LABEL, "1m");
+
+ final String inputPathStr = inputPath.toUri().getPath();
+ System.out.println("inputPathStr = " + inputPathStr);
Review Comment:
@pan3793 I have already removed it, please help review it again.
> Make HadoopArchives support human-friendly units about blocksize and partsize.
> ------------------------------------------------------------------------------
>
> Key: HADOOP-19541
> URL: https://issues.apache.org/jira/browse/HADOOP-19541
> Project: Hadoop Common
> Issue Type: Improvement
> Reporter: fuchaohong
> Priority: Major
> Labels: pull-request-available
>
> You can use the following suffix (case insensitive): k(kilo), m(mega),
> g(giga), t(tera), p(peta), e(exa) to specify the size (such as 128k, 512m,
> 1g, etc.), Or provide complete size in bytes (such as 134217728 for 128 MB).
--
This message was sent by Atlassian Jira
(v8.20.10#820010)
---------------------------------------------------------------------
To unsubscribe, e-mail: common-issues-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-issues-h...@hadoop.apache.org