[ 
https://issues.apache.org/jira/browse/HADOOP-13028?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=15274562#comment-15274562
 ] 

Chris Nauroth commented on HADOOP-13028:
----------------------------------------

Patch v010 addresses all of my prior feedback.

It's fine to keep {{S3AInstrumentation#gauge}}.

I agree that "must be private and have accessor methods" is pedantic.  I think 
we can control this with the Checkstyle 
[HiddenField|http://checkstyle.sourceforge.net/config_coding.html#HiddenField] 
rule settings.  Let's not bother for this patch though.  You've already done a 
lot to clean up Checkstyle warnings.

I agree with updating the JIRA title and closing down the other ones that were 
folded into this patch.

I'm now seeing a failure in {{TestS3ABlocksize#testBlockSize}}:
{code}
testBlockSize(org.apache.hadoop.fs.s3a.TestS3ABlocksize)  Time elapsed: 3.17 
sec  <<< FAILURE!
java.lang.AssertionError: Double default block size in stat(): 
S3AFileStatus{path=s3a://cnauroth-test-aws-s3a/test/testBlockSize/file; 
isDirectory=false; length=1024; replication=1; blocksize=33554432; 
modification_time=1462559933000; access_time=0; owner=; group=; 
permission=rw-rw-rw-; isSymlink=false} expected:<67108864> but was:<33554432>
        at org.junit.Assert.fail(Assert.java:88)
        at org.junit.Assert.failNotEquals(Assert.java:743)
        at org.junit.Assert.assertEquals(Assert.java:118)
        at org.junit.Assert.assertEquals(Assert.java:555)
        at 
org.apache.hadoop.fs.s3a.TestS3ABlocksize.testBlockSize(TestS3ABlocksize.java:65)
{code}

It looks like the test has an expectation that you can update block size in the 
{{Configuration}} of an already-initialized {{S3AFileSystem}}, and it will 
start using the new value.  After this patch, the block size is read from 
{{Configuration}} once and cached.  I think the expectation of the test is a 
little dubious, but that's the behavior that shipped in 2.7.0.  It's probably 
safest to remove this part of the patch for now.

I just have one more request.  Please remove the unused import of 
{{Configuration}} in {{S3AInstrumentation}}.

I'll be +1 after that.  We'll also need a separate trunk patch.

> add low level counter metrics for S3A; use in read performance tests
> --------------------------------------------------------------------
>
>                 Key: HADOOP-13028
>                 URL: https://issues.apache.org/jira/browse/HADOOP-13028
>             Project: Hadoop Common
>          Issue Type: Sub-task
>          Components: fs/s3, metrics
>    Affects Versions: 2.8.0
>            Reporter: Steve Loughran
>            Assignee: Steve Loughran
>         Attachments: HADOOP-13028-001.patch, HADOOP-13028-002.patch, 
> HADOOP-13028-004.patch, HADOOP-13028-005.patch, HADOOP-13028-006.patch, 
> HADOOP-13028-007.patch, HADOOP-13028-008.patch, HADOOP-13028-009.patch, 
> HADOOP-13028-branch-2-008.patch, HADOOP-13028-branch-2-009.patch, 
> HADOOP-13028-branch-2-010.patch, 
> org.apache.hadoop.fs.s3a.scale.TestS3AInputStreamPerformance-output.txt, 
> org.apache.hadoop.fs.s3a.scale.TestS3AInputStreamPerformance-output.txt
>
>
> against S3 (and other object stores), opening connections can be expensive, 
> closing connections may be expensive (a sign of a regression). 
> S3A FS and individual input streams should have counters of the # of 
> open/close/failure+reconnect operations, timers of how long things take. This 
> can be used downstream to measure efficiency of the code (how often 
> connections are being made), connection reliability, etc.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to