hadoop-yetus removed a comment on issue #1820: HADOOP-16830. Add public 
IOStatistics API + S3A implementation
URL: https://github.com/apache/hadoop/pull/1820#issuecomment-589293736
 
 
   :broken_heart: **-1 overall**
   
   
   
   
   
   
   | Vote | Subsystem | Runtime | Comment |
   |:----:|----------:|--------:|:--------|
   | +0 :ok: |  reexec  |   1m 58s |  Docker mode activated.  |
   ||| _ Prechecks _ |
   | +1 :green_heart: |  dupname  |   0m  2s |  No case conflicting files 
found.  |
   | +1 :green_heart: |  @author  |   0m  0s |  The patch does not contain any 
@author tags.  |
   | +1 :green_heart: |  test4tests  |   0m  0s |  The patch appears to include 
13 new or modified test files.  |
   ||| _ trunk Compile Tests _ |
   | +0 :ok: |  mvndep  |   1m 30s |  Maven dependency ordering for branch  |
   | +1 :green_heart: |  mvninstall  |  27m 36s |  trunk passed  |
   | +1 :green_heart: |  compile  |  24m 16s |  trunk passed  |
   | +1 :green_heart: |  checkstyle  |   3m 43s |  trunk passed  |
   | +1 :green_heart: |  mvnsite  |   2m 29s |  trunk passed  |
   | +1 :green_heart: |  shadedclient  |  23m  0s |  branch has no errors when 
building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   1m 27s |  trunk passed  |
   | +0 :ok: |  spotbugs  |   1m  8s |  Used deprecated FindBugs config; 
considering switching to SpotBugs.  |
   | +1 :green_heart: |  findbugs  |   3m 13s |  trunk passed  |
   | -0 :warning: |  patch  |   1m 30s |  Used diff version of patch file. 
Binary files and potentially other changes not applied. Please rebase and 
squash commits if necessary.  |
   ||| _ Patch Compile Tests _ |
   | +0 :ok: |  mvndep  |   0m 20s |  Maven dependency ordering for patch  |
   | +1 :green_heart: |  mvninstall  |   1m 24s |  the patch passed  |
   | +1 :green_heart: |  compile  |  17m  9s |  the patch passed  |
   | +1 :green_heart: |  javac  |  17m  9s |  the patch passed  |
   | -0 :warning: |  checkstyle  |   2m 59s |  root: The patch generated 47 new 
+ 309 unchanged - 19 fixed = 356 total (was 328)  |
   | +1 :green_heart: |  mvnsite  |   2m 11s |  the patch passed  |
   | +1 :green_heart: |  whitespace  |   0m  0s |  The patch has no whitespace 
issues.  |
   | +1 :green_heart: |  shadedclient  |  15m 30s |  patch has no errors when 
building and testing our client artifacts.  |
   | +1 :green_heart: |  javadoc  |   1m 27s |  the patch passed  |
   | -1 :x: |  findbugs  |   1m 17s |  hadoop-tools/hadoop-aws generated 13 new 
+ 0 unchanged - 0 fixed = 13 total (was 0)  |
   ||| _ Other Tests _ |
   | -1 :x: |  unit  |   9m 18s |  hadoop-common in the patch passed.  |
   | +1 :green_heart: |  unit  |   1m 33s |  hadoop-aws in the patch passed.  |
   | +1 :green_heart: |  asflicense  |   0m 47s |  The patch does not generate 
ASF License warnings.  |
   |  |   | 144m  1s |   |
   
   
   | Reason | Tests |
   |-------:|:------|
   | FindBugs | module:hadoop-tools/hadoop-aws |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.policySetCount
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.inputPolicySet(int)
  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.inputPolicySet(int)
  At S3AInstrumentation.java:[line 804] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readExceptions
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readException()
  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readException()
  At S3AInstrumentation.java:[line 741] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readFullyOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readFullyOperationStarted(long,
 long)  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readFullyOperationStarted(long,
 long)  At S3AInstrumentation.java:[line 774] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readsIncomplete
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readOperationCompleted(int,
 int)  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readOperationCompleted(int,
 int)  At S3AInstrumentation.java:[line 785] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readOperationStarted(long,
 long)  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.readOperationStarted(long,
 long)  At S3AInstrumentation.java:[line 763] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.backwardSeekOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekBackwards(long)
  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekBackwards(long)
  At S3AInstrumentation.java:[line 686] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekBackwards(long)
  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekBackwards(long)
  At S3AInstrumentation.java:[line 685] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.forwardSeekOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekForwards(long)
  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekForwards(long)
  At S3AInstrumentation.java:[line 699] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekForwards(long)
  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.seekForwards(long)
  At S3AInstrumentation.java:[line 698] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.aborted 
in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamClose(boolean,
 long)  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamClose(boolean,
 long)  At S3AInstrumentation.java:[line 728] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.closeOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamClose(boolean,
 long)  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamClose(boolean,
 long)  At S3AInstrumentation.java:[line 726] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.closed in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamClose(boolean,
 long)  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamClose(boolean,
 long)  At S3AInstrumentation.java:[line 731] |
   |  |  Increment of volatile field 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.openOperations
 in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamOpened()
  At S3AInstrumentation.java:in 
org.apache.hadoop.fs.s3a.S3AInstrumentation$InputStreamStatisticsImpl.streamOpened()
  At S3AInstrumentation.java:[line 712] |
   | Failed junit tests | hadoop.security.TestFixKerberosTicketOrder |
   |   | hadoop.fs.TestFilterFileSystem |
   
   
   | Subsystem | Report/Notes |
   |----------:|:-------------|
   | Docker | Client=19.03.6 Server=19.03.6 base: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1820/6/artifact/out/Dockerfile
 |
   | GITHUB PR | https://github.com/apache/hadoop/pull/1820 |
   | Optional Tests | dupname asflicense compile javac javadoc mvninstall 
mvnsite unit shadedclient findbugs checkstyle |
   | uname | Linux a86708b0fcf1 4.15.0-74-generic #84-Ubuntu SMP Thu Dec 19 
08:06:28 UTC 2019 x86_64 x86_64 x86_64 GNU/Linux |
   | Build tool | maven |
   | Personality | personality/hadoop.sh |
   | git revision | trunk / 181e6d0 |
   | Default Java | 1.8.0_232 |
   | checkstyle | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1820/6/artifact/out/diff-checkstyle-root.txt
 |
   | findbugs | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1820/6/artifact/out/new-findbugs-hadoop-tools_hadoop-aws.html
 |
   | unit | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1820/6/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt
 |
   |  Test Results | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1820/6/testReport/ |
   | Max. process+thread count | 2397 (vs. ulimit of 5500) |
   | modules | C: hadoop-common-project/hadoop-common hadoop-tools/hadoop-aws 
U: . |
   | Console output | 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1820/6/console |
   | versions | git=2.7.4 maven=3.3.9 findbugs=3.1.0-RC1 |
   | Powered by | Apache Yetus 0.11.1 https://yetus.apache.org |
   
   
   This message was automatically generated.
   
   

----------------------------------------------------------------
This is an automated message from the Apache Git Service.
To respond to the message, please log on to GitHub and use the
URL above to go to the specific comment.
 
For queries about this service, please contact Infrastructure at:
[email protected]


With regards,
Apache Git Services

---------------------------------------------------------------------
To unsubscribe, e-mail: [email protected]
For additional commands, e-mail: [email protected]

Reply via email to