[jira] [Created] (HADOOP-15555) S3AFileStatus to add a serialVersionUID; review & test serialization

2018-06-21 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-1:
---

 Summary: S3AFileStatus to add a serialVersionUID; review & test 
serialization
 Key: HADOOP-1
 URL: https://issues.apache.org/jira/browse/HADOOP-1
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.1.0
Reporter: Steve Loughran


As FileStatus is now serializable, review S3AFilestatus
* add a version field
* review deserialization to see if we need any checks to stop invalid/malicious 
status instances being created



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15556) S3AFileStatus to add a serialVersionUID; review & test serialization

2018-06-21 Thread Steve Loughran (JIRA)
Steve Loughran created HADOOP-15556:
---

 Summary: S3AFileStatus to add a serialVersionUID; review & test 
serialization
 Key: HADOOP-15556
 URL: https://issues.apache.org/jira/browse/HADOOP-15556
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Affects Versions: 3.1.0
Reporter: Steve Loughran


As FileStatus is now serializable, review S3AFilestatus
* add a version field
* review deserialization to see if we need any checks to stop invalid/malicious 
status instances being created



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2018-06-21 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/

[Jun 20, 2018 5:42:13 PM] (haibochen) YARN-8437. Build oom-listener fails on 
older versions. (Miklos Szegedi
[Jun 20, 2018 5:59:33 PM] (miklos.szegedi) YARN-8391. Investigate 
AllocationFileLoaderService.reloadListener
[Jun 20, 2018 6:36:12 PM] (miklos.szegedi) YARN-8440. Typo in YarnConfiguration 
javadoc: "Miniumum request
[Jun 20, 2018 6:40:56 PM] (miklos.szegedi) YARN-7449. Split up class 
TestYarnClient to TestYarnClient and
[Jun 20, 2018 6:55:43 PM] (miklos.szegedi) YARN-8442. Strange characters and 
missing spaces in FairScheduler
[Jun 20, 2018 6:58:18 PM] (miklos.szegedi) YARN-8441. Typo in CSQueueUtils 
local variable names:
[Jun 20, 2018 7:04:44 PM] (miklos.szegedi) MAPREDUCE-7113. Typos in test names 
in TestTaskAttempt:
[Jun 20, 2018 10:45:08 PM] (mackrorysd) HADOOP-14918. Remove the Local Dynamo 
DB test option. Contributed by
[Jun 20, 2018 10:58:26 PM] (xiao) HDFS-13682. Cannot create encryption zone 
after KMS auth token expires.
[Jun 20, 2018 11:43:10 PM] (todd) HADOOP-15551. Avoid use of Arrays.stream in 
Configuration.addTags




-1 overall


The following subsystems voted -1:
asflicense findbugs pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.security.TestRaceWhenRelogin 
   hadoop.security.TestFixKerberosTicketOrder 
   hadoop.hdfs.server.datanode.TestDataNodeMultipleRegistrations 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure 
   hadoop.hdfs.client.impl.TestBlockReaderLocal 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.yarn.server.resourcemanager.applicationsmanager.TestAMRestart 
   
hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageEntities 
   
hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowActivity 
   hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowRun 
   hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageApps 
   
hadoop.yarn.server.timelineservice.storage.flow.TestHBaseStorageFlowRunCompaction
 
   
hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageSchema 
   
hadoop.yarn.server.timelineservice.storage.TestHBaseTimelineStorageDomain 
   
hadoop.yarn.server.timelineservice.reader.TestTimelineReaderWebServicesHBaseStorage
 
   hadoop.mapred.TestMRTimelineEventHandling 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/diff-compile-cc-root.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/diff-compile-javac-root.txt
  [352K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/diff-checkstyle-root.txt
  [4.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/diff-patch-pylint.txt
  [24K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/diff-patch-shellcheck.txt
  [20K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/diff-patch-shelldocs.txt
  [16K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/whitespace-eol.txt
  [9.4M]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/whitespace-tabs.txt
  [1.1M]

   xml:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/xml.txt
  [4.0K]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/branch-findbugs-hadoop-hdds_client.txt
  [56K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/branch-findbugs-hadoop-hdds_container-service.txt
  [48K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/branch-findbugs-hadoop-hdds_server-scm.txt
  [60K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/branch-findbugs-hadoop-hdds_tools.txt
  [12K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/branch-findbugs-hadoop-ozone_client.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/branch-findbugs-hadoop-ozone_common.txt
  [24K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/818/artifact/out/branch-findbugs-hadoop-ozone_objectstore-service.txt
  [4.0K]
   

Apache Hadoop qbt Report: trunk+JDK8 on Windows/x64

2018-06-21 Thread Apache Jenkins Server
For more details, see https://builds.apache.org/job/hadoop-trunk-win/504/

[Jun 20, 2018 5:42:13 PM] (haibochen) YARN-8437. Build oom-listener fails on 
older versions. (Miklos Szegedi
[Jun 20, 2018 5:59:33 PM] (miklos.szegedi) YARN-8391. Investigate 
AllocationFileLoaderService.reloadListener
[Jun 20, 2018 6:36:12 PM] (miklos.szegedi) YARN-8440. Typo in YarnConfiguration 
javadoc: "Miniumum request
[Jun 20, 2018 6:40:56 PM] (miklos.szegedi) YARN-7449. Split up class 
TestYarnClient to TestYarnClient and
[Jun 20, 2018 6:55:43 PM] (miklos.szegedi) YARN-8442. Strange characters and 
missing spaces in FairScheduler
[Jun 20, 2018 6:58:18 PM] (miklos.szegedi) YARN-8441. Typo in CSQueueUtils 
local variable names:
[Jun 20, 2018 7:04:44 PM] (miklos.szegedi) MAPREDUCE-7113. Typos in test names 
in TestTaskAttempt:
[Jun 20, 2018 10:45:08 PM] (mackrorysd) HADOOP-14918. Remove the Local Dynamo 
DB test option. Contributed by
[Jun 20, 2018 10:58:26 PM] (xiao) HDFS-13682. Cannot create encryption zone 
after KMS auth token expires.
[Jun 20, 2018 11:43:10 PM] (todd) HADOOP-15551. Avoid use of Arrays.stream in 
Configuration.addTags




-1 overall


The following subsystems voted -1:
compile mvninstall pathlen unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc javac


The following subsystems are considered long running:
(runtime bigger than 1h 00m 00s)
unit


Specific tests:

Failed junit tests :

   hadoop.crypto.TestCryptoStreamsWithOpensslAesCtrCryptoCodec 
   hadoop.fs.contract.rawlocal.TestRawlocalContractAppend 
   hadoop.fs.TestFileUtil 
   hadoop.fs.TestFsShellCopy 
   hadoop.fs.TestFsShellList 
   hadoop.fs.TestLocalFileSystem 
   hadoop.http.TestHttpServer 
   hadoop.http.TestHttpServerLogs 
   hadoop.io.nativeio.TestNativeIO 
   hadoop.ipc.TestIPC 
   hadoop.ipc.TestSocketFactory 
   hadoop.metrics2.impl.TestStatsDMetrics 
   hadoop.security.TestGroupsCaching 
   hadoop.security.TestSecurityUtil 
   hadoop.security.TestShellBasedUnixGroupsMapping 
   hadoop.security.token.TestDtUtilShell 
   hadoop.util.TestDiskCheckerWithDiskIo 
   hadoop.util.TestNativeCodeLoader 
   hadoop.hdfs.qjournal.server.TestJournalNode 
   hadoop.hdfs.qjournal.server.TestJournalNodeSync 
   hadoop.hdfs.server.blockmanagement.TestBlockManagerSafeMode 
   hadoop.hdfs.server.blockmanagement.TestNameNodePrunesMissingStorages 
   hadoop.hdfs.server.datanode.fsdataset.impl.TestProvidedImpl 
   hadoop.hdfs.server.datanode.TestBlockPoolSliceStorage 
   hadoop.hdfs.server.datanode.TestBlockScanner 
   hadoop.hdfs.server.datanode.TestDataNodeFaultInjector 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure 
   hadoop.hdfs.server.datanode.TestDirectoryScanner 
   hadoop.hdfs.server.datanode.TestNNHandlesCombinedBlockReport 
   hadoop.hdfs.server.diskbalancer.command.TestDiskBalancerCommand 
   hadoop.hdfs.server.diskbalancer.TestDiskBalancerRPC 
   hadoop.hdfs.server.namenode.ha.TestDFSUpgradeWithHA 
   hadoop.hdfs.server.namenode.ha.TestRetryCacheWithHA 
   hadoop.hdfs.server.namenode.metrics.TestNameNodeMetrics 
   hadoop.hdfs.server.namenode.TestNameNodeMXBean 
   hadoop.hdfs.TestDatanodeStartupFixesLegacyStorageIDs 
   hadoop.hdfs.TestDFSShell 
   hadoop.hdfs.TestDFSStripedInputStream 
   hadoop.hdfs.TestDFSStripedInputStreamWithRandomECPolicy 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailure 
   hadoop.hdfs.TestDFSStripedOutputStreamWithFailureWithRandomECPolicy 
   hadoop.hdfs.TestDFSUpgradeFromImage 
   hadoop.hdfs.TestFetchImage 
   hadoop.hdfs.TestFileConcurrentReader 
   hadoop.hdfs.TestHDFSFileSystemContract 
   hadoop.hdfs.TestLeaseRecovery 
   hadoop.hdfs.TestPread 
   hadoop.hdfs.TestSecureEncryptionZoneWithKMS 
   hadoop.hdfs.TestTrashWithSecureEncryptionZones 
   hadoop.hdfs.tools.TestDFSAdmin 
   hadoop.hdfs.web.TestWebHDFS 
   hadoop.hdfs.web.TestWebHdfsUrl 
   hadoop.fs.http.server.TestHttpFSServerWebServer 
   
hadoop.yarn.server.nodemanager.containermanager.launcher.TestContainerLaunch 
   hadoop.yarn.server.nodemanager.containermanager.TestAuxServices 
   hadoop.yarn.server.nodemanager.containermanager.TestContainerManager 
   hadoop.yarn.server.nodemanager.recovery.TestNMLeveldbStateStoreService 
   hadoop.yarn.server.nodemanager.TestContainerExecutor 
   hadoop.yarn.server.nodemanager.TestNodeManagerResync 
   hadoop.yarn.server.nodemanager.TestNodeStatusUpdater 
   hadoop.yarn.server.webproxy.amfilter.TestAmFilter 
   
hadoop.yarn.server.applicationhistoryservice.TestApplicationHistoryServer 
   
hadoop.yarn.server.timeline.security.TestTimelineAuthenticationFilterForV1 
   
hadoop.yarn.server.resourcemanager.scheduler.capacity.conf.TestFSSchedulerConfigurationStore
 
   

Re: HADOOP-14163 proposal for new hadoop.apache.org

2018-06-21 Thread Elek, Marton



Thank you very much to bump up this thread.


About [2]: (Just for the clarification) the content of the proposed 
website is exactly the same as the old one.


About [1]. I believe that the "mvn site" is perfect for the 
documentation but for website creation there are more simple and 
powerful tools.


Hugo has more simple compared to jekyll. Just one binary, without 
dependencies, works everywhere (mac, linux, windows)


Hugo has much more powerful compared to "mvn site". Easier to create/use 
more modern layout/theme, and easier to handle the content (for example 
new release announcements could be generated as part of the release 
process)


I think it's very low risk to try out a new approach for the site (and 
easy to rollback in case of problems)


Marton

ps: I just updated the patch/preview site with the recent releases:

***
* http://hadoop.anzix.net *
***

On 06/21/2018 01:27 AM, Vinod Kumar Vavilapalli wrote:

Got pinged about this offline.

Thanks for keeping at it, Marton!

I think there are two road-blocks here
  (1) Is the mechanism using which the website is built good enough - mvn-site 
/ hugo etc?
  (2) Is the new website good enough?

For (1), I just think we need more committer attention and get feedback rapidly 
and get it in.

For (2), how about we do it in a different way in the interest of progress?
  - We create a hadoop.apache.org/new-site/ where this new site goes.
  - We then modify the existing web-site to say that there is a new 
site/experience that folks can click on a link and navigate to
  - As this new website matures and gets feedback & fixes, we finally pull the 
plug at a later point of time when we think we are good to go.

Thoughts?

+Vinod


On Feb 16, 2018, at 3:10 AM, Elek, Marton  wrote:

Hi,

I would like to bump this thread up.

TLDR; There is a proposed version of a new hadoop site which is available from 
here: https://elek.github.io/hadoop-site-proposal/ and 
https://issues.apache.org/jira/browse/HADOOP-14163

Please let me know what you think about it.


Longer version:

This thread started long time ago to use a more modern hadoop site:

Goals were:

1. To make it easier to manage it (the release entries could be created by a 
script as part of the release process)
2. To use a better look-and-feel
3. Move it out from svn to git

I proposed to:

1. Move the existing site to git and generate it with hugo (which is a single, 
standalone binary)
2. Move both the rendered and source branches to git.
3. (Create a jenkins job to generate the site automatically)

NOTE: this is just about forrest based hadoop.apache.org, NOT about the 
documentation which is generated by mvn-site (as before)


I got multiple valuable feedback and I improved the proposed site according to 
the comments. Allen had some concerns about the used technologies (hugo vs. 
mvn-site) and I answered all the questions why I think mvn-site is the best for 
documentation and hugo is best for generating site.


I would like to finish this effort/jira: I would like to start a discussion 
about using this proposed version and approach as a new site of Apache Hadoop. 
Please let me know what you think.


Thanks a lot,
Marton

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org




-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15554) Improve JIT performance for Configuration parsing

2018-06-21 Thread Todd Lipcon (JIRA)
Todd Lipcon created HADOOP-15554:


 Summary: Improve JIT performance for Configuration parsing
 Key: HADOOP-15554
 URL: https://issues.apache.org/jira/browse/HADOOP-15554
 Project: Hadoop Common
  Issue Type: Improvement
  Components: conf, performance
Affects Versions: 3.0.0
Reporter: Todd Lipcon
Assignee: Todd Lipcon


In investigating a performance regression for small tasks between Hadoop 2 and 
Hadoop 3, we found that the amount of time spent in JIT was significantly 
higher. Using jitwatch we were able to determine that, due to a combination of 
switching from DOM to SAX style parsing and just having more configuration 
key/value pairs, Configuration.loadResource is now getting compiled with the C2 
compiler and taking quite some time. Breaking that very large function up into 
several smaller ones and eliminating some redundant bits of code improves the 
JIT performance measurably.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-15553) DOCO - not clear how to integrate hadoop-aws and aws-sdk jars with S3A

2018-06-21 Thread t oo (JIRA)
t oo created HADOOP-15553:
-

 Summary: DOCO - not clear how to integrate hadoop-aws and aws-sdk 
jars with S3A
 Key: HADOOP-15553
 URL: https://issues.apache.org/jira/browse/HADOOP-15553
 Project: Hadoop Common
  Issue Type: New Feature
  Components: tools
Affects Versions: 2.8.3
Reporter: t oo


Documentation update, a compatibility matrix would be nice, especially since 
upgrading to new aws java sdk does that work :

 

 On a hadoop common 2.7.5 cluster,  I take AWS Java SDK 
[1.7.5|https://mvnrepository.com/artifact/com.amazonaws/aws-java-sdk/1.7.5]  
and Hadoop-aws 
[2.7.5|https://mvnrepository.com/artifact/org.apache.hadoop/hadoop-aws/2.7.5] 
and use them in my Spark app writing to S3a, this is fine.

On another cluster (this time hadoop common 2.8.3 cluster) I take AWS Java SDK 
1.10.6  and Hadoop-aws 2.8.3 and try to use them in my Spark app writing to 
S3a, but get errors about missing classes. Turns out this version of aws java 
sdk is tiny with basically nothing inside.

Couldn't find any documentation helping users on this incompatibility.

 

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org