Re: [DISCUSS] Making 2.10 the last minor 2.x release

2019-11-14 Thread Jonathan Hung
Some other additional items we would need:

   - Mark all fix-versions in YARN/HDFS/MAPREDUCE/HADOOP from 2.11.0 to
   2.10.1
   - Remove 2.11.0 as a version in these projects


Jonathan Hung


On Thu, Nov 14, 2019 at 6:51 PM Jonathan Hung  wrote:

> Hi folks,
>
> Given the release of 2.10.0, and the fact that it's intended to be a
> bridge release to Hadoop 3.x [1], I'm proposing we make 2.10.x the last
> minor release line in branch-2. Currently, the main issue is that there's
> many fixes going into branch-2 (the theoretical 2.11.0) that's not going
> into branch-2.10 (which will become 2.10.1), so the fixes in branch-2 will
> likely never see the light of day unless they are backported to branch-2.10.
>
> To do this, I propose we:
>
>- Delete branch-2.10
>- Rename branch-2 to branch-2.10
>- Set version in the new branch-2.10 to 2.10.1-SNAPSHOT
>
> This way we get all the current branch-2 fixes into the 2.10.x release
> line. Then the commit chain will look like: trunk -> branch-3.2 ->
> branch-3.1 -> branch-2.10 -> branch-2.9 -> branch-2.8
>
> Thoughts?
>
> Jonathan Hung
>
> [1] https://www.mail-archive.com/yarn-dev@hadoop.apache.org/msg29479.html
>


[DISCUSS] Making 2.10 the last minor 2.x release

2019-11-14 Thread Jonathan Hung
Hi folks,

Given the release of 2.10.0, and the fact that it's intended to be a bridge
release to Hadoop 3.x [1], I'm proposing we make 2.10.x the last minor
release line in branch-2. Currently, the main issue is that there's many
fixes going into branch-2 (the theoretical 2.11.0) that's not going into
branch-2.10 (which will become 2.10.1), so the fixes in branch-2 will
likely never see the light of day unless they are backported to branch-2.10.

To do this, I propose we:

   - Delete branch-2.10
   - Rename branch-2 to branch-2.10
   - Set version in the new branch-2.10 to 2.10.1-SNAPSHOT

This way we get all the current branch-2 fixes into the 2.10.x release
line. Then the commit chain will look like: trunk -> branch-3.2 ->
branch-3.1 -> branch-2.10 -> branch-2.9 -> branch-2.8

Thoughts?

Jonathan Hung

[1] https://www.mail-archive.com/yarn-dev@hadoop.apache.org/msg29479.html


[jira] [Created] (HADOOP-16712) Config ha.failover-controller.active-standby-elector.zk.op.retries is not in core-default.xml

2019-11-14 Thread Wei-Chiu Chuang (Jira)
Wei-Chiu Chuang created HADOOP-16712:


 Summary: Config 
ha.failover-controller.active-standby-elector.zk.op.retries is not in 
core-default.xml
 Key: HADOOP-16712
 URL: https://issues.apache.org/jira/browse/HADOOP-16712
 Project: Hadoop Common
  Issue Type: Improvement
Reporter: Wei-Chiu Chuang






--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Created] (HADOOP-16711) With S3Guard + authmode, consider skipping "verifyBuckets" check in S3A fs init()

2019-11-14 Thread Rajesh Balamohan (Jira)
Rajesh Balamohan created HADOOP-16711:
-

 Summary: With S3Guard + authmode, consider skipping 
"verifyBuckets" check in S3A fs init()
 Key: HADOOP-16711
 URL: https://issues.apache.org/jira/browse/HADOOP-16711
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: fs/s3
Reporter: Rajesh Balamohan


When authoritative mode is enabled with s3guard, it would be good to skip 
verifyBuckets call during S3A filesystem init(). This would save call to S3 
during init method.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Re: Questions about Apache Hadoop

2019-11-14 Thread Wei-Chiu Chuang
Please note that Hadoop is essentially an on-prem data infrastructure these
days. To fully leverage Hadoop you typically run other compute engines such
as Hive, Spark, HBase and so on. So YMMV.


Folks on the mailing list, please feel free to add/correct my comments.

• What type of secure connection will there be between Apache
Hadoop and VA systems in terms of secure protocols implemented?

Hadoop implements SSL/TLS up to TLS 1.2 as of Hadoop 3.2 for HTTP-based
connections. SPNEGO is supported too.

Separately, Hadoop's RPC protocol leverages Java SASL APIs to
authenticate/encrypt. Kerberos is supported via Java SASL.

Hadoop's data transfer leverages OpenSSL for data encryption.


There are additional custom protocols (Hadoop Delegation Token) that
authenticates users in a cluster.



• To what extent does Apache Hadoop use a FIPS 140-2 validated
cryptographic module, and what is the certification number?

Hadoop KMS is the cryptographic module in Hadoop. As far as I know it is
not FIPS 140-2 certified. You may need a commercial vendor for that purpose.



• What is the most recent version of Apache Hadoop and its
release date?

We maintain several branches.

https://hadoop.apache.org/releases.html



• Is there a Voluntary Product Accessibility Template (VPAT)
program in place to assess Section 508 compliance?

 I looked it up at Wikipedia and I still don't know what this is. You may
need to hire a consultant to help.


I suspect we don't, otherwise we wouldn't have used light green as the
header of HDFS NameNode web UI.



• What are the main features of Apache Hadoop?

https://hadoop.apache.org/

The Apache™ Hadoop® project develops open-source software for reliable,
scalable, distributed computing.

The Apache Hadoop software library is a framework that allows for the
distributed processing of large data sets across clusters of computers
using simple programming models. It is designed to scale up from single
servers to thousands of machines, each offering local computation and
storage. Rather than rely on hardware to deliver high-availability, the
library itself is designed to detect and handle failures at the application
layer, so delivering a highly-available service on top of a cluster of
computers, each of which may be prone to failures.



• What Cloud Service Provider (CSP) agreements have been set
for Apache Hadoop to be used securely through the cloud?

 Apache Hadoop is an open source project. You need a commercial vendor to
answer this question for you.


• Does Apache Hadoop offer an Application Program Interface
(API)?

Yes -- we offer Java API, C API and RESTFUL API.


• What other apps does Apache Hadoop integrate with?

 Many -- HBase, Hive, Spark, Presto and many others. Search for "Hadoop
ecosystem"

• What level of support does Apache Hadoop offer?

The project itself provides community support.


Commercial support is available via commercial vendors. Cloudera, for
example, offers several levels of technical support and services.


• Does Apache Hadoop leverage other database products?

 No. However, if you use Hive on top of Hadoop, Hive requires a metastore
that runs on a database.

• Is Apache Hadoop available for on-premise deployment?

 Yes.

• Does Apache Hadoop reside on user network?

Not sure what this means.


Thank you for your willingness to help by answering these questions. Please
note that I am working on a tight deadline, and must have my research
completed within three business days. If you would please acknowledge your
initial receipt of my email, it would be greatly appreciated.

Please contact me if you have any questions or concerns.



Best Regards,



Foday B. Fofanah (Contractor)

Senior Security Analyst (Prosphere)

Solution Delivery (Station 116) (005OPB14)

Office of Information and Technology, IT Operations and Services (ITOPS)

Office: (202) 461-4424








On Wed, Nov 13, 2019 at 2:07 PM Fofanah, Foday B. (Prosphere)
 wrote:

> Hello,
>
> I am reaching out to you from the Department of Veterans Affairs (VA)
> where I am part of the team that reviews various information-based products
> from an information security perspective for use within VA. I am reviewing
> information regarding Apache Hadoop and have several questions listed
> below; please respond to the best of your ability so that I may use your
> answers to reach a final determination.
>
> • What type of secure connection will there be between Apache
> Hadoop and VA systems in terms of secure protocols implemented?
>
>
>
> • To what extent does Apache Hadoop use a FIPS 140-2 validated
> cryptographic module, and what is the certification number?
>
>
>
> • What is the most recent version of Apache Hadoop and its
> release date?
>
>
>
> • Is there a Voluntary Product Accessibility Template (VPAT)
> 

[jira] [Created] (HADOOP-16710) testing_azure.md documentation is misleading

2019-11-14 Thread Andras Bokor (Jira)
Andras Bokor created HADOOP-16710:
-

 Summary: testing_azure.md documentation is misleading
 Key: HADOOP-16710
 URL: https://issues.apache.org/jira/browse/HADOOP-16710
 Project: Hadoop Common
  Issue Type: Bug
  Components: fs/azure
Affects Versions: 3.2.0
Reporter: Andras Bokor
Assignee: Andras Bokor


testing_azure.md states that "-Dparallel-tests" will run all the integration 
tests in parallel.

But in fact using -Dparallel-tests without any value actually skips the 
integration tests and runs only the unit tests.

The reason is that to activate a profile which is able to run ITs in parallel 
requires parallel-tests property to have a value (abfs, wasb or 'both'). 
sequential-tests profile defines !parallel-tests as value which means that the 
property should not even be mentioned.

Please check the output of help:active-profiles command:

 
{code:java}
cd hadoop-tools/hadoop-azure
andrasbokor$ mvn help:active-profiles -Dparallel-tests=abfs 
- parallel-tests-abfs (source: org.apache.hadoop:hadoop-azure:3.3.0-SNAPSHOT) 
- os.mac (source: org.apache.hadoop:hadoop-project:3.3.0-SNAPSHOT) 
- hbase1 (source: org.apache.hadoop:hadoop-project:3.3.0-SNAPSHOT) {code}
{code:java}
andrasbokor$ mvn help:active-profiles -Dparallel-tests
- os.mac (source: org.apache.hadoop:hadoop-project:3.3.0-SNAPSHOT)
- hbase1 (source: org.apache.hadoop:hadoop-project:3.3.0-SNAPSHOT)
{code}
{code:java}
mvn help:active-profiles
- sequential-tests (source: org.apache.hadoop:hadoop-azure:3.3.0-SNAPSHOT)
- os.mac (source: org.apache.hadoop:hadoop-project:3.3.0-SNAPSHOT)
- hbase1 (source: org.apache.hadoop:hadoop-project:3.3.0-SNAPSHOT){code}
The help:active-profiles shows that -Dparallel-tests does not add any IT 
related profiles and results in skipping all the integration tests during 
verify phrase.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



[jira] [Reopened] (HADOOP-16683) Disable retry of FailoverOnNetworkExceptionRetry in case of wrapped AccessControlException

2019-11-14 Thread Adam Antal (Jira)


 [ 
https://issues.apache.org/jira/browse/HADOOP-16683?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adam Antal reopened HADOOP-16683:
-

Let's backport this issue to lower branches.

> Disable retry of FailoverOnNetworkExceptionRetry in case of wrapped 
> AccessControlException
> --
>
> Key: HADOOP-16683
> URL: https://issues.apache.org/jira/browse/HADOOP-16683
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: common
>Affects Versions: 3.3.0
>Reporter: Adam Antal
>Assignee: Adam Antal
>Priority: Major
> Fix For: 3.3.0
>
> Attachments: HADOOP-16683.001.patch, HADOOP-16683.002.patch, 
> HADOOP-16683.003.patch
>
>
> Follow up patch on HADOOP-16580.
> We successfully disabled the retry in case of an AccessControlException which 
> has resolved some of the cases, but in other cases AccessControlException is 
> wrapped inside another IOException and you can only get the original 
> exception by calling getCause().
> Let's add this extra case as well.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: common-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: branch2+JDK7 on Linux/x86

2019-11-14 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/

No changes




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

XML :

   Parsing Error(s): 
   
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/empty-configuration.xml
 
   hadoop-tools/hadoop-azure/src/config/checkstyle-suppressions.xml 
   hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/public/crossdomain.xml 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/public/crossdomain.xml
 

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client
 
   Boxed value is unboxed and then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:[line 335] 

Failed junit tests :

   hadoop.util.TestReadWriteDiskValidator 
   hadoop.contrib.bkjournal.TestBookKeeperHACheckpoints 
   hadoop.hdfs.TestDistributedFileSystem 
   hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys 
   hadoop.contrib.bkjournal.TestBookKeeperHACheckpoints 
   hadoop.registry.secure.TestSecureLogins 
   hadoop.yarn.server.timelineservice.security.TestTimelineAuthFilterForV2 
   hadoop.yarn.client.api.impl.TestAMRMClient 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-compile-cc-root-jdk1.7.0_95.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-compile-javac-root-jdk1.7.0_95.txt
  [328K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-compile-cc-root-jdk1.8.0_222.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-compile-javac-root-jdk1.8.0_222.txt
  [308K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-checkstyle-root.txt
  [16M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-patch-hadolint.txt
  [4.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-patch-pylint.txt
  [24K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-patch-shellcheck.txt
  [72K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-patch-shelldocs.txt
  [8.0K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/whitespace-eol.txt
  [12M]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/whitespace-tabs.txt
  [1.3M]

   xml:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/xml.txt
  [12K]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timelineservice-hbase_hadoop-yarn-server-timelineservice-hbase-client-warnings.html
  [8.0K]

   javadoc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-javadoc-javadoc-root-jdk1.7.0_95.txt
  [16K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/diff-javadoc-javadoc-root-jdk1.8.0_222.txt
  [1.1M]

   unit:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/patch-unit-hadoop-common-project_hadoop-common.txt
  [160K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs.txt
  [232K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/505/artifact/out/patch-unit-hadoop-hdfs-project_hadoop-hdfs_src_contrib_bkjournal.txt
  [12K]