[jira] [Created] (HDFS-14274) EC: "hdfs dfs -ls -e" throws NPE When EC Policy of Directory set as "-replicate"

2019-02-12 Thread Souryakanta Dwivedy (JIRA)
Souryakanta Dwivedy created HDFS-14274:
--

 Summary: EC: "hdfs dfs -ls -e" throws NPE When EC Policy of 
Directory set as "-replicate"
 Key: HDFS-14274
 URL: https://issues.apache.org/jira/browse/HDFS-14274
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: erasure-coding
Affects Versions: 3.1.1
Reporter: Souryakanta Dwivedy


"hdfs dfs -ls -e" throws NPE When EC Policy of Directory set as "-replicate"

Steps :-

- Create a Directory 
- Set EC Policy for the Directory as "-replicate" 
- Check the folder details with command "hdfs dfs -ls -e "
 Null Pointer exception will display < ls: java.lang.NullPointerException >
 
Actual Result :- 
 ls: java.lang.NullPointerException
 
Expected Result :-
 Should not throw NullPointerException



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14273) Fix checkstyle issues in BlockLocation's method javadoc

2019-02-12 Thread Shweta (JIRA)
Shweta created HDFS-14273:
-

 Summary: Fix checkstyle issues in BlockLocation's method javadoc
 Key: HDFS-14273
 URL: https://issues.apache.org/jira/browse/HDFS-14273
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Shweta






--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-14269) [SBN read] Observer node switches back to Standby after restart

2019-02-12 Thread Wei-Chiu Chuang (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDFS-14269?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Wei-Chiu Chuang resolved HDFS-14269.

Resolution: Won't Fix

> [SBN read] Observer node switches back to Standby after restart
> ---
>
> Key: HDFS-14269
> URL: https://issues.apache.org/jira/browse/HDFS-14269
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: namenode
>Affects Versions: 3.3.0
>Reporter: Wei-Chiu Chuang
>Priority: Major
>
> The Observer state is not persistent. Once it restarts it becomes Standby 
> node again. Since it does not participate in NameNode failover for now, 
> should we assume an Observer node is always an Observer node? This state 
> should be persisted somewhere, like in Zookeeper.
> CC: [~csun] [~shv]. Like to get inputs from you. Have you discussed this 
> before?



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14272) [SBN read] HDFS command line tools does not guarantee consistency

2019-02-12 Thread Wei-Chiu Chuang (JIRA)
Wei-Chiu Chuang created HDFS-14272:
--

 Summary: [SBN read] HDFS command line tools does not guarantee 
consistency
 Key: HDFS-14272
 URL: https://issues.apache.org/jira/browse/HDFS-14272
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: tools
 Environment: CDH6.1 (Hadoop 3.0.x) + Consistency Reads from Standby + 
SSL + Kerberos + RPC encryption
Reporter: Wei-Chiu Chuang


It is typical for integration tests to create some files and then check its 
existence. For example, like the following simple bash script:
{code:java}
# hdfs dfs -touchz /tmp/abc
# hdfs dfs -ls /tmp/abc
{code}
The test execute HDFS bash command sequentially, but it may fail with 
Consistent Standby Read because the -ls may not find the file.

Analysis: the second bash command, while launched sequentially after the first 
one, is not aware of the state id returned from the first bash command. So 
ObserverNode wouldn't wait for the the edits to get propagated, and thus fails.

I've got a cluster where the Observer has tens of seconds of RPC latency, and 
this becomes very annoying. (I am still trying to figure out why this Observer 
has such a long RPC latency. But that's another story.)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14271) [SBN read] StandbyException is logged if Observer is the first NameNode

2019-02-12 Thread Wei-Chiu Chuang (JIRA)
Wei-Chiu Chuang created HDFS-14271:
--

 Summary: [SBN read] StandbyException is logged if Observer is the 
first NameNode
 Key: HDFS-14271
 URL: https://issues.apache.org/jira/browse/HDFS-14271
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: hdfs
Affects Versions: 3.3.0
Reporter: Wei-Chiu Chuang


If I transition the first NameNode into Observer state, and then I create a 
file from command line, it prints the following StandbyException log message, 
as if the command failed. But it actually completed successfully:
{noformat}
[root@weichiu-sbsr-1 ~]# hdfs dfs -touchz /tmp/abf
19/02/12 16:35:17 INFO retry.RetryInvocationHandler: 
org.apache.hadoop.ipc.RemoteException(org.apache.hadoop.ipc.StandbyException): 
Operation category WRITE is not supported in state observer. Visit 
https://s.apache.org/sbnn-error
at 
org.apache.hadoop.hdfs.server.namenode.ha.StandbyState.checkOperation(StandbyState.java:98)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode$NameNodeHAContext.checkOperation(NameNode.java:1987)
at 
org.apache.hadoop.hdfs.server.namenode.FSNamesystem.checkOperation(FSNamesystem.java:1424)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.create(NameNodeRpcServer.java:762)
at 
org.apache.hadoop.hdfs.protocolPB.ClientNamenodeProtocolServerSideTranslatorPB.create(ClientNamenodeProtocolServerSideTranslatorPB.java:458)
at 
org.apache.hadoop.hdfs.protocol.proto.ClientNamenodeProtocolProtos$ClientNamenodeProtocol$2.callBlockingMethod(ClientNamenodeProtocolProtos.java)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server$ProtoBufRpcInvoker.call(ProtobufRpcEngine.java:530)
at org.apache.hadoop.ipc.RPC$Server.call(RPC.java:1070)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:918)
at org.apache.hadoop.ipc.Server$RpcCall.run(Server.java:853)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1729)
at org.apache.hadoop.ipc.Server$Handler.run(Server.java:2782)
, while invoking $Proxy4.create over 
[weichiu-sbsr-1.gce.cloudera.com/172.31.121.145:8020,weichiu-sbsr-2.gce.cloudera.com/172.31.121.140:8020].
 Trying to failover immediately.
{noformat}

This is unlike the case when the first NameNode is the Standby, where this 
StandbyException is suppressed.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14270) StateId and TrasactionId not present in Trace level logging

2019-02-12 Thread Shweta (JIRA)
Shweta created HDFS-14270:
-

 Summary: StateId and TrasactionId not present in Trace level 
logging
 Key: HDFS-14270
 URL: https://issues.apache.org/jira/browse/HDFS-14270
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode
Reporter: Shweta


While running the command "hdfs --loglevel TRACE dfs -ls /" it was seen that 
stateId and TransactionId do not appear in the logs. 
How does one see the 
stateId and TransactionId in the logs? Is there a different approach?



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: DFSClient/DistriburedFileSystem fault injection?

2019-02-12 Thread Stephen Loughran
Sergey -you trying to simulate failures client side, or do you have an NN
Which actually injects failures all the way up the IPC stack?

as if its just client, couldn't registering a fault-injecting client as
fs.hdfs.impl could do that

FWIW, in the s3a connector we have the "inconsistent" s3 client which
mimics some symptoms of delayed consistency; it has a path, a probability
of happening
and a delay before things become visible. This is in the main hadoop-aws
JAR, and is turned on by a configuration switch (yes, it prints a big
warning). With a single switch to turn it on, its trivial to enable it in
tests

On Mon, Feb 11, 2019 at 11:42 PM Sergey Shelukhin
 wrote:

> Hi.
> I've been looking for a client-side solution for fault injection in HDFS.
> We had a naturally unstable HDFS cluster that helped uncover a lot of
> issues in HBase; now that it has been stabilized, we miss it already :)
>
> To keep testing without actually disrupting others' use of HDFS or having
> to deploy a new version, I was thinking about having a client-side schema
> (e.g. fhdfs) map to a wrapper over the standard DFS that would inject
> failures and delays according to some configs, similar to
> https://hadoop.apache.org/docs/r2.7.2/hadoop-project-dist/hadoop-hdfs/FaultInjectFramework.html
>
> However I wonder if something like this exists already?
>
> -
> To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
> For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
>
>


[jira] [Created] (HDFS-14269) [SBN read] Observer node switches back to Standby after restart

2019-02-12 Thread Wei-Chiu Chuang (JIRA)
Wei-Chiu Chuang created HDFS-14269:
--

 Summary: [SBN read] Observer node switches back to Standby after 
restart
 Key: HDFS-14269
 URL: https://issues.apache.org/jira/browse/HDFS-14269
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Affects Versions: 3.3.0
Reporter: Wei-Chiu Chuang


The Observer state is not persistent. Once it restarts it becomes Standby node 
again. Since it does not participate in NameNode failover for now, should we 
assume an Observer node is always an Observer node? This state should be 
persisted somewhere, like in Zookeeper.

CC: [~csun] [~shv]. Like to get inputs from you. Have you discussed this before?



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1063) Implement OM init in secure cluster

2019-02-12 Thread Xiaoyu Yao (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1063?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao resolved HDDS-1063.
--
Resolution: Duplicate

> Implement OM init in secure cluster
> ---
>
> Key: HDDS-1063
> URL: https://issues.apache.org/jira/browse/HDDS-1063
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>Reporter: Ajay Kumar
>Assignee: Ajay Kumar
>Priority: Major
>
> Implement OM init in secure cluster.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-581) Bootstrap DN with private/public key pair

2019-02-12 Thread Xiaoyu Yao (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-581?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao resolved HDDS-581.
-
Resolution: Duplicate

> Bootstrap DN with private/public key pair
> -
>
> Key: HDDS-581
> URL: https://issues.apache.org/jira/browse/HDDS-581
> Project: Hadoop Distributed Data Store
>  Issue Type: Sub-task
>Reporter: Xiaoyu Yao
>Assignee: Ajay Kumar
>Priority: Major
> Attachments: HDDS-581-HDDS-4.00.patch
>
>
> This will create public/private key pair for HDDS datanode if there isn't one 
> available during secure dn startup.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: DFSClient/DistriburedFileSystem fault injection?

2019-02-12 Thread Wei-Chiu Chuang
For what is worth,
At Cloudera we have an internal fault injection tool that is based on Chaos
Monkey.
We use it to kill disks or kill nodes, for example. So it doesn't
instrument HDFS directly.

On Mon, Feb 11, 2019 at 3:42 PM Sergey Shelukhin
 wrote:

> Hi.
> I've been looking for a client-side solution for fault injection in HDFS.
> We had a naturally unstable HDFS cluster that helped uncover a lot of
> issues in HBase; now that it has been stabilized, we miss it already :)
>
> To keep testing without actually disrupting others' use of HDFS or having
> to deploy a new version, I was thinking about having a client-side schema
> (e.g. fhdfs) map to a wrapper over the standard DFS that would inject
> failures and delays according to some configs, similar to
> https://hadoop.apache.org/docs/r2.7.2/hadoop-project-dist/hadoop-hdfs/FaultInjectFramework.html
>
> However I wonder if something like this exists already?
>
> -
> To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
> For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
>
>


RE: DFSClient/DistriburedFileSystem fault injection?

2019-02-12 Thread Sergey Shelukhin
Adding the user list :)

-Original Message-
From: Sergey Shelukhin 
Sent: Monday, February 11, 2019 3:42 PM
To: hdfs-dev@hadoop.apache.org
Subject: DFSClient/DistriburedFileSystem fault injection?

Hi.
I've been looking for a client-side solution for fault injection in HDFS.
We had a naturally unstable HDFS cluster that helped uncover a lot of issues in 
HBase; now that it has been stabilized, we miss it already :)

To keep testing without actually disrupting others' use of HDFS or having to 
deploy a new version, I was thinking about having a client-side schema (e.g. 
fhdfs) map to a wrapper over the standard DFS that would inject failures and 
delays according to some configs, similar to 
https://hadoop.apache.org/docs/r2.7.2/hadoop-project-dist/hadoop-hdfs/FaultInjectFramework.html
 

However I wonder if something like this exists already?

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14268) RBF: Fix the location of the DNs in getDatanodeReport()

2019-02-12 Thread JIRA
Íñigo Goiri created HDFS-14268:
--

 Summary: RBF: Fix the location of the DNs in getDatanodeReport()
 Key: HDFS-14268
 URL: https://issues.apache.org/jira/browse/HDFS-14268
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Íñigo Goiri
Assignee: Íñigo Goiri


When getting all the DNs in the federation, the Router queries each of the 
subclusters and aggregates them assigning the subcluster id to the location. 
This query uses a {{HashSet}} which provides a "random" order for the results.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14267) Add test_libhdfs_ops to libhdfs tests, mark libhdfs_read/write.c as examples

2019-02-12 Thread Sahil Takiar (JIRA)
Sahil Takiar created HDFS-14267:
---

 Summary: Add test_libhdfs_ops to libhdfs tests, mark 
libhdfs_read/write.c as examples
 Key: HDFS-14267
 URL: https://issues.apache.org/jira/browse/HDFS-14267
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: libhdfs, native, test
Reporter: Sahil Takiar


{{test_libhdfs_ops.c}} provides test coverage for basic operations against 
libhdfs, but currently has to be run manually (e.g. {{mvn install}} does not 
run these tests). The goal of this patch is to add {{test_libhdfs_ops.c}} to 
the list of tests that are automatically run for libhdfs.

It looks like {{test_libhdfs_ops.c}} was used in conjunction with 
{{hadoop-hdfs-project/hadoop-hdfs/src/main/native/tests/test-libhdfs.sh}} to 
run some tests against a mini DFS cluster. Now that the 
{{NativeMiniDfsCluster}} exists, it makes more sense to use that rather than 
rely on an external bash script to start a mini DFS cluster.

The {{libhdfs-tests}} directory (which contains {{test_libhdfs_ops.c}}) 
contains two other files: {{test_libhdfs_read.c}} and {{test_libhdfs_write.c}}. 
At some point, these files might have been used in conjunction with 
{{test-libhdfs.sh}} to run some tests manually. However, they (1) largely 
overlap with the test coverage provided by {{test_libhdfs_ops.c}} and (2) are 
not designed to be run as unit tests. Thus I suggest we move these two files 
into a new folder called {{libhdfs-examples}} and use them to further document 
how users of libhdfs can use the API. We can move {{test-libhdfs.sh}} into the 
examples folder as well given that example files probably require the script to 
actually work.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1094) Performance testing infrastructure : Special handling for zero-filled chunks on the Datanode

2019-02-12 Thread Supratim Deka (JIRA)
Supratim Deka created HDDS-1094:
---

 Summary: Performance testing infrastructure : Special handling for 
zero-filled chunks on the Datanode
 Key: HDDS-1094
 URL: https://issues.apache.org/jira/browse/HDDS-1094
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
  Components: Ozone Datanode
Reporter: Supratim Deka


Goal:
Make Ozone chunk Read/Write operations CPU/network bound for specially 
constructed performance micro benchmarks.
Remove disk bandwidth and latency constraints - running ozone data path against 
extreme low-latency & high throughput storage will expose performance 
bottlenecks in the flow. But low-latency storage(NVME flash drives, Storage 
class memory etc) is expensive and availability is limited. Is there a 
workaround which achieves similar running conditions for the software without 
actually having the low latency storage? At least for specially constructed 
datasets -  for example zero-filled blocks (*not* zero-length blocks).

Required characteristics of the solution:
No changes in Ozone client, OM and SCM. Changes limited to Datanode, Minimal 
footprint in datanode code.

Possible High level Approach:
The ChunkManager and ChunkUtils can enable writeChunk for zero-filled chunks to 
be dropped without actually writing to the local filesystem. Similarly, if 
readChunk can construct a zero-filled buffer without reading from the local 
filesystem whenever it detects a zero-filled chunk. Specifics of how to detect 
and record a zero-filled chunk can be discussed on this jira. Also discuss how 
to control this behaviour and make it available only for internal testing.




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-1003) Intermittent IO exceptions encountered during pre-commit tests

2019-02-12 Thread Mukul Kumar Singh (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-1003?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mukul Kumar Singh resolved HDDS-1003.
-
Resolution: Duplicate

> Intermittent IO exceptions encountered during pre-commit tests
> --
>
> Key: HDDS-1003
> URL: https://issues.apache.org/jira/browse/HDDS-1003
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>Reporter: Supratim Deka
>Priority: Major
>
> stack trace from 
> https://builds.apache.org/job/PreCommit-HDDS-Build/2095/testReport/org.apache.hadoop.ozone.client.rpc/TestOzoneRpcClient/testPutKey/
> java.io.IOException: Unexpected Storage Container Exception: 
> java.util.concurrent.ExecutionException: 
> org.apache.ratis.thirdparty.io.grpc.StatusRuntimeException: UNAVAILABLE: io 
> exception
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.writeChunkToContainer(BlockOutputStream.java:622)
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.writeChunk(BlockOutputStream.java:464)
>   at 
> org.apache.hadoop.hdds.scm.storage.BlockOutputStream.close(BlockOutputStream.java:480)
>   at 
> org.apache.hadoop.ozone.client.io.BlockOutputStreamEntry.close(BlockOutputStreamEntry.java:137)
>   at 
> org.apache.hadoop.ozone.client.io.KeyOutputStream.handleFlushOrClose(KeyOutputStream.java:481)
>   at 
> org.apache.hadoop.ozone.client.io.KeyOutputStream.handleWrite(KeyOutputStream.java:314)
>   at 
> org.apache.hadoop.ozone.client.io.KeyOutputStream.write(KeyOutputStream.java:258)
>   at 
> org.apache.hadoop.ozone.client.io.OzoneOutputStream.write(OzoneOutputStream.java:49)
>   at java.io.OutputStream.write(OutputStream.java:75)
>   at 
> org.apache.hadoop.ozone.client.rpc.TestOzoneRpcClientAbstract.testPutKey(TestOzoneRpcClientAbstract.java:522)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:168)
>   at org.junit.rules.RunRules.evaluate(RunRules.java:20)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
>   at 
> org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
>   at 
> org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
>   at 
> org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2019-02-12 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/

[Feb 11, 2019 3:17:08 AM] (bharat) HDDS-1078. TestRatisPipelineProvider failing 
because of node count
[Feb 11, 2019 3:34:53 AM] (bharat) HDDS-1077. TestSecureOzoneCluster does not 
config OM HTTP keytab.
[Feb 11, 2019 6:42:21 AM] (wwei) YARN-9252. Allocation Tag Namespace support in 
Distributed Shell.
[Feb 11, 2019 7:53:50 AM] (wwei) YARN-8555. Parameterize 
TestSchedulingRequestContainerAllocation(Async)
[Feb 11, 2019 11:18:26 AM] (elek) HDDS-1017. Use distributed tracing to 
indentify performance problems in
[Feb 11, 2019 2:38:25 PM] (msingh) HDDS-1040. Add blockade Tests for client 
failures. Contributed by
[Feb 11, 2019 6:09:44 PM] (inigoiri) HDFS-14260. Replace synchronized method in 
BlockReceiver with atomic
[Feb 11, 2019 7:02:32 PM] (weichiu) HDFS-14261. Kerberize JournalNodeSyncer 
unit test. Contributed by Siyao
[Feb 11, 2019 8:09:14 PM] (xyao) HDDS-1075. Fix 
CertificateUtil#parseRSAPublicKey charsetName.
[Feb 11, 2019 9:07:47 PM] (iwasakims) HADOOP-16098. Fix javadoc warnings in 
hadoop-aws. Contributed by
[Feb 11, 2019 10:54:30 PM] (sumasai) YARN-9229. Document docker registry 
deployment with NFS Gateway.
[Feb 11, 2019 11:27:43 PM] (gifuma) YARN-996. REST API support for node 
resource configuration. Contributed




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

Failed junit tests :

   hadoop.util.TestDiskChecker 
   hadoop.hdfs.server.namenode.TestPersistentStoragePolicySatisfier 
   hadoop.hdfs.TestFileCorruption 
   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.hdfs.server.datanode.TestDataNodeVolumeFailure 
   hadoop.hdfs.qjournal.server.TestJournalNodeSync 
   
hadoop.yarn.server.resourcemanager.scheduler.constraint.TestPlacementConstraintsUtil
 
   
hadoop.yarn.server.resourcemanager.scheduler.capacity.TestCapacityScheduler 
   hadoop.mapred.TestJobCounters 
   hadoop.yarn.sls.TestSLSStreamAMSynth 
   hadoop.yarn.sls.appmaster.TestAMSimulator 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/diff-compile-cc-root.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/diff-compile-javac-root.txt
  [336K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/diff-checkstyle-root.txt
  [17M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/diff-patch-hadolint.txt
  [8.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/diff-patch-pylint.txt
  [144K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/diff-patch-shellcheck.txt
  [20K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/diff-patch-shelldocs.txt
  [12K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/whitespace-eol.txt
  [9.6M]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/whitespace-tabs.txt
  [1.1M]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-hdds_client.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-hdds_container-service.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-hdds_framework.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-hdds_server-scm.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-hdds_tools.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-ozone_client.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-ozone_common.txt
  [4.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-ozone_objectstore-service.txt
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/1045/artifact/out/branch-findbugs-hadoop-ozone_ozone-manager.txt
  [4.0K]
   

[jira] [Created] (HDDS-1093) Configuration tab in OM ui is not displaying the correct values

2019-02-12 Thread Sandeep Nemuri (JIRA)
Sandeep Nemuri created HDDS-1093:


 Summary: Configuration tab in OM ui is not displaying the correct 
values
 Key: HDDS-1093
 URL: https://issues.apache.org/jira/browse/HDDS-1093
 Project: Hadoop Distributed Data Store
  Issue Type: Bug
  Components: OM
Reporter: Sandeep Nemuri
 Attachments: image-2019-02-12-19-47-18-332.png

Configuration tab in OM ui is not displaying the correct/configured values, 
rather it is displaying the default values.

!image-2019-02-12-19-47-18-332.png!


{code:java}
[hdfs@freonnode10 hadoop]$ curl -s http://freonnode10:9874/conf | grep 
ozone.om.handler.count.key
ozone.om.handler.count.key40falseozone-site.xml
{code}




--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1092) Use Java 11 JRE to run Ozone in containers

2019-02-12 Thread Elek, Marton (JIRA)
Elek, Marton created HDDS-1092:
--

 Summary: Use Java 11 JRE to run Ozone in containers
 Key: HDDS-1092
 URL: https://issues.apache.org/jira/browse/HDDS-1092
 Project: Hadoop Distributed Data Store
  Issue Type: Sub-task
Reporter: Elek, Marton
Assignee: Elek, Marton


As of now we use opendk 1.8.0 in the Ozone containers.

Java 9 and Java 10 introduces advanced support for the resource management of 
the containers and not all of them are available from the latest release of 
1.8.0. (see this blog for more details: 
https://medium.com/adorsys/jvm-memory-settings-in-a-container-environment-64b0840e1d9e)

I propose to switch to use Java 11 in the containers and test everything with 
Java 11 at runtime.

Note: this issue is just about the runtime jdk not about the compile time JDK.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1091) Provide centos based apache/hadoop-runner base image for ozone.

2019-02-12 Thread Elek, Marton (JIRA)
Elek, Marton created HDDS-1091:
--

 Summary: Provide centos based apache/hadoop-runner base image for 
ozone.
 Key: HDDS-1091
 URL: https://issues.apache.org/jira/browse/HDDS-1091
 Project: Hadoop Distributed Data Store
  Issue Type: Sub-task
Reporter: Elek, Marton
Assignee: Elek, Marton


Our [current base 
image|https://github.com/apache/hadoop/blob/docker-hadoop-runner/Dockerfile] 
[is based on debian 
|https://github.com/apache/hadoop/blob/docker-hadoop-runner/Dockerfile](openjdk:8-jdk).

Based on informal statistics the majority of the hadoop users and the potential 
Ozone users use centos/redhat. I propose to test Ozone with centos base image 
instead of debian one.

There are two options here:

1. Provide two base images (one for debian one for centos)
 2. Update the apache/hadoop-runner to use centos instead of debian

(1) is more generic but it requires more work to maintain. (2) is more simple, 
but we need to test docker-hadoop-2 and docker-hadoop-3 branches to be sure 
that they work together with centos base image.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: branch2+JDK7 on Linux/x86

2019-02-12 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/

No changes




-1 overall


The following subsystems voted -1:
asflicense findbugs hadolint pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

XML :

   Parsing Error(s): 
   
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/conf/empty-configuration.xml
 
   hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/public/crossdomain.xml 
   
hadoop-yarn-project/hadoop-yarn/hadoop-yarn-ui/src/main/webapp/public/crossdomain.xml
 

FindBugs :

   module:hadoop-common-project/hadoop-common 
   Class org.apache.hadoop.fs.GlobalStorageStatistics defines non-transient 
non-serializable instance field map In GlobalStorageStatistics.java:instance 
field map In GlobalStorageStatistics.java 

FindBugs :

   module:hadoop-hdfs-project/hadoop-hdfs 
   Dead store to state in 
org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(OutputStream,
 INodeSymlink) At 
FSImageFormatPBINode.java:org.apache.hadoop.hdfs.server.namenode.FSImageFormatPBINode$Saver.save(OutputStream,
 INodeSymlink) At FSImageFormatPBINode.java:[line 623] 

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-server/hadoop-yarn-server-timelineservice-hbase/hadoop-yarn-server-timelineservice-hbase-client
 
   Boxed value is unboxed and then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:then immediately reboxed in 
org.apache.hadoop.yarn.server.timelineservice.storage.common.ColumnRWHelper.readResultsWithTimestamps(Result,
 byte[], byte[], KeyConverter, ValueConverter, boolean) At 
ColumnRWHelper.java:[line 335] 

Failed junit tests :

   hadoop.hdfs.qjournal.server.TestJournalNodeRespectsBindHostKeys 
   hadoop.registry.secure.TestSecureLogins 
   hadoop.yarn.server.nodemanager.containermanager.TestContainerManager 
   hadoop.yarn.server.timelineservice.security.TestTimelineAuthFilterForV2 
   hadoop.yarn.client.api.impl.TestAMRMProxy 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-compile-cc-root-jdk1.7.0_95.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-compile-javac-root-jdk1.7.0_95.txt
  [328K]

   cc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-compile-cc-root-jdk1.8.0_191.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-compile-javac-root-jdk1.8.0_191.txt
  [308K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-checkstyle-root.txt
  [16M]

   hadolint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-patch-hadolint.txt
  [4.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-patch-pylint.txt
  [24K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-patch-shellcheck.txt
  [72K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-patch-shelldocs.txt
  [8.0K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/whitespace-eol.txt
  [12M]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/whitespace-tabs.txt
  [1.2M]

   xml:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/xml.txt
  [8.0K]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/branch-findbugs-hadoop-common-project_hadoop-common-warnings.html
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/branch-findbugs-hadoop-hdfs-project_hadoop-hdfs-warnings.html
  [8.0K]
   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-server_hadoop-yarn-server-timelineservice-hbase_hadoop-yarn-server-timelineservice-hbase-client-warnings.html
  [8.0K]

   javadoc:

   
https://builds.apache.org/job/hadoop-qbt-branch2-java7-linux-x86/230/artifact/out/diff-javadoc-javadoc-root-jdk1.7.0_95.txt
  [16K]
   

[jira] [Created] (HDDS-1090) Add a recipe to the ozone docs about using Spark, OzoneFS, and K8s

2019-02-12 Thread Elek, Marton (JIRA)
Elek, Marton created HDDS-1090:
--

 Summary: Add a recipe to the ozone docs about using Spark, 
OzoneFS, and K8s
 Key: HDDS-1090
 URL: https://issues.apache.org/jira/browse/HDDS-1090
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
Reporter: Elek, Marton
Assignee: Elek, Marton


With the isolated classloader (HDDS-922) it's possible to use the regular spark 
distribution (which includes hadoop-2.7) and ozonefs (based on hadoop-3.2) 
together.

In HDDS-846 we started to add '_recipes_' to the documentation to provide 
how-to style pages for different use-cases.

I would add a new page about using the following components together:
 * Spark2.4-hadoop-2.7
 * Ozonefs (isolated classloader)
 * Kubernetes spark executor



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDDS-1089) Disable OzoneFSStorageStatistics for hadoop versions older than 2.8

2019-02-12 Thread Elek, Marton (JIRA)
Elek, Marton created HDDS-1089:
--

 Summary: Disable OzoneFSStorageStatistics for hadoop versions 
older than 2.8
 Key: HDDS-1089
 URL: https://issues.apache.org/jira/browse/HDDS-1089
 Project: Hadoop Distributed Data Store
  Issue Type: Improvement
  Components: Ozone Filesystem
Reporter: Elek, Marton
Assignee: Elek, Marton


HDDS-1033 introduced OzoneFSStorageStatistics for OzoneFileSystem. It uses the 
StorageStatistics which is introduced in HADOOP-13065 (available from the 
hadoop2.8/3.0).

Using older hadoop (for example hadoop-2.7 which is included in the spark 
distributions) is not possible any more even with using the isolated class 
loader (introduced in HDDS-922).

Fortunately it can be fixed:
 # We can support null in storageStatistics field with checking everywhere 
before call it.
 # We can create a new constructor of OzoneClientAdapterImpl without using 
OzoneFSStorageStatistics): If OzoneFSStorageStatistics is not in the 
method/constructor signature we don't need to load it.
 # We can check the availability of HADOOP-13065 and if the classes are not in 
the classpath we can skip the initialization of the OzoneFSStorageStatistics



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org