[jira] [Commented] (HADOOP-11828) Implement the Hitchhiker erasure coding algorithm

2016-01-21 Thread jack liuquan (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110481#comment-15110481
 ] 

jack liuquan commented on HADOOP-11828:
---

Hi [~zhz],
Thanks for your review!
bq.For an util class with all static methods, we don't need a constructor.
I add a private constructor for checkstyle issue, just reference to code of 
{{DumpUtil}} class in {{rawcoder}}
bq.The getPiggyBacksFromInput method is fairly complex and deserves a unit 
test. A ASCII illustration would also be very helpful, similar to Figure 4 in 
the Hitchhiker paper .
Although {{getPiggyBacksFromInput }} is fairly complex, there only one running 
branch in it. I think current unit test cases are good to cover it. I will add 
a ASCII illustration for it.
bq.3.Maybe I'm missing something, but how do we guarantee the length of inputs 
passed to performCoding is always numDataUnits * subPacketSize?
As Rashmi said, subPacketSize is always 2 in Hitchhiker. I think we can 
guarantee it when we preparing block chunks.

> Implement the Hitchhiker erasure coding algorithm
> -
>
> Key: HADOOP-11828
> URL: https://issues.apache.org/jira/browse/HADOOP-11828
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Zhe Zhang
>Assignee: jack liuquan
> Attachments: 7715-hitchhikerXOR-v2-testcode.patch, 
> 7715-hitchhikerXOR-v2.patch, HADOOP-11828-hitchhikerXOR-V3.patch, 
> HADOOP-11828-hitchhikerXOR-V4.patch, HADOOP-11828-hitchhikerXOR-V5.patch, 
> HADOOP-11828-hitchhikerXOR-V6.patch, HADOOP-11828-hitchhikerXOR-V7.patch, 
> HDFS-7715-hhxor-decoder.patch, HDFS-7715-hhxor-encoder.patch
>
>
> [Hitchhiker | 
> http://www.eecs.berkeley.edu/~nihar/publications/Hitchhiker_SIGCOMM14.pdf] is 
> a new erasure coding algorithm developed as a research project at UC 
> Berkeley. It has been shown to reduce network traffic and disk I/O by 25%-45% 
> during data reconstruction while retaining the same storage capacity and 
> failure tolerance capability as RS codes. This JIRA aims to introduce 
> Hitchhiker to the HDFS-EC framework, as one of the pluggable codec algorithms.
> The existing implementation is based on HDFS-RAID. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HADOOP-12726) Unsupported operations should throw UnsupportedOperationException

2016-01-21 Thread Daniel Templeton (JIRA)
Daniel Templeton created HADOOP-12726:
-

 Summary: Unsupported operations should throw 
UnsupportedOperationException
 Key: HADOOP-12726
 URL: https://issues.apache.org/jira/browse/HADOOP-12726
 Project: Hadoop Common
  Issue Type: Improvement
  Components: fs
Affects Versions: 2.7.1
Reporter: Daniel Templeton
Assignee: Daniel Templeton


In the {{FileSystem}} implementation classes, unsupported operations throw 
{{new IOException("Not supported")}}, which makes it needlessly difficult to 
distinguish an actual error from an unsupported operation.  They should instead 
throw {{new UnsupportedOperationException()}}.

It's possible that this anti-idiom is used elsewhere in the code base.  This 
JIRA should include finding and cleaning up those instances as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12712) Fix some cmake plugin and native build warnings

2016-01-21 Thread Varun Vasudev (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12712?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110401#comment-15110401
 ] 

Varun Vasudev commented on HADOOP-12712:


Thanks for catching this [~cmccabe]. My apologies for letting this slip 
through. Like Andrew, I suspect some implicit return was carried out by the 
compiler. I've cherry-picked the patch to 2.8 as well.

> Fix some cmake plugin and native build warnings
> ---
>
> Key: HADOOP-12712
> URL: https://issues.apache.org/jira/browse/HADOOP-12712
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: native
>Affects Versions: 2.4.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 2.8.0, 2.9.0
>
> Attachments: HADOOP-12712.001.patch, HADOOP-12712.002.patch, 
> HADOOP-12712.003.patch
>
>
> Fix some native build warnings



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12712) Fix some cmake plugin and native build warnings

2016-01-21 Thread Varun Vasudev (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12712?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Varun Vasudev updated HADOOP-12712:
---
Fix Version/s: 2.8.0

> Fix some cmake plugin and native build warnings
> ---
>
> Key: HADOOP-12712
> URL: https://issues.apache.org/jira/browse/HADOOP-12712
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: native
>Affects Versions: 2.4.0
>Reporter: Colin Patrick McCabe
>Assignee: Colin Patrick McCabe
>Priority: Minor
> Fix For: 2.8.0, 2.9.0
>
> Attachments: HADOOP-12712.001.patch, HADOOP-12712.002.patch, 
> HADOOP-12712.003.patch
>
>
> Fix some native build warnings



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12605) Fix intermittent failure of TestIPC.testIpcWithReaderQueuing

2016-01-21 Thread Masatake Iwasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12605?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Masatake Iwasaki updated HADOOP-12605:
--
  Resolution: Fixed
   Fix Version/s: 2.9.0
Target Version/s: 2.9.0
  Status: Resolved  (was: Patch Available)

Committed this to trunk and branch-2. Thanks, [~xiaochen] and 
[~ste...@apache.org].

> Fix intermittent failure of TestIPC.testIpcWithReaderQueuing
> 
>
> Key: HADOOP-12605
> URL: https://issues.apache.org/jira/browse/HADOOP-12605
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: test
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Minor
> Fix For: 2.9.0
>
> Attachments: HADOOP-12605.001.patch, HADOOP-12605.002.patch, 
> HADOOP-12605.003.patch, HADOOP-12605.004.patch, HADOOP-12605.005.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12605) Fix intermittent failure of TestIPC.testIpcWithReaderQueuing

2016-01-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12605?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110328#comment-15110328
 ] 

Hudson commented on HADOOP-12605:
-

FAILURE: Integrated in Hadoop-trunk-Commit #9146 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9146/])
HADOOP-12605. Fix intermittent failure of (iwasakims: rev 
9eec6cbedcc53e6dd306a4a578a22937ae239260)
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/ipc/TestIPC.java


> Fix intermittent failure of TestIPC.testIpcWithReaderQueuing
> 
>
> Key: HADOOP-12605
> URL: https://issues.apache.org/jira/browse/HADOOP-12605
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: test
>Reporter: Masatake Iwasaki
>Assignee: Masatake Iwasaki
>Priority: Minor
> Attachments: HADOOP-12605.001.patch, HADOOP-12605.002.patch, 
> HADOOP-12605.003.patch, HADOOP-12605.004.patch, HADOOP-12605.005.patch
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11505) Various native parts use bswap incorrectly and unportably

2016-01-21 Thread Alan Burlison (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11505?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110257#comment-15110257
 ] 

Alan Burlison commented on HADOOP-11505:


Ah, it's been pushed since my last sync, I'll resync - thanks.

> Various native parts use bswap incorrectly and unportably
> -
>
> Key: HADOOP-11505
> URL: https://issues.apache.org/jira/browse/HADOOP-11505
> Project: Hadoop Common
>  Issue Type: Bug
>Affects Versions: 3.0.0
>Reporter: Colin Patrick McCabe
>Assignee: Alan Burlison
> Fix For: 3.0.0
>
> Attachments: HADOOP-11505.001.patch, HADOOP-11505.003.patch, 
> HADOOP-11505.004.patch, HADOOP-11505.005.patch, HADOOP-11505.006.patch, 
> HADOOP-11505.007.patch
>
>
> hadoop-mapreduce-client-nativetask fails to use x86 optimizations in some 
> cases.  Also, on some alternate, non-x86, non-ARM architectures the generated 
> code is incorrect.  Thanks to Steve Loughran and Edward Nevill for finding 
> this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HADOOP-12727) Minor cleanups needed for CMake 3.X

2016-01-21 Thread Alan Burlison (JIRA)
Alan Burlison created HADOOP-12727:
--

 Summary: Minor cleanups needed for CMake 3.X
 Key: HADOOP-12727
 URL: https://issues.apache.org/jira/browse/HADOOP-12727
 Project: Hadoop Common
  Issue Type: Sub-task
  Components: native
Affects Versions: 2.7.1
Reporter: Alan Burlison
Assignee: Alan Burlison
Priority: Minor


On switching from CMake 2.8.6 to 3.3.2 a couple of minor issues popped up:
\\
\\
* There's a syntax error in 
{{hadoop-common-project/hadoop-common/src/CMakeLists.txt}} that generates a 
warning in 3.X

* {{CMAKE_SHARED_LINKER_FLAGS}} is being incorrectly set in 
{{hadoop-common-project/hadoop-common/HadoopCommon.cmake}} - despite the name 
it contains the flags passed to {{ar}} not to the linker. 2.8.6 ignores the 
incorrect flags, 3.3.2 doesn't and building static libraries fails as a result. 
See http://public.kitware.com/pipermail/cmake/2016-January/062447.html

Patch to follow.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12043) Display warning if defaultFs is not set when running fs commands.

2016-01-21 Thread Kazuho Fujii (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12043?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110717#comment-15110717
 ] 

Kazuho Fujii commented on HADOOP-12043:
---

Hi,

This function is useful. But, I think "fs.defaultFs" in the warning message 
should be the correct name "fs.defaultFS". A user who does not set fs.defaultFS 
property or mistypes the property name gets confused when seeing the message. 
At least I was misled.

How do you think?

> Display warning if defaultFs is not set when running fs commands.
> -
>
> Key: HADOOP-12043
> URL: https://issues.apache.org/jira/browse/HADOOP-12043
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs
>Affects Versions: 2.7.0
>Reporter: Lei (Eddy) Xu
>Assignee: Lei (Eddy) Xu
>Priority: Minor
> Fix For: 2.8.0
>
> Attachments: HDFS-8322.000.patch, HDFS-8322.001.patch, 
> HDFS-8322.002.patch, HDFS-8322.003.patch, HDFS-8322.003.patch, 
> HDFS-8322.004.patch, HDFS-8322.005.patch, HDFS-8322.006.patch
>
>
> Using {{LocalFileSystem}} is rarely the intention of running {{hadoop fs 
> -ls}}.
> This JIRA proposes displaying a warning message if hadoop fs -ls is showing 
> the local filesystem or using default fs.  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11683) Need a plugin API to translate long principal names to local OS user names arbitrarily

2016-01-21 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11683?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110802#comment-15110802
 ] 

Allen Wittenauer commented on HADOOP-11683:
---

I think it is important to recognize that principal -> username conversion 
happens all over the stack.  For example, every single web UI is going to be 
needing this functionality.  Unless I missed something, the way this code is 
written will require the mapping code+configuration to be present on every 
single node in a way that every single process is going to need access.  To 
make matters worse, if a non-Java AM decides to provide user auth (think 
Slider), it doesn't appear to have a way to access this functionality without 
using JNI.  

> Need a plugin API to translate long principal names to local OS user names 
> arbitrarily
> --
>
> Key: HADOOP-11683
> URL: https://issues.apache.org/jira/browse/HADOOP-11683
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.6.0
>Reporter: Sunny Cheung
>Assignee: roger mak
> Attachments: HADOOP-11683.001.patch, HADOOP-11683.002.patch, 
> HADOOP-11683.003.patch
>
>
> We need a plugin API to translate long principal names (e.g. 
> john@example.com) to local OS user names (e.g. user123456) arbitrarily.
> For some organizations the name translation is straightforward (e.g. 
> john@example.com to john_doe), and the hadoop.security.auth_to_local 
> configurable mapping is sufficient to resolve this (see HADOOP-6526). 
> However, in some other cases the name translation is arbitrary and cannot be 
> generalized by a set of translation rules easily.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12423) Handle failure of registering shutdownhook by ShutdownHookManager in static block

2016-01-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12423?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110731#comment-15110731
 ] 

Hudson commented on HADOOP-12423:
-

SUCCESS: Integrated in Hadoop-trunk-Commit #9148 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9148/])
HADOOP-12423. Handle failure of registering shutdownhook by (vinayakumarb: rev 
446987e20aacd870e1bc3f950b2bd4bbda1f9571)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/util/ShutdownHookManager.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> Handle failure of registering shutdownhook by ShutdownHookManager in static 
> block
> -
>
> Key: HADOOP-12423
> URL: https://issues.apache.org/jira/browse/HADOOP-12423
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 2.7.1
>Reporter: Abhishek Agarwal
>Assignee: Abhishek Agarwal
>Priority: Minor
> Fix For: 2.8.0
>
> Attachments: HADOOP-12423.patch
>
>
> If JVM is under shutdown, static method in ShutdownHookManager will throw an 
> IllegalStateException. This exception should be caught and ignored while 
> registering the hooks. 
> Stack trace: 
> {noformat}
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.hadoop.util.ShutdownHookManager
>at 
> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2639) 
> ~[stormjar.jar:1.4.0-SNAPSHOT]
>at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2612) 
> ~[stormjar.jar:1.4.0-SNAPSHOT]
>at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370) 
> ~[stormjar.jar:1.4.0-SNAPSHOT]
>...
>...
>at 
> backtype.storm.daemon.executor$fn__6647$fn__6659.invoke(executor.clj:692) 
> ~[storm-core-0.9.5.jar:0.9.5]
>at backtype.storm.util$async_loop$fn__459.invoke(util.clj:461) 
> ~[storm-core-0.9.5.jar:0.9.5]
>at clojure.lang.AFn.run(AFn.java:24) [clojure-1.5.1.jar:na]
>at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12423) Handle failure of registering shutdownhook by ShutdownHookManager in static block

2016-01-21 Thread Vinayakumar B (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12423?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Vinayakumar B updated HADOOP-12423:
---
   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

Committed to trunk, branch-2 and branch-2.8
thanks for the contribution [~abhishek.agarwal]
thanks [~lakshman] for review.

> Handle failure of registering shutdownhook by ShutdownHookManager in static 
> block
> -
>
> Key: HADOOP-12423
> URL: https://issues.apache.org/jira/browse/HADOOP-12423
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: fs
>Affects Versions: 2.7.1
>Reporter: Abhishek Agarwal
>Assignee: Abhishek Agarwal
>Priority: Minor
> Fix For: 2.8.0
>
> Attachments: HADOOP-12423.patch
>
>
> If JVM is under shutdown, static method in ShutdownHookManager will throw an 
> IllegalStateException. This exception should be caught and ignored while 
> registering the hooks. 
> Stack trace: 
> {noformat}
> java.lang.NoClassDefFoundError: Could not initialize class 
> org.apache.hadoop.util.ShutdownHookManager
>at 
> org.apache.hadoop.fs.FileSystem$Cache.getInternal(FileSystem.java:2639) 
> ~[stormjar.jar:1.4.0-SNAPSHOT]
>at org.apache.hadoop.fs.FileSystem$Cache.get(FileSystem.java:2612) 
> ~[stormjar.jar:1.4.0-SNAPSHOT]
>at org.apache.hadoop.fs.FileSystem.get(FileSystem.java:370) 
> ~[stormjar.jar:1.4.0-SNAPSHOT]
>...
>...
>at 
> backtype.storm.daemon.executor$fn__6647$fn__6659.invoke(executor.clj:692) 
> ~[storm-core-0.9.5.jar:0.9.5]
>at backtype.storm.util$async_loop$fn__459.invoke(util.clj:461) 
> ~[storm-core-0.9.5.jar:0.9.5]
>at clojure.lang.AFn.run(AFn.java:24) [clojure-1.5.1.jar:na]
>at java.lang.Thread.run(Thread.java:745) [na:1.8.0_60]
> {noformat}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-7161) Remove unnecessary oro dependency

2016-01-21 Thread Akira AJISAKA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-7161?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15110966#comment-15110966
 ] 

Akira AJISAKA commented on HADOOP-7161:
---

+1, committing this.

> Remove unnecessary oro dependency
> -
>
> Key: HADOOP-7161
> URL: https://issues.apache.org/jira/browse/HADOOP-7161
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 0.23.0
>Reporter: Todd Lipcon
>Assignee: Sean Busbey
>Priority: Minor
>  Labels: beginner
> Fix For: 3.0.0
>
> Attachments: HADOOP-7161.1.patch
>
>
> Best I can tell we never use the "oro" dependency, but it's been in ivy.xml 
> forever. Does anyone know any reason we might need it?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-7161) Remove unnecessary oro package from dependency management section

2016-01-21 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-7161?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HADOOP-7161:
--
   Resolution: Fixed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

Committed this to trunk, branch-2, and branch-2.8. Thanks [~busbey] for the 
contribution!

> Remove unnecessary oro package from dependency management section
> -
>
> Key: HADOOP-7161
> URL: https://issues.apache.org/jira/browse/HADOOP-7161
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 0.23.0, 2.7.1
>Reporter: Todd Lipcon
>Assignee: Sean Busbey
>Priority: Minor
>  Labels: newbie
> Fix For: 2.8.0
>
> Attachments: HADOOP-7161.1.patch
>
>
> Best I can tell we never use the "oro" dependency, but it's been in ivy.xml 
> forever. Does anyone know any reason we might need it?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-11828) Implement the Hitchhiker erasure coding algorithm

2016-01-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-11828?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111085#comment-15111085
 ] 

Hudson commented on HADOOP-11828:
-

FAILURE: Integrated in Hadoop-trunk-Commit #9153 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9153/])
HADOOP-11828. Implement the Hitchhiker erasure coding algorithm. (zhz: rev 
1bb31fb22e6f8e6df8e9ff4e94adf20308b4c743)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/coder/util/HHUtil.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/coder/AbstractErasureDecoder.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/coder/HHXORErasureDecodingStep.java
* hadoop-common-project/hadoop-common/CHANGES.txt
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/erasurecode/coder/TestHHErasureCoderBase.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/coder/HHXORErasureDecoder.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/coder/AbstractHHErasureCodingStep.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/erasurecode/coder/TestErasureCoderBase.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/coder/HHXORErasureEncodingStep.java
* 
hadoop-common-project/hadoop-common/src/test/java/org/apache/hadoop/io/erasurecode/coder/TestHHXORErasureCoder.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/io/erasurecode/coder/HHXORErasureEncoder.java


> Implement the Hitchhiker erasure coding algorithm
> -
>
> Key: HADOOP-11828
> URL: https://issues.apache.org/jira/browse/HADOOP-11828
> Project: Hadoop Common
>  Issue Type: Sub-task
>Affects Versions: 3.0.0
>Reporter: Zhe Zhang
>Assignee: jack liuquan
> Fix For: 3.0.0
>
> Attachments: 7715-hitchhikerXOR-v2-testcode.patch, 
> 7715-hitchhikerXOR-v2.patch, HADOOP-11828-hitchhikerXOR-V3.patch, 
> HADOOP-11828-hitchhikerXOR-V4.patch, HADOOP-11828-hitchhikerXOR-V5.patch, 
> HADOOP-11828-hitchhikerXOR-V6.patch, HADOOP-11828-hitchhikerXOR-V7.patch, 
> HADOOP-11828-v8.patch, HDFS-7715-hhxor-decoder.patch, 
> HDFS-7715-hhxor-encoder.patch
>
>
> [Hitchhiker | 
> http://www.eecs.berkeley.edu/~nihar/publications/Hitchhiker_SIGCOMM14.pdf] is 
> a new erasure coding algorithm developed as a research project at UC 
> Berkeley. It has been shown to reduce network traffic and disk I/O by 25%-45% 
> during data reconstruction while retaining the same storage capacity and 
> failure tolerance capability as RS codes. This JIRA aims to introduce 
> Hitchhiker to the HDFS-EC framework, as one of the pluggable codec algorithms.
> The existing implementation is based on HDFS-RAID. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11828) Implement the Hitchhiker erasure coding algorithm

2016-01-21 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HADOOP-11828:
---
Affects Version/s: 3.0.0

> Implement the Hitchhiker erasure coding algorithm
> -
>
> Key: HADOOP-11828
> URL: https://issues.apache.org/jira/browse/HADOOP-11828
> Project: Hadoop Common
>  Issue Type: Sub-task
>Affects Versions: 3.0.0
>Reporter: Zhe Zhang
>Assignee: jack liuquan
> Fix For: 3.0.0
>
> Attachments: 7715-hitchhikerXOR-v2-testcode.patch, 
> 7715-hitchhikerXOR-v2.patch, HADOOP-11828-hitchhikerXOR-V3.patch, 
> HADOOP-11828-hitchhikerXOR-V4.patch, HADOOP-11828-hitchhikerXOR-V5.patch, 
> HADOOP-11828-hitchhikerXOR-V6.patch, HADOOP-11828-hitchhikerXOR-V7.patch, 
> HADOOP-11828-v8.patch, HDFS-7715-hhxor-decoder.patch, 
> HDFS-7715-hhxor-encoder.patch
>
>
> [Hitchhiker | 
> http://www.eecs.berkeley.edu/~nihar/publications/Hitchhiker_SIGCOMM14.pdf] is 
> a new erasure coding algorithm developed as a research project at UC 
> Berkeley. It has been shown to reduce network traffic and disk I/O by 25%-45% 
> during data reconstruction while retaining the same storage capacity and 
> failure tolerance capability as RS codes. This JIRA aims to introduce 
> Hitchhiker to the HDFS-EC framework, as one of the pluggable codec algorithms.
> The existing implementation is based on HDFS-RAID. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11828) Implement the Hitchhiker erasure coding algorithm

2016-01-21 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HADOOP-11828:
---
  Resolution: Fixed
Hadoop Flags: Reviewed
   Fix Version/s: 3.0.0
Target Version/s: 3.0.0
  Status: Resolved  (was: Patch Available)

> Implement the Hitchhiker erasure coding algorithm
> -
>
> Key: HADOOP-11828
> URL: https://issues.apache.org/jira/browse/HADOOP-11828
> Project: Hadoop Common
>  Issue Type: Sub-task
>Affects Versions: 3.0.0
>Reporter: Zhe Zhang
>Assignee: jack liuquan
> Fix For: 3.0.0
>
> Attachments: 7715-hitchhikerXOR-v2-testcode.patch, 
> 7715-hitchhikerXOR-v2.patch, HADOOP-11828-hitchhikerXOR-V3.patch, 
> HADOOP-11828-hitchhikerXOR-V4.patch, HADOOP-11828-hitchhikerXOR-V5.patch, 
> HADOOP-11828-hitchhikerXOR-V6.patch, HADOOP-11828-hitchhikerXOR-V7.patch, 
> HADOOP-11828-v8.patch, HDFS-7715-hhxor-decoder.patch, 
> HDFS-7715-hhxor-encoder.patch
>
>
> [Hitchhiker | 
> http://www.eecs.berkeley.edu/~nihar/publications/Hitchhiker_SIGCOMM14.pdf] is 
> a new erasure coding algorithm developed as a research project at UC 
> Berkeley. It has been shown to reduce network traffic and disk I/O by 25%-45% 
> during data reconstruction while retaining the same storage capacity and 
> failure tolerance capability as RS codes. This JIRA aims to introduce 
> Hitchhiker to the HDFS-EC framework, as one of the pluggable codec algorithms.
> The existing implementation is based on HDFS-RAID. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-7161) Remove unnecessary oro package from dependency management section

2016-01-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-7161?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111029#comment-15111029
 ] 

Hudson commented on HADOOP-7161:


FAILURE: Integrated in Hadoop-trunk-Commit #9152 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9152/])
HADOOP-7161. Remove unnecessary oro package from dependency management 
(aajisaka: rev 2ac39ca762fd7e47b7b9a7b3308b31138b0f59be)
* hadoop-common-project/hadoop-common/CHANGES.txt
* hadoop-project/pom.xml


> Remove unnecessary oro package from dependency management section
> -
>
> Key: HADOOP-7161
> URL: https://issues.apache.org/jira/browse/HADOOP-7161
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 0.23.0, 2.7.1
>Reporter: Todd Lipcon
>Assignee: Sean Busbey
>Priority: Minor
>  Labels: newbie
> Fix For: 2.8.0
>
> Attachments: HADOOP-7161.1.patch
>
>
> Best I can tell we never use the "oro" dependency, but it's been in ivy.xml 
> forever. Does anyone know any reason we might need it?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-11828) Implement the Hitchhiker erasure coding algorithm

2016-01-21 Thread Zhe Zhang (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-11828?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Zhe Zhang updated HADOOP-11828:
---
Attachment: HADOOP-11828-v8.patch

Thanks for the discussions Rashmi and Jack.

I just added the license header to {{HHXORErasureDecodingStep}} and committed 
the v8 patch.

This is really exciting work! Let's plan on getting some performance results.

> Implement the Hitchhiker erasure coding algorithm
> -
>
> Key: HADOOP-11828
> URL: https://issues.apache.org/jira/browse/HADOOP-11828
> Project: Hadoop Common
>  Issue Type: Sub-task
>Reporter: Zhe Zhang
>Assignee: jack liuquan
> Attachments: 7715-hitchhikerXOR-v2-testcode.patch, 
> 7715-hitchhikerXOR-v2.patch, HADOOP-11828-hitchhikerXOR-V3.patch, 
> HADOOP-11828-hitchhikerXOR-V4.patch, HADOOP-11828-hitchhikerXOR-V5.patch, 
> HADOOP-11828-hitchhikerXOR-V6.patch, HADOOP-11828-hitchhikerXOR-V7.patch, 
> HADOOP-11828-v8.patch, HDFS-7715-hhxor-decoder.patch, 
> HDFS-7715-hhxor-encoder.patch
>
>
> [Hitchhiker | 
> http://www.eecs.berkeley.edu/~nihar/publications/Hitchhiker_SIGCOMM14.pdf] is 
> a new erasure coding algorithm developed as a research project at UC 
> Berkeley. It has been shown to reduce network traffic and disk I/O by 25%-45% 
> during data reconstruction while retaining the same storage capacity and 
> failure tolerance capability as RS codes. This JIRA aims to introduce 
> Hitchhiker to the HDFS-EC framework, as one of the pluggable codec algorithms.
> The existing implementation is based on HDFS-RAID. 



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-7161) Remove unnecessary oro package from dependency management section

2016-01-21 Thread Akira AJISAKA (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-7161?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Akira AJISAKA updated HADOOP-7161:
--
Affects Version/s: 2.7.1
   Labels: newbie  (was: beginner)
 Hadoop Flags: Reviewed
Fix Version/s: (was: 3.0.0)
  Summary: Remove unnecessary oro package from dependency 
management section  (was: Remove unnecessary oro dependency)

> Remove unnecessary oro package from dependency management section
> -
>
> Key: HADOOP-7161
> URL: https://issues.apache.org/jira/browse/HADOOP-7161
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: build
>Affects Versions: 0.23.0, 2.7.1
>Reporter: Todd Lipcon
>Assignee: Sean Busbey
>Priority: Minor
>  Labels: newbie
> Attachments: HADOOP-7161.1.patch
>
>
> Best I can tell we never use the "oro" dependency, but it's been in ivy.xml 
> forever. Does anyone know any reason we might need it?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12726) Unsupported operations should throw UnsupportedOperationException

2016-01-21 Thread Steve Loughran (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12726?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111051#comment-15111051
 ] 

Steve Loughran commented on HADOOP-12726:
-

in favour for all FS operations. In an ideal world, you'd even be able to tell 
in advance what was unsupported, but as it may depend on the specific FS 
instance, that's a tough one, especially for rawfs

> Unsupported operations should throw UnsupportedOperationException
> -
>
> Key: HADOOP-12726
> URL: https://issues.apache.org/jira/browse/HADOOP-12726
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: fs
>Affects Versions: 2.7.1
>Reporter: Daniel Templeton
>Assignee: Daniel Templeton
>
> In the {{FileSystem}} implementation classes, unsupported operations throw 
> {{new IOException("Not supported")}}, which makes it needlessly difficult to 
> distinguish an actual error from an unsupported operation.  They should 
> instead throw {{new UnsupportedOperationException()}}.
> It's possible that this anti-idiom is used elsewhere in the code base.  This 
> JIRA should include finding and cleaning up those instances as well.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12728) "hadoop jar my.jar" should probably prioritize my.jar in the classpath by default

2016-01-21 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12728?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-12728:
--
Target Version/s: 3.0.0
Hadoop Flags: Incompatible change

> "hadoop jar my.jar" should probably prioritize my.jar in the classpath by 
> default
> -
>
> Key: HADOOP-12728
> URL: https://issues.apache.org/jira/browse/HADOOP-12728
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: scripts
>Affects Versions: 2.7.1
>Reporter: Ovidiu Gheorghioiu
>Assignee: Allen Wittenauer
>Priority: Minor
>
> Found this surprising behavior when testing a dev version of a jar that was 
> already in the hadoop classpath:"hadoop jar ./my.jar" used the system 
> my.jar, which was an old version that did not contain my bug fix.  Since 
> "hadoop jar" is the rough equivalent of running an executable, it should use 
> the  version passed on the command line.
> Even worse than my case (which took a while to figure out with log messages) 
> is when I'd be testing that the new version works the same as the old in some 
> use case. Then I'd think it did, even though the new version was completely 
> broken.
> Allen mentioned verbally that there are some tricky aspects to this, but to 
> open this issue for tracking / brainstorming.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12651) Replace dev-support with wrappers to Yetus

2016-01-21 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111225#comment-15111225
 ] 

Allen Wittenauer commented on HADOOP-12651:
---

Assuming the +1 is still good, I'll go ahead and commit, update the wiki pages, 
and drop a note to common-dev so people aren't taken by surprise.

Thanks for the review!

> Replace dev-support with wrappers to Yetus
> --
>
> Key: HADOOP-12651
> URL: https://issues.apache.org/jira/browse/HADOOP-12651
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>Assignee: Allen Wittenauer
> Attachments: HADOOP-12651.00.patch, HADOOP-12651.01.patch, 
> HADOOP-12651.02.patch, HADOOP-12651.03.patch, HADOOP-12651.04.patch
>
>
> Now that Yetus has had a release, we should rip out the components that make 
> it up from dev-support and replace them with wrappers.  The wrappers should:
> * default to a sane version
> * allow for version overrides via an env var
> * download into patchprocess
> * execute with the given parameters
> Marking this as an incompatible change, since we should also remove the 
> filename extensions and move these into a bin directory for better 
> maintenance towards the future.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12651) Replace dev-support with wrappers to Yetus

2016-01-21 Thread Allen Wittenauer (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12651?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Allen Wittenauer updated HADOOP-12651:
--
   Resolution: Fixed
Fix Version/s: 3.0.0
   Status: Resolved  (was: Patch Available)

> Replace dev-support with wrappers to Yetus
> --
>
> Key: HADOOP-12651
> URL: https://issues.apache.org/jira/browse/HADOOP-12651
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>Assignee: Allen Wittenauer
> Fix For: 3.0.0
>
> Attachments: HADOOP-12651.00.patch, HADOOP-12651.01.patch, 
> HADOOP-12651.02.patch, HADOOP-12651.03.patch, HADOOP-12651.04.patch
>
>
> Now that Yetus has had a release, we should rip out the components that make 
> it up from dev-support and replace them with wrappers.  The wrappers should:
> * default to a sane version
> * allow for version overrides via an env var
> * download into patchprocess
> * execute with the given parameters
> Marking this as an incompatible change, since we should also remove the 
> filename extensions and move these into a bin directory for better 
> maintenance towards the future.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12659) Incorrect usage of config parameters in token manager of KMS

2016-01-21 Thread Xiaoyu Yao (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12659?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Xiaoyu Yao updated HADOOP-12659:

   Resolution: Fixed
 Hadoop Flags: Reviewed
Fix Version/s: 2.8.0
   Status: Resolved  (was: Patch Available)

Thanks [~liuml07] for the contribution and all for the reviews. I've committed 
the patch to trunk, branch-2 and branch-2.8.

> Incorrect usage of config parameters in token manager of KMS
> 
>
> Key: HADOOP-12659
> URL: https://issues.apache.org/jira/browse/HADOOP-12659
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.7.1, 2.6.2
>Reporter: Tianyin Xu
>Assignee: Mingliang Liu
> Fix For: 2.8.0
>
> Attachments: HADOOP-12659.000.patch
>
>
> Hi, the usage of the following configs of Key Management Server (KMS) are 
> problematic: 
> {{hadoop.kms.authentication.delegation-token.renew-interval.sec}}
> {{hadoop.kms.authentication.delegation-token.removal-scan-interval.sec}}
> The name indicates that the units are {{sec}}, and the online doc shows that 
> the default values are {{86400}} and {{3600}}, respectively.
> https://hadoop.apache.org/docs/stable/hadoop-kms/index.html
> which is also defined in
> {code:title=DelegationTokenManager.java|borderStyle=solid}
>  55   public static final String RENEW_INTERVAL = PREFIX + 
> "renew-interval.sec";
>  56   public static final long RENEW_INTERVAL_DEFAULT = 24 * 60 * 60;
>  ...
>  58   public static final String REMOVAL_SCAN_INTERVAL = PREFIX +
>  59   "removal-scan-interval.sec";
>  60   public static final long REMOVAL_SCAN_INTERVAL_DEFAULT = 60 * 60;
> {code}
> However, in {{DelegationTokenManager.java}} and 
> {{ZKDelegationTokenSecretManager.java}}, these two parameters are used 
> incorrectly.
> 1. *{{DelegationTokenManager.java}}*
> {code}
>  70   conf.getLong(RENEW_INTERVAL, RENEW_INTERVAL_DEFAULT) * 1000,
>  71   conf.getLong(REMOVAL_SCAN_INTERVAL, 
>  72   REMOVAL_SCAN_INTERVAL_DEFAULT * 1000));
> {code}
> Apparently, at Line 72, {{REMOVAL_SCAN_INTERVAL}} should be used in the same 
> way as {{RENEW_INTERVAL}}, like
> {code}
> 72c72
> <   REMOVAL_SCAN_INTERVAL_DEFAULT * 1000));
> ---
> >   REMOVAL_SCAN_INTERVAL_DEFAULT) * 1000);
> {code}
> Currently, the unit of 
> {{hadoop.kms.authentication.delegation-token.removal-scan-interval.sec}} is 
> not {{sec}} but {{millisec}}.
> 2. *{{ZKDelegationTokenSecretManager.java}}*
> {code}
> 142 conf.getLong(DelegationTokenManager.RENEW_INTERVAL,
> 143 DelegationTokenManager.RENEW_INTERVAL_DEFAULT * 1000),
> 144 conf.getLong(DelegationTokenManager.REMOVAL_SCAN_INTERVAL,
> 145 DelegationTokenManager.REMOVAL_SCAN_INTERVAL_DEFAULT) * 1000);
> {code}
>  The situation is the opposite in this class that 
> {{hadoop.kms.authentication.delegation-token.renew-interval.sec}} is wrong 
> but the other is correct...
> A patch should be like
> {code}
> 143c143
> < DelegationTokenManager.RENEW_INTERVAL_DEFAULT * 1000),
> ---
> > DelegationTokenManager.RENEW_INTERVAL_DEFAULT) * 1000,
> {code}
> Thanks!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12469) distcp should not ignore the ignoreFailures option

2016-01-21 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12469?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu updated HADOOP-12469:
---
Attachment: HADOOP-12469.002.patch

The v2 patch rebases from {{trunk}} branch to resolve trivial conflicts.

> distcp should not ignore the ignoreFailures option
> --
>
> Key: HADOOP-12469
> URL: https://issues.apache.org/jira/browse/HADOOP-12469
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: tools/distcp
>Affects Versions: 2.7.1
>Reporter: Gera Shegalov
>Assignee: Mingliang Liu
>Priority: Critical
> Attachments: HADOOP-12469.000.patch, HADOOP-12469.001.patch, 
> HADOOP-12469.002.patch
>
>
> {{RetriableFileCopyCommand.CopyReadException}} is double-wrapped via
> # via {{RetriableCommand::execute}}
> # via {{CopyMapper#copyFileWithRetry}}
> before {{CopyMapper::handleFailure}} tests 
> {code}
> if (ignoreFailures && exception.getCause() instanceof
> RetriableFileCopyCommand.CopyReadException
> {code}
> which is always false.
> Orthogonally, ignoring failures should be mutually exclusive with the atomic 
> option otherwise an incomplete dir is eligible for commit defeating the 
> purpose.
>  



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12659) Incorrect usage of config parameters in token manager of KMS

2016-01-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111546#comment-15111546
 ] 

Hudson commented on HADOOP-12659:
-

FAILURE: Integrated in Hadoop-trunk-Commit #9157 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9157/])
HADOOP-12659. Incorrect usage of config parameters in token manager of (xyao: 
rev f3427d3766d7101d0d1c37d6281918551d221ebe)
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/ZKDelegationTokenSecretManager.java
* 
hadoop-common-project/hadoop-common/src/main/java/org/apache/hadoop/security/token/delegation/web/DelegationTokenManager.java
* hadoop-common-project/hadoop-common/CHANGES.txt


> Incorrect usage of config parameters in token manager of KMS
> 
>
> Key: HADOOP-12659
> URL: https://issues.apache.org/jira/browse/HADOOP-12659
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.7.1, 2.6.2
>Reporter: Tianyin Xu
>Assignee: Mingliang Liu
> Fix For: 2.8.0
>
> Attachments: HADOOP-12659.000.patch
>
>
> Hi, the usage of the following configs of Key Management Server (KMS) are 
> problematic: 
> {{hadoop.kms.authentication.delegation-token.renew-interval.sec}}
> {{hadoop.kms.authentication.delegation-token.removal-scan-interval.sec}}
> The name indicates that the units are {{sec}}, and the online doc shows that 
> the default values are {{86400}} and {{3600}}, respectively.
> https://hadoop.apache.org/docs/stable/hadoop-kms/index.html
> which is also defined in
> {code:title=DelegationTokenManager.java|borderStyle=solid}
>  55   public static final String RENEW_INTERVAL = PREFIX + 
> "renew-interval.sec";
>  56   public static final long RENEW_INTERVAL_DEFAULT = 24 * 60 * 60;
>  ...
>  58   public static final String REMOVAL_SCAN_INTERVAL = PREFIX +
>  59   "removal-scan-interval.sec";
>  60   public static final long REMOVAL_SCAN_INTERVAL_DEFAULT = 60 * 60;
> {code}
> However, in {{DelegationTokenManager.java}} and 
> {{ZKDelegationTokenSecretManager.java}}, these two parameters are used 
> incorrectly.
> 1. *{{DelegationTokenManager.java}}*
> {code}
>  70   conf.getLong(RENEW_INTERVAL, RENEW_INTERVAL_DEFAULT) * 1000,
>  71   conf.getLong(REMOVAL_SCAN_INTERVAL, 
>  72   REMOVAL_SCAN_INTERVAL_DEFAULT * 1000));
> {code}
> Apparently, at Line 72, {{REMOVAL_SCAN_INTERVAL}} should be used in the same 
> way as {{RENEW_INTERVAL}}, like
> {code}
> 72c72
> <   REMOVAL_SCAN_INTERVAL_DEFAULT * 1000));
> ---
> >   REMOVAL_SCAN_INTERVAL_DEFAULT) * 1000);
> {code}
> Currently, the unit of 
> {{hadoop.kms.authentication.delegation-token.removal-scan-interval.sec}} is 
> not {{sec}} but {{millisec}}.
> 2. *{{ZKDelegationTokenSecretManager.java}}*
> {code}
> 142 conf.getLong(DelegationTokenManager.RENEW_INTERVAL,
> 143 DelegationTokenManager.RENEW_INTERVAL_DEFAULT * 1000),
> 144 conf.getLong(DelegationTokenManager.REMOVAL_SCAN_INTERVAL,
> 145 DelegationTokenManager.REMOVAL_SCAN_INTERVAL_DEFAULT) * 1000);
> {code}
>  The situation is the opposite in this class that 
> {{hadoop.kms.authentication.delegation-token.renew-interval.sec}} is wrong 
> but the other is correct...
> A patch should be like
> {code}
> 143c143
> < DelegationTokenManager.RENEW_INTERVAL_DEFAULT * 1000),
> ---
> > DelegationTokenManager.RENEW_INTERVAL_DEFAULT) * 1000,
> {code}
> Thanks!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12651) Replace dev-support with wrappers to Yetus

2016-01-21 Thread Hudson (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12651?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111372#comment-15111372
 ] 

Hudson commented on HADOOP-12651:
-

FAILURE: Integrated in Hadoop-trunk-Commit #9156 (See 
[https://builds.apache.org/job/Hadoop-trunk-Commit/9156/])
HADOOP-12651. Replace dev-support with wrappers to Yetus (aw) (aw: rev 
8cecad2d564d89d0f56e98de80a6f9b0bcb39d7a)
* dev-support/smart-apply-patch.sh
* hadoop-common-project/hadoop-common/pom.xml
* BUILDING.txt
* dev-support/test-patch.d/shellcheck.sh
* dev-support/test-patch.sh
* dev-support/test-patch.d/checkstyle.sh
* dev-support/test-patch.d/whitespace.sh
* dev-support/shelldocs.py
* dev-support/releasedocmaker.py
HADOOP-12651. Replace dev-support with wrappers to Yetus (aw) (aw: rev 
2a867355dfce880bb82257508862ec26bc7f16b9)
* dev-support/README.md
* dev-support/bin/test-patch
* dev-support/bin/releasedocmaker
* dev-support/bin/shelldocs
* dev-support/bin/yetus-wrapper
* dev-support/bin/smart-apply-patch


> Replace dev-support with wrappers to Yetus
> --
>
> Key: HADOOP-12651
> URL: https://issues.apache.org/jira/browse/HADOOP-12651
> Project: Hadoop Common
>  Issue Type: New Feature
>  Components: scripts
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>Assignee: Allen Wittenauer
> Fix For: 3.0.0
>
> Attachments: HADOOP-12651.00.patch, HADOOP-12651.01.patch, 
> HADOOP-12651.02.patch, HADOOP-12651.03.patch, HADOOP-12651.04.patch
>
>
> Now that Yetus has had a release, we should rip out the components that make 
> it up from dev-support and replace them with wrappers.  The wrappers should:
> * default to a sane version
> * allow for version overrides via an env var
> * download into patchprocess
> * execute with the given parameters
> Marking this as an incompatible change, since we should also remove the 
> filename extensions and move these into a bin directory for better 
> maintenance towards the future.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12659) Incorrect usage of config parameters in token manager of KMS

2016-01-21 Thread Mingliang Liu (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12659?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111383#comment-15111383
 ] 

Mingliang Liu commented on HADOOP-12659:


Thank you [~asuresh] and [~xyao] for the review.

> Incorrect usage of config parameters in token manager of KMS
> 
>
> Key: HADOOP-12659
> URL: https://issues.apache.org/jira/browse/HADOOP-12659
> Project: Hadoop Common
>  Issue Type: Bug
>  Components: security
>Affects Versions: 2.7.1, 2.6.2
>Reporter: Tianyin Xu
>Assignee: Mingliang Liu
> Fix For: 2.8.0
>
> Attachments: HADOOP-12659.000.patch
>
>
> Hi, the usage of the following configs of Key Management Server (KMS) are 
> problematic: 
> {{hadoop.kms.authentication.delegation-token.renew-interval.sec}}
> {{hadoop.kms.authentication.delegation-token.removal-scan-interval.sec}}
> The name indicates that the units are {{sec}}, and the online doc shows that 
> the default values are {{86400}} and {{3600}}, respectively.
> https://hadoop.apache.org/docs/stable/hadoop-kms/index.html
> which is also defined in
> {code:title=DelegationTokenManager.java|borderStyle=solid}
>  55   public static final String RENEW_INTERVAL = PREFIX + 
> "renew-interval.sec";
>  56   public static final long RENEW_INTERVAL_DEFAULT = 24 * 60 * 60;
>  ...
>  58   public static final String REMOVAL_SCAN_INTERVAL = PREFIX +
>  59   "removal-scan-interval.sec";
>  60   public static final long REMOVAL_SCAN_INTERVAL_DEFAULT = 60 * 60;
> {code}
> However, in {{DelegationTokenManager.java}} and 
> {{ZKDelegationTokenSecretManager.java}}, these two parameters are used 
> incorrectly.
> 1. *{{DelegationTokenManager.java}}*
> {code}
>  70   conf.getLong(RENEW_INTERVAL, RENEW_INTERVAL_DEFAULT) * 1000,
>  71   conf.getLong(REMOVAL_SCAN_INTERVAL, 
>  72   REMOVAL_SCAN_INTERVAL_DEFAULT * 1000));
> {code}
> Apparently, at Line 72, {{REMOVAL_SCAN_INTERVAL}} should be used in the same 
> way as {{RENEW_INTERVAL}}, like
> {code}
> 72c72
> <   REMOVAL_SCAN_INTERVAL_DEFAULT * 1000));
> ---
> >   REMOVAL_SCAN_INTERVAL_DEFAULT) * 1000);
> {code}
> Currently, the unit of 
> {{hadoop.kms.authentication.delegation-token.removal-scan-interval.sec}} is 
> not {{sec}} but {{millisec}}.
> 2. *{{ZKDelegationTokenSecretManager.java}}*
> {code}
> 142 conf.getLong(DelegationTokenManager.RENEW_INTERVAL,
> 143 DelegationTokenManager.RENEW_INTERVAL_DEFAULT * 1000),
> 144 conf.getLong(DelegationTokenManager.REMOVAL_SCAN_INTERVAL,
> 145 DelegationTokenManager.REMOVAL_SCAN_INTERVAL_DEFAULT) * 1000);
> {code}
>  The situation is the opposite in this class that 
> {{hadoop.kms.authentication.delegation-token.renew-interval.sec}} is wrong 
> but the other is correct...
> A patch should be like
> {code}
> 143c143
> < DelegationTokenManager.RENEW_INTERVAL_DEFAULT * 1000),
> ---
> > DelegationTokenManager.RENEW_INTERVAL_DEFAULT) * 1000,
> {code}
> Thanks!



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12469) distcp should not ignore the ignoreFailures option

2016-01-21 Thread Hadoop QA (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12469?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111677#comment-15111677
 ] 

Hadoop QA commented on HADOOP-12469:


| (/) *{color:green}+1 overall{color}* |
\\
\\
|| Vote || Subsystem || Runtime || Comment ||
| {color:blue}0{color} | {color:blue} reexec {color} | {color:blue} 0m 0s 
{color} | {color:blue} Docker mode activated. {color} |
| {color:green}+1{color} | {color:green} @author {color} | {color:green} 0m 0s 
{color} | {color:green} The patch does not contain any @author tags. {color} |
| {color:green}+1{color} | {color:green} test4tests {color} | {color:green} 0m 
0s {color} | {color:green} The patch appears to include 1 new or modified test 
files. {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 9m 
1s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 27s 
{color} | {color:green} trunk passed with JDK v1.8.0_66 {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 17s 
{color} | {color:green} trunk passed with JDK v1.7.0_91 {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
10s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 23s 
{color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
13s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 
30s {color} | {color:green} trunk passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 13s 
{color} | {color:green} trunk passed with JDK v1.8.0_66 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 15s 
{color} | {color:green} trunk passed with JDK v1.7.0_91 {color} |
| {color:green}+1{color} | {color:green} mvninstall {color} | {color:green} 0m 
18s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 11s 
{color} | {color:green} the patch passed with JDK v1.8.0_66 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 11s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} compile {color} | {color:green} 0m 14s 
{color} | {color:green} the patch passed with JDK v1.7.0_91 {color} |
| {color:green}+1{color} | {color:green} javac {color} | {color:green} 0m 14s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} checkstyle {color} | {color:green} 0m 
9s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvnsite {color} | {color:green} 0m 20s 
{color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} mvneclipse {color} | {color:green} 0m 
10s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} whitespace {color} | {color:green} 0m 
0s {color} | {color:green} Patch has no whitespace issues. {color} |
| {color:green}+1{color} | {color:green} findbugs {color} | {color:green} 0m 
37s {color} | {color:green} the patch passed {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 10s 
{color} | {color:green} the patch passed with JDK v1.8.0_66 {color} |
| {color:green}+1{color} | {color:green} javadoc {color} | {color:green} 0m 12s 
{color} | {color:green} the patch passed with JDK v1.7.0_91 {color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 7m 19s 
{color} | {color:green} hadoop-distcp in the patch passed with JDK v1.8.0_66. 
{color} |
| {color:green}+1{color} | {color:green} unit {color} | {color:green} 7m 31s 
{color} | {color:green} hadoop-distcp in the patch passed with JDK v1.7.0_91. 
{color} |
| {color:green}+1{color} | {color:green} asflicense {color} | {color:green} 0m 
27s {color} | {color:green} Patch does not generate ASF License warnings. 
{color} |
| {color:black}{color} | {color:black} {color} | {color:black} 30m 32s {color} 
| {color:black} {color} |
\\
\\
|| Subsystem || Report/Notes ||
| Docker |  Image:yetus/hadoop:0ca8df7 |
| JIRA Patch URL | 
https://issues.apache.org/jira/secure/attachment/12783685/HADOOP-12469.002.patch
 |
| JIRA Issue | HADOOP-12469 |
| Optional Tests |  asflicense  compile  javac  javadoc  mvninstall  mvnsite  
unit  findbugs  checkstyle  |
| uname | Linux 7c4655b03f00 3.13.0-36-lowlatency #63-Ubuntu SMP PREEMPT Wed 
Sep 3 21:56:12 UTC 2014 x86_64 x86_64 x86_64 GNU/Linux |
| Build tool | maven |
| Personality | /testptch/hadoop/patchprocess/precommit/personality/provided.sh 
|
| git revision | trunk / f3427d3 |
| Default Java | 1.7.0_91 |
| Multi-JDK versions |  

[jira] [Commented] (HADOOP-12699) TestKMS#testKMSProvider intermittently fails during 'test rollover draining'

2016-01-21 Thread Xiao Chen (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12699?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111838#comment-15111838
 ] 

Xiao Chen commented on HADOOP-12699:


Zhe talked to me offline today regarding the sequence to reproduce the failure. 
Below is the order of the events happened to result in an old key being 
retrieved. Note that the race is only in the 
{{EagerKeyGeneratorKeyProviderCryptoExtension}} in {{KMSWebApp}}, so steps are 
only focused on the {{ValueQueue}} here. 
# Caller {{generateEncryptedKey}}, resulting in the server side KMSProvider's 
{{ValueQueue}} fired up an async filling task - a {{NamedRunnable}} is added 
into the {{UniqueKeyBlockingQueue}}(queue).
# The {{NamedRunnable}} in step 1 is executed, filling in the local 
{{retEdeks}}, but not yet {{keyQueue.addAll}}. Code 
[here|https://github.com/apache/hadoop/blob/trunk/hadoop-common-project/hadoop-kms/src/main/java/org/apache/hadoop/crypto/key/kms/server/EagerKeyGeneratorKeyProviderCryptoExtension.java#L83].
# Caller {{rollNewVersion}}, {{ValueQueue}} is drained. That is, the queue of 
the being rolled name in {{LoadingCache}}(keyQueues) is cleared.
# The {{NamedRunnable}} continues to run, adding the old keys from local 
variable into {{keyQueue}}.
# Caller side {{generateEncryptedKey}}, poll from the keyQueue, get old key.

As a side note, simply not using the local variable wouldn't help, because even 
if

> TestKMS#testKMSProvider intermittently fails during 'test rollover draining'
> 
>
> Key: HADOOP-12699
> URL: https://issues.apache.org/jira/browse/HADOOP-12699
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Xiao Chen
>Assignee: Xiao Chen
> Attachments: HADOOP-12699.01.patch, HADOOP-12699.02.patch, 
> HADOOP-12699.03.patch, HADOOP-12699.04.patch, HADOOP-12699.06.patch, 
> HADOOP-12699.07.patch, HADOOP-12699.repro.2, HADOOP-12699.repro.patch
>
>
> I've seen several failures of testKMSProvider, all failed in the following 
> snippet:
> {code}
> // test rollover draining
> KeyProviderCryptoExtension kpce = KeyProviderCryptoExtension.
> createKeyProviderCryptoExtension(kp);
> .
> EncryptedKeyVersion ekv1 = kpce.generateEncryptedKey("k6");
> kpce.rollNewVersion("k6");
> EncryptedKeyVersion ekv2 = kpce.generateEncryptedKey("k6");
> Assert.assertNotEquals(ekv1.getEncryptionKeyVersionName(),
> ekv2.getEncryptionKeyVersionName());
> {code}
> with error message
> {quote}Values should be different. Actual: k6@0{quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12588) Fix intermittent test failure of TestGangliaMetrics

2016-01-21 Thread Masatake Iwasaki (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12588?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Masatake Iwasaki updated HADOOP-12588:
--
Status: Open  (was: Patch Available)

> Fix intermittent test failure of TestGangliaMetrics
> ---
>
> Key: HADOOP-12588
> URL: https://issues.apache.org/jira/browse/HADOOP-12588
> Project: Hadoop Common
>  Issue Type: Bug
>Reporter: Tsuyoshi Ozawa
>Assignee: Masatake Iwasaki
> Attachments: HADOOP-12588.001.patch, HADOOP-12588.addendum.02.patch, 
> HADOOP-12588.addendum.patch
>
>
> Jenkins found this test failure on HADOOP-11149.
> {quote}
> Tests run: 2, Failures: 1, Errors: 0, Skipped: 0, Time elapsed: 0.773 sec <<< 
> FAILURE! - in org.apache.hadoop.metrics2.impl.TestGangliaMetrics
> testGangliaMetrics2(org.apache.hadoop.metrics2.impl.TestGangliaMetrics)  Time 
> elapsed: 0.39 sec  <<< FAILURE!
> java.lang.AssertionError: Missing metrics: test.s1rec.Xxx
>   at org.junit.Assert.fail(Assert.java:88)
>   at org.junit.Assert.assertTrue(Assert.java:41)
>   at 
> org.apache.hadoop.metrics2.impl.TestGangliaMetrics.checkMetrics(TestGangliaMetrics.java:159)
>   at 
> org.apache.hadoop.metrics2.impl.TestGangliaMetrics.testGangliaMetrics2(TestGangliaMetrics.java:137)
> {quote}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12668) Modify HDFS embeded jetty server logic in HttpServer2.java to exclude weak Ciphers through ssl-server.conf

2016-01-21 Thread John Zhuge (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12668?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=15111702#comment-15111702
 ] 

John Zhuge commented on HADOOP-12668:
-

+1 for the patch

> Modify HDFS embeded jetty server logic in HttpServer2.java to exclude weak 
> Ciphers through ssl-server.conf
> --
>
> Key: HADOOP-12668
> URL: https://issues.apache.org/jira/browse/HADOOP-12668
> Project: Hadoop Common
>  Issue Type: Improvement
>  Components: security
>Affects Versions: 2.7.1
>Reporter: Vijay Singh
>Assignee: Vijay Singh
>Priority: Critical
>  Labels: common, ha, hadoop, hdfs, security
> Attachments: Hadoop-12668.006.patch
>
>   Original Estimate: 24h
>  Remaining Estimate: 24h
>
> Currently Embeded jetty Server used across all hadoop services is configured 
> through ssl-server.xml file from their respective configuration section. 
> However, the SSL/TLS protocol being used for this jetty servers can be 
> downgraded to weak cipher suites. This code changes aims to add following 
> functionality:
> 1) Add logic in hadoop common (HttpServer2.java and associated interfaces) to 
> spawn jetty servers with ability to exclude weak cipher suites. I propose we 
> make this though ssl-server.xml and hence each service can choose to disable 
> specific ciphers.
> 2) Modify DFSUtil.java used by HDFS code to supply new parameter 
> ssl.server.exclude.cipher.list for hadoop-common code, so it can exclude the 
> ciphers supplied through this key.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Updated] (HADOOP-12725) RPC call benchmark and optimization in different SASL QOP level

2016-01-21 Thread Kai Zheng (JIRA)

 [ 
https://issues.apache.org/jira/browse/HADOOP-12725?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Kai Zheng updated HADOOP-12725:
---
Summary: RPC call benchmark and optimization in different SASL QOP level  
(was: RPC call benchmark in different SASL QOP level)

> RPC call benchmark and optimization in different SASL QOP level
> ---
>
> Key: HADOOP-12725
> URL: https://issues.apache.org/jira/browse/HADOOP-12725
> Project: Hadoop Common
>  Issue Type: Improvement
>Reporter: Kai Zheng
>Assignee: Kai Zheng
>
> This would implement a benchmark tool to measure and compare the performance 
> of Hadoop IPC/RPC call when security is enabled and different SASL 
> QOP(Quality of Protection) is enforced. Given the data collected by this 
> benchmark, it would then be able to know if any performance concern when 
> considering to enforce privacy, integration, or authenticy protection level, 
> and do optimization accordingly.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Comment Edited] (HADOOP-12563) Updated utility to create/modify token files

2016-01-21 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=1574#comment-1574
 ] 

Allen Wittenauer edited comment on HADOOP-12563 at 1/21/16 7:44 PM:


I'm still playing around a bit, but some feedback:

* Non-existent file handling should throw a message as well as give a failure 
exit code:
{code}
$ rm foo
$ hadoop dtutil get foo
$ echo $?
0
{code}

* fetchdt needs to default to use the java-based serialization for backward 
compatibility. Simple script that shows it:

{code}
#!/bin/bash

URL=hdfs://$(hostname):9000

hdfs fetchdt dt-fetchdt
hadoop dtutil get ${URL} dt-util-default
hadoop dtutil get ${URL} -format java dt-util-java
hadoop dtutil get ${URL} -format protobuf dt-util-proto
ls -l dt-*
{code}



was (Author: aw):
I'm still playing around a bit, but some feedback:

* Non-existent file handling should throw a message as well as give a failure 
exit code:
{code}
$ rm foo
$ hadoop dtutil get foo
$ echo $?
0
{code}

* fetchdt needs to default to use the java-based serialization for backward 
compatibility.

{code}
#!/bin/bash

URL=hdfs://$(hostname):9000

hdfs fetchdt dt-fetchdt
hadoop dtutil get ${URL} dt-util-default
hadoop dtutil get ${URL} -format java dt-util-java
hadoop dtutil get ${URL} -format protobuf dt-util-proto
ls -l dt-*
{code}


> Updated utility to create/modify token files
> 
>
> Key: HADOOP-12563
> URL: https://issues.apache.org/jira/browse/HADOOP-12563
> Project: Hadoop Common
>  Issue Type: New Feature
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>Assignee: Matthew Paduano
> Attachments: HADOOP-12563.01.patch, HADOOP-12563.02.patch, 
> HADOOP-12563.03.patch, HADOOP-12563.04.patch, HADOOP-12563.05.patch, 
> HADOOP-12563.06.patch, example_dtutil_commands_and_output.txt, 
> generalized_token_case.pdf
>
>
> hdfs fetchdt is missing some critical features and is geared almost 
> exclusively towards HDFS operations.  Additionally, the token files that are 
> created use Java serializations which are hard/impossible to deal with in 
> other languages. It should be replaced with a better utility in common that 
> can read/write protobuf-based token files, has enough flexibility to be used 
> with other services, and offers key functionality such as append and rename. 
> The old version file format should still be supported for backward 
> compatibility, but will be effectively deprecated.
> A follow-on JIRA will deprecrate fetchdt.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Commented] (HADOOP-12563) Updated utility to create/modify token files

2016-01-21 Thread Allen Wittenauer (JIRA)

[ 
https://issues.apache.org/jira/browse/HADOOP-12563?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel=1574#comment-1574
 ] 

Allen Wittenauer commented on HADOOP-12563:
---

I'm still playing around a bit, but some feedback:

* Non-existent file handling should throw a message as well as give a failure 
exit code:
{code}
$ rm foo
$ hadoop dtutil get foo
$ echo $?
0
{code}

* fetchdt needs to default to use the java-based serialization for backward 
compatibility.

{code}
#!/bin/bash

URL=hdfs://$(hostname):9000

hdfs fetchdt dt-fetchdt
hadoop dtutil get ${URL} dt-util-default
hadoop dtutil get ${URL} -format java dt-util-java
hadoop dtutil get ${URL} -format protobuf dt-util-proto
ls -l dt-*
{code}


> Updated utility to create/modify token files
> 
>
> Key: HADOOP-12563
> URL: https://issues.apache.org/jira/browse/HADOOP-12563
> Project: Hadoop Common
>  Issue Type: New Feature
>Affects Versions: 3.0.0
>Reporter: Allen Wittenauer
>Assignee: Matthew Paduano
> Attachments: HADOOP-12563.01.patch, HADOOP-12563.02.patch, 
> HADOOP-12563.03.patch, HADOOP-12563.04.patch, HADOOP-12563.05.patch, 
> HADOOP-12563.06.patch, example_dtutil_commands_and_output.txt, 
> generalized_token_case.pdf
>
>
> hdfs fetchdt is missing some critical features and is geared almost 
> exclusively towards HDFS operations.  Additionally, the token files that are 
> created use Java serializations which are hard/impossible to deal with in 
> other languages. It should be replaced with a better utility in common that 
> can read/write protobuf-based token files, has enough flexibility to be used 
> with other services, and offers key functionality such as append and rename. 
> The old version file format should still be supported for backward 
> compatibility, but will be effectively deprecated.
> A follow-on JIRA will deprecrate fetchdt.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HADOOP-12728) "hadoop jar my.jar" should probably prioritize my.jar in the classpath by default

2016-01-21 Thread Ovidiu Gheorghioiu (JIRA)
Ovidiu Gheorghioiu created HADOOP-12728:
---

 Summary: "hadoop jar my.jar" should probably prioritize my.jar in 
the classpath by default
 Key: HADOOP-12728
 URL: https://issues.apache.org/jira/browse/HADOOP-12728
 Project: Hadoop Common
  Issue Type: Improvement
  Components: scripts
Affects Versions: 2.7.1
Reporter: Ovidiu Gheorghioiu
Assignee: Allen Wittenauer
Priority: Minor


Found this surprising behavior when testing a dev version of a jar that was 
already in the hadoop classpath:"hadoop jar ./my.jar" used the system 
my.jar, which was an old version that did not contain my bug fix.  Since 
"hadoop jar" is the rough equivalent of running an executable, it should use 
the  version passed on the command line.

Even worse than my case (which took a while to figure out with log messages) is 
when I'd be testing that the new version works the same as the old in some use 
case. Then I'd think it did, even though the new version was completely broken.

Allen mentioned verbally that there are some tricky aspects to this, but to 
open this issue for tracking / brainstorming.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)