[jira] [Created] (HDFS-13827) Distcp job is failing due to Invalid arguments

2018-08-14 Thread Sudhansu Bhuyan (JIRA)
Sudhansu Bhuyan created HDFS-13827:
--

 Summary: Distcp job is failing due to Invalid arguments
 Key: HDFS-13827
 URL: https://issues.apache.org/jira/browse/HDFS-13827
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: distcp
Affects Versions: 2.7.1
Reporter: Sudhansu Bhuyan


18/08/14 18:35:59 ERROR tools.DistCp: Invalid arguments:
java.lang.IllegalArgumentException: Neither source file listing nor source 
paths present
 at 
org.apache.hadoop.tools.OptionsParser.parseSourceAndTargetPaths(OptionsParser.java:348)
 at org.apache.hadoop.tools.OptionsParser.parse(OptionsParser.java:89)
 at org.apache.hadoop.tools.DistCp.run(DistCp.java:117)
 at org.apache.hadoop.util.ToolRunner.run(ToolRunner.java:76)
 at org.apache.hadoop.tools.DistCp.main(DistCp.java:462)
Invalid arguments: Neither source file listing nor source paths present
usage: distcp OPTIONS [source_path...] 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-13826) Add a hidden configuration for NameNode to generate fake block locations

2018-08-14 Thread Todd Lipcon (JIRA)
Todd Lipcon created HDFS-13826:
--

 Summary: Add a hidden configuration for NameNode to generate fake 
block locations
 Key: HDFS-13826
 URL: https://issues.apache.org/jira/browse/HDFS-13826
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: namenode
Reporter: Todd Lipcon
Assignee: Todd Lipcon


In doing testing and benchmarking of the NameNode and dependent systems, it's 
often useful to be able to use an fsimage provided by some production system in 
a controlled environment without actually having access to any of the data. For 
example, while doing some recent work on Apache Impala I was trying to optimize 
the transmission and storage of block locations and tokens and measure the 
results based on metadata from a production user. In order to achieve this, it 
would be useful for the NN to expose a developer-only (undocumented) 
configuration to generate fake block locations and return them to callers. The 
"fake" locations should be randomly distributed across a fixed set of fake 
datanodes.



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: Hadoop 3.2 Release Plan proposal

2018-08-14 Thread Sunil G
Hi All,

Thanks for the feedbacks. Inline with earlier communication dated 17th July
2018, I would like to provide some updates.

We are approaching previously proposed feature freeze date (Aug 21, about 7
days from today).
If any features in branch which are targeted to 3.2.0, please reply to this
email thread.
Steve has mentioned about the s3 features which will come close to Code
Freeze Date (Aug 31st).

*Here's an updated 3.2.0 feature status:*

1. Merged & Completed features:

- (Wangda) YARN-8561: Hadoop Submarine project for DeepLearning workloads
Initial cut.
- (Uma) HDFS-10285: HDFS Storage Policy Satisfier

2. Features close to finish:

- (Naga/Sunil) YARN-3409: Node Attributes support in YARN. Major patches
are all in, only one last
patch is in review state.
- (Sunil) YARN-7494: Multi Node scheduling support in Capacity Scheduler.
Close to commit.
- (Chandni/Eric) YARN-7512: Support service upgrade via YARN Service API
and CLI. 2 patches are pending
which will be closed by Feature freeze date.
- (Rohith) YARN-5742: Serve aggregated logs of historical apps from ATSv2.
Patch in progress.
- (Virajit) HDFS-12615: Router-based HDFS federation. Improvement works.
- (Steve) S3Guard Phase III, S3a phase V, Support Windows Azure Storage. In
progress.

3. Tentative features:

- (Haibo Chen) YARN-1011: Resource overcommitment. Looks challenging to be
done before Aug 2018.
- (Eric) YARN-7129: Application Catalog for YARN applications. Challenging
as more discussions are on-going.

*Summary of 3.2.0 issues status:*

39 Blocker and Critical issues [1] are open, I am checking with owners to
get status on each of them to get in by Code Freeze date.

[1] project in (YARN, HADOOP, MAPREDUCE, HDFS) AND priority in (Blocker,
Critical) AND resolution = Unresolved AND "Target Version/s" = 3.2.0 ORDER
BY priority DESC

Thanks,
Sunil

On Fri, Jul 20, 2018 at 8:03 AM Sunil G  wrote:

> Thanks Subru for the thoughts.
> One of the main reason for a major release is to push out critical
> features with a faster cadence to the users. If we are pulling more and
> more different types of features to a minor release, that branch will
> become more destabilized and it may be tough to say that 3.1.2 is stable
> that 3.1.1 for eg. We always tend to improve and stabilize features in
> subsequent minor release.
> For few companies, it makes sense to push out these new features faster to
> make a reach to the users. Adding to the point to the backporting issues, I
> agree that its a pain and we can workaround that with some git scripts. If
> we can make such scripts available to committers, backport will be
> seem-less across branches and we can achieve the faster release cadence
> also.
>
> Thoughts?
>
> - Sunil
>
>
> On Fri, Jul 20, 2018 at 3:37 AM Subru Krishnan  wrote:
>
>> Thanks Sunil for volunteering to lead the release effort. I am generally
>> supportive of a release but -1 on a 3.2 (prefer a 3.1.x) as feel we
>> already
>> have too many branches to be maintained. I already see many commits are in
>> different branches with no apparent rationale, for e.g: 3.1 has commits
>> which are absent in 3.0 etc.
>>
>> Additionally AFAIK 3.x has not been deployed in any major production
>> setting so the cost of adding features should be minimal.
>>
>> Thoughts?
>>
>> -Subru
>>
>> On Thu, Jul 19, 2018 at 12:31 AM, Sunil G  wrote:
>>
>> > Thanks Steve, Aaron, Wangda for sharing thoughts.
>> >
>> > Yes, important changes and features are much needed, hence we will be
>> > keeping the door open for them as possible. Also considering few more
>> > offline requests from other folks, I think extending the timeframe by
>> > couple of weeks makes sense (including a second RC buffer) and this
>> should
>> > ideally help us to ship this by September itself.
>> >
>> > Revised dates (I will be updating same in Roadmap wiki as well)
>> >
>> > - Feature freeze date : all features to merge by August 21, 2018.
>> >
>> > - Code freeze date : blockers/critical only, no improvements and non
>> > blocker/critical
>> >
>> > bug-fixes  August 31, 2018.
>> >
>> > - Release date: September 15, 2018
>> >
>> > Thank Eric and Zian, I think Wangda has already answered your questions.
>> >
>> > Thanks
>> > Sunil
>> >
>> >
>> > On Thu, Jul 19, 2018 at 12:13 PM Wangda Tan 
>> wrote:
>> >
>> > > Thanks Sunil for volunteering to be RM of 3.2 release, +1 for that.
>> > >
>> > > To concerns from Steve,
>> > >
>> > > It is a good idea to keep the door open to get important changes /
>> > > features in before cutoff. I would prefer to keep the proposed release
>> > date
>> > > to make sure things can happen earlier instead of last minute and we
>> all
>> > > know that releases are always get delayed :). I'm also fine if we want
>> > get
>> > > another several weeks time.
>> > >
>> > > Regarding of 3.3 release, I would suggest doing that before
>> thanksgiving.
>> > > Do you think is it good or too early / late?
>> > >
>> > > Eric,
>> > >
>> > > The YARN-8220 will be 

[jira] [Created] (HDFS-13825) HDFS Uses very outdated okhttp library

2018-08-14 Thread Ben Parker (JIRA)
Ben Parker created HDFS-13825:
-

 Summary: HDFS Uses very outdated okhttp library
 Key: HDFS-13825
 URL: https://issues.apache.org/jira/browse/HDFS-13825
 Project: Hadoop HDFS
  Issue Type: Improvement
Affects Versions: 3.0.3
Reporter: Ben Parker


HDFS Client uses okHttp library version 2.7.4 which is two years out of date.

[https://mvnrepository.com/artifact/com.squareup.okhttp/okhttp]

The updates for this library have been moved to a new package here:

[https://mvnrepository.com/artifact/com.squareup.okhttp3/okhttp]

 

This causes dependancy management problems for services that use HDFS.

For example trying to use okHttp in code that runs on Amazon EMR gives you 
Method not found errors due to the new version being kicked out in favour of 
the one used by HDFS.

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Apache Hadoop qbt Report: trunk+JDK8 on Linux/x86

2018-08-14 Thread Apache Jenkins Server
For more details, see 
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/

[Aug 13, 2018 7:17:52 AM] (msingh) HDDS-308. SCM should identify a container 
with pending deletes using
[Aug 13, 2018 8:52:55 AM] (sunilg) YARN-8561. [Submarine] Initial 
implementation: Training job submission
[Aug 13, 2018 9:32:56 AM] (drankye) HDFS-13668. FSPermissionChecker may throws 
AIOOE when check inode
[Aug 13, 2018 10:57:45 AM] (ewan.higgs) HADOOP-15645. 
ITestS3GuardToolLocal.testDiffCommand fails if bucket has
[Aug 13, 2018 5:08:58 PM] (aengineer) HDDS-346. ozoneShell show the new volume 
info after updateVolume command
[Aug 13, 2018 5:40:31 PM] (xiao) HADOOP-15638. KMS Accept Queue Size default 
changed from 500 to 128 in
[Aug 13, 2018 6:35:19 PM] (arp) HDFS-13823. NameNode UI : "Utilities -> Browse 
the file system -> open a
[Aug 13, 2018 7:47:49 PM] (xyao) HDDS-324. Use pipeline name as Ratis groupID 
to allow datanode to report
[Aug 13, 2018 8:50:00 PM] (eyang) YARN-7417. Remove duplicated code from 
IndexedFileAggregatedLogsBlock   
[Aug 13, 2018 11:12:37 PM] (weichiu) HDFS-13813. Exit NameNode if dangling 
child inode is detected when
[Aug 14, 2018 12:36:13 AM] (weichiu) HDFS-13738. fsck -list-corruptfileblocks 
has infinite loop if user is




-1 overall


The following subsystems voted -1:
asflicense findbugs pathlen unit xml


The following subsystems voted -1 but
were configured to be filtered/ignored:
cc checkstyle javac javadoc pylint shellcheck shelldocs whitespace


The following subsystems are considered long running:
(runtime bigger than 1h  0m  0s)
unit


Specific tests:

FindBugs :

   
module:hadoop-yarn-project/hadoop-yarn/hadoop-yarn-applications/hadoop-yarn-submarine
 
   Unread field:FSBasedSubmarineStorageImpl.java:[line 39] 
   Found reliance on default encoding in 
org.apache.hadoop.yarn.submarine.runtimes.yarnservice.YarnServiceJobSubmitter.generateCommandLaunchScript(RunJobParameters,
 TaskType, Component):in 
org.apache.hadoop.yarn.submarine.runtimes.yarnservice.YarnServiceJobSubmitter.generateCommandLaunchScript(RunJobParameters,
 TaskType, Component): new java.io.FileWriter(File) At 
YarnServiceJobSubmitter.java:[line 192] 
   
org.apache.hadoop.yarn.submarine.runtimes.yarnservice.YarnServiceJobSubmitter.generateCommandLaunchScript(RunJobParameters,
 TaskType, Component) may fail to clean up java.io.Writer on checked exception 
Obligation to clean up resource created at YarnServiceJobSubmitter.java:to 
clean up java.io.Writer on checked exception Obligation to clean up resource 
created at YarnServiceJobSubmitter.java:[line 192] is not discharged 
   
org.apache.hadoop.yarn.submarine.runtimes.yarnservice.YarnServiceUtils.getComponentArrayJson(String,
 int, String) concatenates strings using + in a loop At 
YarnServiceUtils.java:using + in a loop At YarnServiceUtils.java:[line 72] 

Failed CTEST tests :

   test_test_libhdfs_threaded_hdfs_static 
   test_libhdfs_threaded_hdfspp_test_shim_static 

Failed junit tests :

   hadoop.hdfs.web.TestWebHdfsTimeouts 
   hadoop.yarn.server.resourcemanager.applicationsmanager.TestAMRestart 
   hadoop.yarn.applications.distributedshell.TestDistributedShell 
   hadoop.mapred.TestMRTimelineEventHandling 
   hadoop.tools.util.TestProducerConsumer 
   hadoop.yarn.sls.TestSLSRunner 
  

   cc:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/diff-compile-cc-root.txt
  [4.0K]

   javac:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/diff-compile-javac-root.txt
  [328K]

   checkstyle:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/diff-checkstyle-root.txt
  [4.0K]

   pathlen:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/pathlen.txt
  [12K]

   pylint:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/diff-patch-pylint.txt
  [24K]

   shellcheck:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/diff-patch-shellcheck.txt
  [20K]

   shelldocs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/diff-patch-shelldocs.txt
  [16K]

   whitespace:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/whitespace-eol.txt
  [9.4M]
   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/whitespace-tabs.txt
  [1.1M]

   xml:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/xml.txt
  [4.0K]

   findbugs:

   
https://builds.apache.org/job/hadoop-qbt-trunk-java8-linux-x86/868/artifact/out/branch-findbugs-hadoop-yarn-project_hadoop-yarn_hadoop-yarn-applications_hadoop-yarn-submarine-warnings.html
  [12K]
   

[jira] [Resolved] (HDFS-13031) To detect fsimage corruption on the spot

2018-08-14 Thread Adam Antal (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDFS-13031?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Adam Antal resolved HDFS-13031.
---
Resolution: Won't Fix

Created HDFS-13818 as follow-up Jira for the issue.

> To detect fsimage corruption on the spot
> 
>
> Key: HDFS-13031
> URL: https://issues.apache.org/jira/browse/HDFS-13031
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: hdfs
> Environment:  
>Reporter: Yongjun Zhang
>Assignee: Adam Antal
>Priority: Major
>
> Since we fixed HDFS-9406, there are new cases reported from the field that 
> similar fsimage corruption happens. We need good fsimage + editlogs to replay 
> to reproduce the corruption. However, usually when the corruption is detected 
> (at later NN restart), the good fsimage is already deleted.
> We need to have a way to detect fsimage corruption on the spot. Currently 
> what I think we could do is:
>  # after SNN creates a new fsimage, it spawn a new modified NN process (NN 
> with some new command line args) to just load the fsimage and do nothing 
> else. 
>  # If the process failed, the currently running SNN will do either a) backup 
> the fsimage + editlogs or b) no longer do checkpointing. And it need to 
> somehow raise a flag to user that the fsimage is corrupt.
> In step 2, if we do a, we need to introduce new NN->JN API to backup 
> editlogs; if we do b, it changes SNN's behavior, and kind of not compatible. 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDDS-295) TestCloseContainerByPipeline is failing because of timeout

2018-08-14 Thread Shashikant Banerjee (JIRA)


 [ 
https://issues.apache.org/jira/browse/HDDS-295?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Shashikant Banerjee resolved HDDS-295.
--
   Resolution: Duplicate
Fix Version/s: 0.2.1

> TestCloseContainerByPipeline is failing because of timeout
> --
>
> Key: HDDS-295
> URL: https://issues.apache.org/jira/browse/HDDS-295
> Project: Hadoop Distributed Data Store
>  Issue Type: Bug
>  Components: SCM
>Reporter: Mukul Kumar Singh
>Assignee: Shashikant Banerjee
>Priority: Major
> Fix For: 0.2.1
>
>
> The test is failing because the test is timing out waiting for the container 
> to be closed.
> The details are logged at 
> https://builds.apache.org/job/PreCommit-HDDS-Build/627/testReport/



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-13824) Number of Dead nodes is not showing in the Overview and Subclusters pages. However Live nodes are relecting properly

2018-08-14 Thread Soumyapn (JIRA)
Soumyapn created HDFS-13824:
---

 Summary: Number of Dead nodes is not showing in the Overview and 
Subclusters pages. However Live nodes are relecting properly
 Key: HDFS-13824
 URL: https://issues.apache.org/jira/browse/HDFS-13824
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: federation
Affects Versions: 3.1.0
Reporter: Soumyapn
 Attachments: image-2018-08-14-11-47-05-025.png

Scenario:

Suppose we have 2 nameservices with 3 Datanodes each. 

If we make 2 DN's down, then the Datanodes page, Live nodes field in Overview 
and Live in Subclusters page is reflected to 4.

But the Deadnodes field in Overview and Subclusters page is showing as 0. It is 
not reflected.

!image-2018-08-14-11-47-05-025.png!

 



--
This message was sent by Atlassian JIRA
(v7.6.3#76005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org