Re: [VOTE] Hadoop 3.2.x EOL

2023-12-05 Thread Mingliang Liu
+1

On Tue, Dec 5, 2023 at 8:09 PM Xiaoqiao He  wrote:

> Dear Hadoop devs,
>
> Given the feedback from the discussion thread [1], I'd like to start
> an official thread for the community to vote on release line 3.2 EOL.
>
> It will include,
> a. An official announcement informs no further regular Hadoop 3.2.x
> releases.
> b. Issues which target 3.2.5 will not be fixed.
>
> This vote will run for 7 days and conclude by Dec 13, 2023.
>
> I’ll start with my +1.
>
> Best Regards,
> - He Xiaoqiao
>
> [1] https://lists.apache.org/thread/bbf546c6jz0og3xcl9l3qfjo93b65szr
>


Re: [ANNOUNCE] New Hadoop Committer - Simbarashe Dzinamarira

2023-10-03 Thread Mingliang Liu
Congratulations!

On Tue, Oct 3, 2023 at 8:55 AM Ayush Saxena  wrote:

> Congratulations!!!
>
> -Ayush
>
> > On 03-Oct-2023, at 5:42 AM, Erik Krogen  wrote:
> >
> > Congratulations Simba! Thanks for the great work you've done on making
> HDFS
> > more scalable!
> >
> >> On Mon, Oct 2, 2023 at 4:31 PM Iñigo Goiri  wrote:
> >>
> >> I am pleased to announce that Simbarashe Dzinamarira has been elected
> as a
> >> committer on the Apache Hadoop project.
> >> We appreciate all of Simbarashe's work, and look forward to his
> continued
> >> contributions.
> >>
> >> Congratulations and welcome !
> >>
> >> Best Regards,
> >> Inigo Goiri
> >> (On behalf of the Apache Hadoop PMC)
> >>
>
> -
> To unsubscribe, e-mail: general-unsubscr...@hadoop.apache.org
> For additional commands, e-mail: general-h...@hadoop.apache.org
>
>


Re: [DISCUSS] JIRA Public Signup Disabled

2022-11-24 Thread Mingliang Liu
Thanks Ayush for taking care of this. I think option 1 sounds good.

> On Nov 22, 2022, at 2:51 PM, Ayush Saxena  wrote:
> 
> Hi Folks,
> Just driving the attention towards the recent change from Infra, which
> disables new people from creating a Jira account, in order to prevent spams
> around JIRA.
> So, the new Jira account creation request needs to be routed via the PMC of
> the project.
> So, we have 2 options, which I can think of:
> 1. Update the contribution guidelines to route such requests to private@
> 2. Create a dedicated ML for it. A couple of projects which I know did that.
> 
> The Infra page: https://infra.apache.org/jira-guidelines.html
> 
> Let me know what folks think, if nothing, I will go with the 1st option and
> update the contribution guidelines mostly by next week or a week after that.
> 
> -Ayuhs


-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-15938) Fix java doc in FSEditLog

2021-04-01 Thread Mingliang Liu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15938?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-15938.
--
Fix Version/s: 3.4.0
 Hadoop Flags: Reviewed
   Resolution: Fixed

> Fix java doc in FSEditLog
> -
>
> Key: HDFS-15938
> URL: https://issues.apache.org/jira/browse/HDFS-15938
> Project: Hadoop HDFS
>  Issue Type: Wish
>Reporter: tomscut
>Assignee: tomscut
>Priority: Minor
>  Labels: pull-request-available
> Fix For: 3.4.0
>
>  Time Spent: 50m
>  Remaining Estimate: 0h
>
> Fix java doc in 
> org.apache.hadoop.hdfs.server.namenode.FSEditLog#logAddCacheDirectiveInfo.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-15624) Fix the SetQuotaByStorageTypeOp problem after updating hadoop

2021-02-02 Thread Mingliang Liu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15624?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-15624.
--
Fix Version/s: 3.4.0
 Hadoop Flags: Reviewed
   Resolution: Fixed

Committed to trunk branch. Thank you [~huangtianhua] and su xu for your 
contribution. Thank you [~ayushtkn] and [~vinayakumarb] for your helpful review.

>  Fix the SetQuotaByStorageTypeOp problem after updating hadoop 
> ---
>
> Key: HDFS-15624
> URL: https://issues.apache.org/jira/browse/HDFS-15624
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: hdfs
>Affects Versions: 3.4.0
>Reporter: YaYun Wang
>Assignee: huangtianhua
>Priority: Major
>  Labels: pull-request-available, release-blocker
> Fix For: 3.4.0
>
>  Time Spent: 9.5h
>  Remaining Estimate: 0h
>
> HDFS-15025 adds a new storage Type NVDIMM, changes the ordinal() of the enum 
> of StorageType. And, setting the quota by storageType depends on the 
> ordinal(), therefore, it may cause the setting of quota to be invalid after 
> upgrade.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[ANNOUNCE] Mukund Thakur is a new Apache Hadoop Committer

2021-02-02 Thread Mingliang Liu
Hi all,

It's my pleasure to announce that Mukund Thakur has been elected as a
committer on the Apache Hadoop project recognizing his continued
contributions to the project. He has accepted the invitation.


His involvement in the hadoop object store modules, hadoop common and
distcp tool is a great addition to the Hadoop project. His reviews are
comprehensive with high standards to "approve a pull request". The granting
of committership to him will enable better productivity.


Please join me in congratulating him. Welcome aboard, Mukund!



Mingliang Liu

(on behalf of the Apache Hadoop PMC)


Re: [VOTE] Moving Ozone to a separated Apache project

2020-09-29 Thread Mingliang Liu
+1

On Tue, Sep 29, 2020 at 3:44 PM Wangda Tan  wrote:

> +1,
>
> Thanks,
> Wangda Tan
>
> On Tue, Sep 29, 2020 at 10:10 AM Aravindan Vijayan
>  wrote:
>
> > +1, thank you Marton.
> >
> > On Tue, Sep 29, 2020 at 9:17 AM Bharat Viswanadham 
> > wrote:
> >
> > > +1
> > > Thank You @Elek, Marton  for driving this.
> > >
> > >
> > > Thanks,
> > > Bharat
> > >
> > >
> > > On Mon, Sep 28, 2020 at 10:54 AM Vivek Ratnavel <
> > vivekratna...@apache.org>
> > > wrote:
> > >
> > > > +1 for moving Ozone to a separated Top-Level Apache Project.
> > > >
> > > > Thanks,
> > > > Vivek Subramanian
> > > >
> > > > On Mon, Sep 28, 2020 at 8:30 AM Hanisha Koneru
> > > > 
> > > > wrote:
> > > >
> > > > > +1
> > > > >
> > > > > Thanks,
> > > > > Hanisha
> > > > >
> > > > > > On Sep 27, 2020, at 11:48 PM, Akira Ajisaka  >
> > > > wrote:
> > > > > >
> > > > > > +1
> > > > > >
> > > > > > Thanks,
> > > > > > Akira
> > > > > >
> > > > > > On Fri, Sep 25, 2020 at 3:00 PM Elek, Marton  > >  > > > > e...@apache.org>> wrote:
> > > > > >>
> > > > > >> Hi all,
> > > > > >>
> > > > > >> Thank you for all the feedback and requests,
> > > > > >>
> > > > > >> As we discussed in the previous thread(s) [1], Ozone is proposed
> > to
> > > > be a
> > > > > >> separated Apache Top Level Project (TLP)
> > > > > >>
> > > > > >> The proposal with all the details, motivation and history is
> here:
> > > > > >>
> > > > > >>
> > > > >
> > > >
> > >
> >
> https://cwiki.apache.org/confluence/display/HADOOP/Ozone+Hadoop+subproject+to+Apache+TLP+proposal
> > > > > >>
> > > > > >> This voting runs for 7 days and will be concluded at 2nd of
> > October,
> > > > 6AM
> > > > > >> GMT.
> > > > > >>
> > > > > >> Thanks,
> > > > > >> Marton Elek
> > > > > >>
> > > > > >> [1]:
> > > > > >>
> > > > >
> > > >
> > >
> >
> https://lists.apache.org/thread.html/rc6c79463330b3e993e24a564c6817aca1d290f186a1206c43ff0436a%40%3Chdfs-dev.hadoop.apache.org%3E
> > > > > >>
> > > > > >>
> > > -
> > > > > >> To unsubscribe, e-mail: yarn-dev-unsubscr...@hadoop.apache.org
> > > >  > > > > yarn-dev-unsubscr...@hadoop.apache.org>
> > > > > >> For additional commands, e-mail:
> yarn-dev-h...@hadoop.apache.org
> > > > > 
> > > > > >>
> > > > > >
> > > > > >
> > -
> > > > > > To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
> > > > > 
> > > > > > For additional commands, e-mail:
> common-dev-h...@hadoop.apache.org
> > > > > 
> > > > >
> > > >
> > >
> >
> >
> > --
> > Thanks & Regards,
> > Aravindan
> >
>


[jira] [Created] (HDFS-15595) stSnapshotCommands.testMaxSnapshotLimit fails in trunk

2020-09-23 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-15595:


 Summary: stSnapshotCommands.testMaxSnapshotLimit fails in trunk
 Key: HDFS-15595
 URL: https://issues.apache.org/jira/browse/HDFS-15595
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs, snapshots, test
Reporter: Mingliang Liu


See 
[this|https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2326/1/testReport/org.apache.hadoop.hdfs/TestSnapshotCommands/testMaxSnapshotLimit/]
 for a sample error.

Sample error stack:
{quote}
Error Message
The real output is: createSnapshot: Failed to create snapshot: there are 
already 4 snapshot(s) and the per directory snapshot limit is 3
.
 It should contain: Failed to add snapshot: there are already 3 snapshot(s) and 
the max snapshot limit is 3
Stacktrace
java.lang.AssertionError: 
The real output is: createSnapshot: Failed to create snapshot: there are 
already 4 snapshot(s) and the per directory snapshot limit is 3
.
 It should contain: Failed to add snapshot: there are already 3 snapshot(s) and 
the max snapshot limit is 3
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.apache.hadoop.hdfs.DFSTestUtil.toolRun(DFSTestUtil.java:1934)
at org.apache.hadoop.hdfs.DFSTestUtil.FsShellRun(DFSTestUtil.java:1942)
at 
org.apache.hadoop.hdfs.TestSnapshotCommands.testMaxSnapshotLimit(TestSnapshotCommands.java:130)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at 
org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
{quote}

I can also reproduce this locally.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [VOTE] Release Apache Hadoop 2.10.1 (RC0)

2020-09-21 Thread Mingliang Liu
+1 (binding)

1. Download binary and check signature / checksum successfully
2. Create a 3 node cluster in Docker containers and start the HDFS/YARN
services
3. Verify the running version
4. Run simple HDFS/YARN client/admin commands and verify the output
5. Run example programs wordcount and grep
6. Check NN/DN/RM status and service logs

Thanks,

On Mon, Sep 21, 2020 at 12:34 AM Vinayakumar B 
wrote:

> Apologies for delayed voting.
>
> +1 (Binding)
>
> 1. Verified the signature. Signatures are good.
> 2. Verified the sha512 checksum
>(I tried to use the sha512sum tool to verify. But ".sha512"
> files doesn't have the expected format. I guess these files are regenerated
> using "gpg --print-md", not the one used by createrelease scripts,
> 'sha512sum --tag'. It would be easy to verify using tool if the .sha512
> have the proper format)
> 3. src tar have the "patchprocess" directory, may be need to backport the
> fix in trunk to branch-2.10 to avoid this in future.
> 4. Deployed 3 node docker cluster and ran basic jobs. All-Ok.
>
> -Vinay
>
>
> On Mon, Sep 21, 2020 at 5:24 AM Wei-Chiu Chuang 
> wrote:
>
> > +1 (binding)
> >
> > I did a security scan for the 2.10.1 RC0 and it looks fine to me.
> >
> > Checked recent critical/blocker HDFS issues that are not in 2.10.1. It
> > looks mostly fine. Most of them are Hadoop 3.x features (EC, ... etc) but
> > there is one worth attention:
> >
> >
> >1. HDFS-14674  [SBN
> >read] Got an unexpected txid when tail editlog.
> >2.
> >   1. But looking at the jira, it doesn't apply to 2.x so I think we
> are
> >   good there.
> >   2.
> >   3.
> >   4. I wanted to do an API compat check but didn't finish it yet. If
> >   someone can do it quickly that would be great. (Does anyone know
> > of a cloud
> >   service that we can quickly do a Java API compat check?)
> >
> > Cheers,
> > Wei-Chiu
> >
> > On Sun, Sep 20, 2020 at 9:25 AM Sunil Govindan 
> wrote:
> >
> > > +1 (binding)
> > >
> > > - verified checksum and sign. Shows as a Good signature from "Masatake
> > > Iwasaki (CODE SIGNING KEY) "
> > > - built from source
> > > - ran basic MR job and looks good
> > > - UI also seems fine
> > >
> > > Thanks,
> > > Sunil
> > >
> > > On Sun, Sep 20, 2020 at 11:38 AM Masatake Iwasaki <
> > > iwasak...@oss.nttdata.co.jp> wrote:
> > >
> > > > The RC0 got 2 binding +1's and 2 non-binging +1's [1].
> > > >
> > > > Based on the discussion about release vote [2],
> > > > bylaws[3] defines the periods in minimum terms.
> > > > We can extend it if there is not enough activity.
> > > >
> > > > I would like to extend the period to 7 days,
> > > > until Monday September 21 at 10:00 am PDT.
> > > >
> > > > I will appreciate additional votes.
> > > >
> > > > Thanks,
> > > > Masatake Iwasaki
> > > >
> > > > [1]
> > > >
> > >
> >
> https://lists.apache.org/thread.html/r16a7f36315a0673c7d522c41065e7ef9c9ee15c76ffcb5db80931002%40%3Ccommon-dev.hadoop.apache.org%3E
> > > > [2]
> > > >
> > >
> >
> https://lists.apache.org/thread.html/e392b902273ee0c14ba34d72c44630e05f54cb3976109af510592ea2%401403330080%40%3Ccommon-dev.hadoop.apache.org%3E
> > > > [3] https://hadoop.apache.org/bylaws.html
> > > >
> > > > On 2020/09/15 2:59, Masatake Iwasaki wrote:
> > > > > Hi folks,
> > > > >
> > > > > This is the first release candidate for the second release of
> Apache
> > > > Hadoop 2.10.
> > > > > It contains 218 fixes/improvements since 2.10.0 [1].
> > > > >
> > > > > The RC0 artifacts are at:
> > > > > http://home.apache.org/~iwasakims/hadoop-2.10.1-RC0/
> > > > >
> > > > > RC tag is release-2.10.1-RC0:
> > > > > https://github.com/apache/hadoop/tree/release-2.10.1-RC0
> > > > >
> > > > > The maven artifacts are hosted here:
> > > > >
> > >
> https://repository.apache.org/content/repositories/orgapachehadoop-1279/
> > > > >
> > > > > My public key is available here:
> > > > > https://dist.apache.org/repos/dist/release/hadoop/common/KEYS
> > > > >
> > > > > The vote will run for 5 days, until Saturday, September 19 at 10:00
> > am
> > > > PDT.
> > > > >
> > > > > [1]
> > > >
> > >
> >
> https://issues.apache.org/jira/issues/?jql=project%20in%20(HDFS%2C%20YARN%2C%20HADOOP%2C%20MAPREDUCE)%20AND%20resolution%20%3D%20Fixed%20AND%20fixVersion%20%3D%202.10.1
> > > > >
> > > > > Thanks,
> > > > > Masatake Iwasaki
> > > > >
> > > > >
> -
> > > > > To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
> > > > > For additional commands, e-mail: common-dev-h...@hadoop.apache.org
> > > > >
> > > >
> > > > -
> > > > To unsubscribe, e-mail: mapreduce-dev-unsubscr...@hadoop.apache.org
> > > > For additional commands, e-mail:
> mapreduce-dev-h...@hadoop.apache.org
> > > >
> > > >
> > >
> >
>


Re: Unstable Unit Tests in Trunk

2020-09-09 Thread Mingliang Liu
Thanks Eric.

I also see some intermittent test failures in recent builds.

Maybe we can mark those failing tests as @Flaky, and/or add the RetryRule()
to them. Thoughts?

On Tue, Sep 1, 2020 at 10:48 AM Eric Badger
 wrote:

> While putting up patches for HADOOP-17169
>  I noticed that the
> unit tests in trunk, specifically in HDFS, are incredibly unstable. Every
> time I put up a new patch, 4-8 unit tests failed with failures that were
> completely unrelated to the patch. I'm pretty confident in that since the
> patch is simply changing variable names. I also ran the unit tests locally
> and they would pass (or fail intermittently).
>
> Is there an effort to stabilize the unit tests? I don't know if these are
> bugs or if they're bad tests. But in either case, it's bad for the
> stability of the project.
>
> Eric
>


-- 
L


Re: How to manually trigger a PreCommit build for a github PR?

2020-09-09 Thread Mingliang Liu
> you mean to say rerun jenkins, that means it ran once, you want to run it
again without any code changes?
Yes.

Thank you Ayush!

On Tue, Sep 8, 2020 at 6:20 PM Ayush Saxena  wrote:

> Hi Mingliang,
> If by running Pre-Commit without any code change, you mean to say rerun
> jenkins, that means it ran once, you want to run it again without any code
> changes?
> If so,
> You can go to that PR build and click on replay. For example :
> Once logged in, if you click on the below link :
> https://ci-hadoop.apache.org/job/hadoop-multibranch/job/PR-2281/2/
>
> You can see an option 'replay' on the left, Clicking on it will rerun the
> build.
>
> Secondly, If the last build isn't available, then I think creating an empty
> commit is the only way as Dinesh also suggested.
>
> -Ayush
>
> On Wed, 9 Sep 2020 at 03:43, Mingliang Liu  wrote:
>
> > Thanks Dinesh. This is very helpful.
> >
> > I will add this to the wiki page
> > <https://cwiki.apache.org/confluence/display/HADOOP/GitHub+Integration>
> if
> > this is the suggested way of doing it.
> >
> >
> >
> > On Tue, Sep 8, 2020 at 2:54 PM Dinesh Chitlangia <
> dchitlan...@cloudera.com
> > >
> > wrote:
> >
> > > You could try doing an empty commit.
> > >
> > > git commit --allow-empty -m 'trigger new CI check' && git push
> > >
> > >
> > > Thanks,
> > > Dinesh
> > >
> > >
> > >
> > > On Tue, Sep 8, 2020 at 5:39 PM Mingliang Liu 
> wrote:
> > >
> > >> Hi,
> > >>
> > >> To trigger a PreCommit build without code change, I can make the JIRA
> > >> status "Patch Available" and provide the JIRA number to "Build With
> > >> Parameters" link
> > >> <
> > >>
> >
> https://ci-hadoop.apache.org/view/Hadoop/job/PreCommit-HADOOP-Build/build?delay=0sec
> > >> >
> > >> .
> > >>
> > >> Not sure how to do that for a PR without a real commit to the PR
> branch?
> > >>
> > >> Thanks,
> > >>
> > >
> >
>


Re: How to manually trigger a PreCommit build for a github PR?

2020-09-08 Thread Mingliang Liu
Thanks Dinesh. This is very helpful.

I will add this to the wiki page
<https://cwiki.apache.org/confluence/display/HADOOP/GitHub+Integration> if
this is the suggested way of doing it.



On Tue, Sep 8, 2020 at 2:54 PM Dinesh Chitlangia 
wrote:

> You could try doing an empty commit.
>
> git commit --allow-empty -m 'trigger new CI check' && git push
>
>
> Thanks,
> Dinesh
>
>
>
> On Tue, Sep 8, 2020 at 5:39 PM Mingliang Liu  wrote:
>
>> Hi,
>>
>> To trigger a PreCommit build without code change, I can make the JIRA
>> status "Patch Available" and provide the JIRA number to "Build With
>> Parameters" link
>> <
>> https://ci-hadoop.apache.org/view/Hadoop/job/PreCommit-HADOOP-Build/build?delay=0sec
>> >
>> .
>>
>> Not sure how to do that for a PR without a real commit to the PR branch?
>>
>> Thanks,
>>
>


How to manually trigger a PreCommit build for a github PR?

2020-09-08 Thread Mingliang Liu
Hi,

To trigger a PreCommit build without code change, I can make the JIRA
status "Patch Available" and provide the JIRA number to "Build With
Parameters" link

.

Not sure how to do that for a PR without a real commit to the PR branch?

Thanks,


Re: [DISCUSS] Hadoop 2.10.1 release

2020-08-31 Thread Mingliang Liu
I can see how I can help, but I can not take the RM role this time.

Thanks,

On Mon, Aug 31, 2020 at 12:15 PM Wei-Chiu Chuang
 wrote:

> Hello,
>
> I see that Masatake graciously agreed to volunteer with the Hadoop 2.10.1
> release work in the 2.9 branch EOL discussion thread
> https://s.apache.org/hadoop2.9eold
>
> Anyone else likes to contribute also?
>
> Thanks
>


-- 
L


Re: [VOTE] End of Life Hadoop 2.9

2020-08-31 Thread Mingliang Liu
+1 (binding)

p.s. We should still maintain 2.10 for Hadoop 2 users and encourage them to
upgrade to Hadoop 3.

On Mon, Aug 31, 2020 at 12:10 PM Wei-Chiu Chuang  wrote:

> Dear fellow Hadoop developers,
>
> Given the overwhelming feedback from the discussion thread
> https://s.apache.org/hadoop2.9eold, I'd like to start an official vote
> thread for the community to vote and start the 2.9 EOL process.
>
> What this entails:
>
> (1) an official announcement that no further regular Hadoop 2.9.x releases
> will be made after 2.9.2 (which was GA on 11/19/2019)
> (2) resolve JIRAs that specifically target 2.9.3 as won't fix.
>
>
> This vote will run for 7 days and will conclude by September 7th, 12:00pm
> pacific time.
> Committers are eligible to cast binding votes. Non-committers are welcomed
> to cast non-binding votes.
>
> Here is my vote, +1
>


-- 
L


[jira] [Resolved] (HDFS-15546) Remove duplicate initializeGenericKeys method when creating DFSZKFailoverController

2020-08-28 Thread Mingliang Liu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15546?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-15546.
--
Resolution: Not A Problem

Thanks for good discussion [~jianghuazhu] and [~chaosun]

Since this is not committed to any branch, I have cleared the "Fix Version/s" 
and marked this Jira as "Not a Problem". A Jira which is "Resolved" with fix 
versions only apply to code change.

> Remove duplicate initializeGenericKeys method when creating 
> DFSZKFailoverController
> ---
>
> Key: HDFS-15546
> URL: https://issues.apache.org/jira/browse/HDFS-15546
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: jianghua zhu
>Assignee: jianghua zhu
>Priority: Major
>  Labels: pull-request-available
> Attachments: HDFS-15546.001.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> When calling the DFSZKFailoverController#create() method, call 
> NameNode#initializeGenericKeys() twice.
> First call:
> DFSZKFailoverController#create() {
> ...
> NameNode.initializeGenericKeys(localNNConf, nsId, nnId);
> ...
> }
> The second call:
> NNHAServiceTarget construction method:
> NNHAServiceTarget() {
> ...
> NameNode.initializeGenericKeys(targetConf, nsId, nnId);
> ...
> }
> In fact, the parameters passed in the two calls are the same. Here you can 
> remove the first call without affecting the program itself.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Reopened] (HDFS-15546) Remove duplicate initializeGenericKeys method when creating DFSZKFailoverController

2020-08-28 Thread Mingliang Liu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15546?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu reopened HDFS-15546:
--

> Remove duplicate initializeGenericKeys method when creating 
> DFSZKFailoverController
> ---
>
> Key: HDFS-15546
> URL: https://issues.apache.org/jira/browse/HDFS-15546
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: jianghua zhu
>Assignee: jianghua zhu
>Priority: Major
>  Labels: pull-request-available
> Attachments: HDFS-15546.001.patch
>
>  Time Spent: 1.5h
>  Remaining Estimate: 0h
>
> When calling the DFSZKFailoverController#create() method, call 
> NameNode#initializeGenericKeys() twice.
> First call:
> DFSZKFailoverController#create() {
> ...
> NameNode.initializeGenericKeys(localNNConf, nsId, nnId);
> ...
> }
> The second call:
> NNHAServiceTarget construction method:
> NNHAServiceTarget() {
> ...
> NameNode.initializeGenericKeys(targetConf, nsId, nnId);
> ...
> }
> In fact, the parameters passed in the two calls are the same. Here you can 
> remove the first call without affecting the program itself.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [DISCUSS] fate of branch-2.9

2020-08-27 Thread Mingliang Liu
> are there any Hadoop branch-2 releases planned, ever? If so I'll need to
backport my s3a directory compatibility patch to whatever is still live.

The branch-2 is gone. I think you mean branch-2.10, Steve.

Many HBase users are still using Hadoop 2, so I hope Hadoop 2.10.x should
still be released at least every 12 months. If there is no volunteer for
2.10.1 RM, I can see how I can help.

Thanks,

On Thu, Aug 27, 2020 at 8:55 AM John Zhuge  wrote:

> +1
>
> On Thu, Aug 27, 2020 at 6:01 AM Ayush Saxena  wrote:
>
> > +1
> >
> > -Ayush
> >
> > > On 27-Aug-2020, at 6:24 PM, Steve Loughran  >
> > wrote:
> > >
> > > 
> > >
> > > +1
> > >
> > > are there any Hadoop branch-2 releases planned, ever? If so I'll need
> to
> > backport my s3a directory compatibility patch to whatever is still live.
> > >
> > >
> > >> On Thu, 27 Aug 2020 at 06:55, Wei-Chiu Chuang 
> > wrote:
> > >> Bump up this thread after 6 months.
> > >>
> > >> Is anyone still interested in the 2.9 release line? Or are we good to
> > start
> > >> the EOL process? The 2.9.2 was released in Nov 2018.
> > >>
> > >> I'd really like to see the community to converge to fewer release
> lines
> > and
> > >> make more frequent releases in each line.
> > >>
> > >> Thanks,
> > >> Weichiu
> > >>
> > >>
> > >> On Fri, Mar 6, 2020 at 5:47 PM Wei-Chiu Chuang 
> > wrote:
> > >>
> > >> > I think that's a great suggestion.
> > >> > Currently, we make 1 minor release per year, and within each minor
> > release
> > >> > we bring up 1 thousand to 2 thousand commits in it compared with the
> > >> > previous one.
> > >> > I can totally understand it is a big bite for users to swallow.
> > Having a
> > >> > more frequent release cycle, plus LTS and non-LTS releases should
> > help with
> > >> > this. (Of course we will need to make the release preparation much
> > easier,
> > >> > which is currently a pain)
> > >> >
> > >> > I am happy to discuss the release model further in the dev ML. LTS
> > v.s.
> > >> > non-LTS is one suggestion.
> > >> >
> > >> > Another similar issue: In the past Hadoop strived to
> > >> > maintain compatibility. However, this is no longer sustainable as
> > more CVEs
> > >> > coming from our dependencies: netty, jetty, jackson ... etc.
> > >> > In many cases, updating the dependencies brings breaking changes.
> More
> > >> > recently, especially in Hadoop 3.x, I started to make the effort to
> > update
> > >> > dependencies much more frequently. How do users feel about this
> > change?
> > >> >
> > >> > On Thu, Mar 5, 2020 at 7:58 AM Igor Dvorzhak  >
> > >> > wrote:
> > >> >
> > >> >> Maybe Hadoop will benefit from adopting a similar release and
> support
> > >> >> strategy as Java? I.e. designate some releases as LTS and support
> > them for
> > >> >> 2 (?) years (it seems that 2.7.x branch was de-facto LTS), other
> > non-LTS
> > >> >> releases will be supported for 6 months (or until next release).
> This
> > >> >> should allow to reduce maintenance cost of non-LTS release and
> > provide
> > >> >> conservative users desired stability by allowing them to wait for
> > new LTS
> > >> >> release and upgrading to it.
> > >> >>
> > >> >> On Thu, Mar 5, 2020 at 1:26 AM Rupert Mazzucco <
> > rupert.mazzu...@gmail.com>
> > >> >> wrote:
> > >> >>
> > >> >>> After recently jumping from 2.7.7 to 2.10 without issue myself, I
> > vote
> > >> >>> for keeping only the 2.10 line.
> > >> >>> It would seem all other 2.x branches can upgrade to a 2.10.x
> easily
> > if
> > >> >>> they feel like upgrading at all,
> > >> >>> unlike a jump to 3.x, which may require more planning.
> > >> >>>
> > >> >>> I also vote for having only one main 3.x branch. Why are there
> > 3.1.x and
> > >> >>> 3.2.x seemingly competing,
> > >> >>> and now 3.3.x? For a community that does not have the resources to
> > >> >>> manage multiple release lines,
> > >> >>> you guys sure like to multiply release lines a lot.
> > >> >>>
> > >> >>> Cheers
> > >> >>> Rupert
> > >> >>>
> > >> >>> Am Mi., 4. März 2020 um 19:40 Uhr schrieb Wei-Chiu Chuang
> > >> >>> :
> > >> >>>
> > >>  Forwarding the discussion thread from the dev mailing lists to
> the
> > user
> > >>  mailing lists.
> > >> 
> > >>  I'd like to get an idea of how many users are still on Hadoop
> 2.9.
> > >>  Please share your thoughts.
> > >> 
> > >>  On Mon, Mar 2, 2020 at 6:30 PM Sree Vaddi
> > >>   wrote:
> > >> 
> > >> > +1
> > >> >
> > >> > Sent from Yahoo Mail on Android
> > >> >
> > >> >   On Mon, Mar 2, 2020 at 5:12 PM, Wei-Chiu Chuang<
> > weic...@apache.org>
> > >> > wrote:   Hi,
> > >> >
> > >> > Following the discussion to end branch-2.8, I want to start a
> > >> > discussion
> > >> > around what's next with branch-2.9. I am hesitant to use the
> word
> > "end
> > >> > of
> > >> > life" but consider these facts:
> > >> >
> > >> > * 2.9.0 was released Dec 17, 2017.
> > >> > * 2.9.2, the last 2.9.x release, went out Nov 19 

Re: [DISCUSS] fate of branch-2.9

2020-08-27 Thread Mingliang Liu
+1 for putting 2.9 lines to EOL.

Let's focus on 2.10 releases for Hadoop 2. Also is there any plan for
2.10.1? It has been 11 months since 2.10 first release.

Thanks,

On Wed, Aug 26, 2020 at 10:57 PM Wei-Chiu Chuang  wrote:

> Bump up this thread after 6 months.
>
> Is anyone still interested in the 2.9 release line? Or are we good to start
> the EOL process? The 2.9.2 was released in Nov 2018.
>
> I'd really like to see the community to converge to fewer release lines and
> make more frequent releases in each line.
>
> Thanks,
> Weichiu
>
>
> On Fri, Mar 6, 2020 at 5:47 PM Wei-Chiu Chuang  wrote:
>
> > I think that's a great suggestion.
> > Currently, we make 1 minor release per year, and within each minor
> release
> > we bring up 1 thousand to 2 thousand commits in it compared with the
> > previous one.
> > I can totally understand it is a big bite for users to swallow. Having a
> > more frequent release cycle, plus LTS and non-LTS releases should help
> with
> > this. (Of course we will need to make the release preparation much
> easier,
> > which is currently a pain)
> >
> > I am happy to discuss the release model further in the dev ML. LTS v.s.
> > non-LTS is one suggestion.
> >
> > Another similar issue: In the past Hadoop strived to
> > maintain compatibility. However, this is no longer sustainable as more
> CVEs
> > coming from our dependencies: netty, jetty, jackson ... etc.
> > In many cases, updating the dependencies brings breaking changes. More
> > recently, especially in Hadoop 3.x, I started to make the effort to
> update
> > dependencies much more frequently. How do users feel about this change?
> >
> > On Thu, Mar 5, 2020 at 7:58 AM Igor Dvorzhak 
> > wrote:
> >
> >> Maybe Hadoop will benefit from adopting a similar release and support
> >> strategy as Java? I.e. designate some releases as LTS and support them
> for
> >> 2 (?) years (it seems that 2.7.x branch was de-facto LTS), other non-LTS
> >> releases will be supported for 6 months (or until next release). This
> >> should allow to reduce maintenance cost of non-LTS release and provide
> >> conservative users desired stability by allowing them to wait for new
> LTS
> >> release and upgrading to it.
> >>
> >> On Thu, Mar 5, 2020 at 1:26 AM Rupert Mazzucco <
> rupert.mazzu...@gmail.com>
> >> wrote:
> >>
> >>> After recently jumping from 2.7.7 to 2.10 without issue myself, I vote
> >>> for keeping only the 2.10 line.
> >>> It would seem all other 2.x branches can upgrade to a 2.10.x easily if
> >>> they feel like upgrading at all,
> >>> unlike a jump to 3.x, which may require more planning.
> >>>
> >>> I also vote for having only one main 3.x branch. Why are there 3.1.x
> and
> >>> 3.2.x seemingly competing,
> >>> and now 3.3.x? For a community that does not have the resources to
> >>> manage multiple release lines,
> >>> you guys sure like to multiply release lines a lot.
> >>>
> >>> Cheers
> >>> Rupert
> >>>
> >>> Am Mi., 4. März 2020 um 19:40 Uhr schrieb Wei-Chiu Chuang
> >>> :
> >>>
>  Forwarding the discussion thread from the dev mailing lists to the
> user
>  mailing lists.
> 
>  I'd like to get an idea of how many users are still on Hadoop 2.9.
>  Please share your thoughts.
> 
>  On Mon, Mar 2, 2020 at 6:30 PM Sree Vaddi
>   wrote:
> 
> > +1
> >
> > Sent from Yahoo Mail on Android
> >
> >   On Mon, Mar 2, 2020 at 5:12 PM, Wei-Chiu Chuang >
> > wrote:   Hi,
> >
> > Following the discussion to end branch-2.8, I want to start a
> > discussion
> > around what's next with branch-2.9. I am hesitant to use the word
> "end
> > of
> > life" but consider these facts:
> >
> > * 2.9.0 was released Dec 17, 2017.
> > * 2.9.2, the last 2.9.x release, went out Nov 19 2018, which is more
> > than
> > 15 months ago.
> > * no one seems to be interested in being the release manager for
> 2.9.3.
> > * Most if not all of the active Hadoop contributors are using Hadoop
> > 2.10
> > or Hadoop 3.x.
> > * We as a community do not have the cycle to manage multiple release
> > line,
> > especially since Hadoop 3.3.0 is coming out soon.
> >
> > It is perhaps the time to gradually reduce our footprint in Hadoop
> > 2.x, and
> > encourage people to upgrade to Hadoop 3.x
> >
> > Thoughts?
> >
> >
>


-- 
L


[jira] [Created] (HDFS-15499) Exclude aws-java-sdk-bundle from httpfs

2020-07-29 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-15499:


 Summary: Exclude aws-java-sdk-bundle from httpfs
 Key: HDFS-15499
 URL: https://issues.apache.org/jira/browse/HDFS-15499
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: httpfs
Reporter: Mingliang Liu


In [[HADOOP-14040]] we use shaded aws-sdk uber-JAR for instead of s3 jar in 
hadoop-project/pom.xml. After that, we should update httpfs `pom.xml` to 
exclude the correct jar dependency.



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [DISCUSS] making Ozone a separate Apache project

2020-05-18 Thread Mingliang Liu
+1

On Mon, May 18, 2020 at 12:37 AM Elek, Marton  wrote:

>
>
> > One question, for the committers who contributed to Ozone before and got
> > the committer-role in the past (like me), will they carry the
> > committer-role to the new repo?
>
>
> In short: yes.
>
>
> In more details:
>
> This discussion (if there is an agreement) should be followed by a next
> discussion + vote about a very specific proposal which should contain
> all the technical information (including committer list)
>
> I support the the same approach what we followed with Submarine:
>
> ALL the existing (Hadoop) committers should have a free / opt-in
> opportunity to be a committer in Ozone.
>
> (After proposal is created on the wiki, you can add your name, or
> request to be added. But as the initial list can be created based on
> statistics from the Jira, your name can be already there ;-) )
>
>
>
> Marton
>
> -
> To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
> For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
>
>

-- 
L


[jira] [Created] (HDFS-15327) TestDataNodeErasureCodingMetrics.testReconstructionBytesPartialGroup3 fails intermittently

2020-05-02 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-15327:


 Summary: 
TestDataNodeErasureCodingMetrics.testReconstructionBytesPartialGroup3 fails 
intermittently
 Key: HDFS-15327
 URL: https://issues.apache.org/jira/browse/HDFS-15327
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, erasure-coding, test
Affects Versions: 3.4.0
Reporter: Mingliang Liu


Sample stack trace:

{code}
Error Message
ecReconstructionBytesRead should be  expected:<6501170> but was:<0>
Stacktrace
java.lang.AssertionError: ecReconstructionBytesRead should be  
expected:<6501170> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeErasureCodingMetrics.testReconstructionBytesPartialGroup3(TestDataNodeErasureCodingMetrics.java:150)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
{code}

Sample failing build:
- https://builds.apache.org/job/PreCommit-HDFS-Build/29226/testReport/



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-15324) TestRefreshCallQueue::testRefresh fails intermittently

2020-05-02 Thread Mingliang Liu (Jira)


 [ 
https://issues.apache.org/jira/browse/HDFS-15324?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-15324.
--
Resolution: Duplicate

> TestRefreshCallQueue::testRefresh fails intermittently
> --
>
> Key: HDFS-15324
> URL: https://issues.apache.org/jira/browse/HDFS-15324
> Project: Hadoop HDFS
>  Issue Type: Test
>  Components: namenode, test
>Affects Versions: 3.4.0
>        Reporter: Mingliang Liu
>Priority: Major
>
> This test seems flaky. It failed previously see HDFS-10253 and now it is 
> failing intermittently in {{trunk}}. Not sure we should use the same fix to 
> Mock as it was used in HDFS-10253.
> Sample failing builds:
> - 
> https://builds.apache.org/job/hadoop-multibranch/job/PR-1992/2/testReport/org.apache.hadoop/TestRefreshCallQueue/testRefresh/
> - 
> https://builds.apache.org/job/PreCommit-HDFS-Build/29221/testReport/org.apache.hadoop/TestRefreshCallQueue/testRefresh/
> Sample failing stack is:
> {code}
> Error Message
> org.apache.hadoop.TestRefreshCallQueue$MockCallQueue could not be constructed.
> Stacktrace
> java.lang.RuntimeException: 
> org.apache.hadoop.TestRefreshCallQueue$MockCallQueue could not be constructed.
>   at 
> org.apache.hadoop.ipc.CallQueueManager.createCallQueueInstance(CallQueueManager.java:193)
>   at 
> org.apache.hadoop.ipc.CallQueueManager.(CallQueueManager.java:83)
>   at org.apache.hadoop.ipc.Server.(Server.java:3087)
>   at org.apache.hadoop.ipc.RPC$Server.(RPC.java:1039)
>   at 
> org.apache.hadoop.ipc.ProtobufRpcEngine$Server.(ProtobufRpcEngine.java:427)
>   at 
> org.apache.hadoop.ipc.ProtobufRpcEngine.getServer(ProtobufRpcEngine.java:347)
>   at org.apache.hadoop.ipc.RPC$Builder.build(RPC.java:848)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.(NameNodeRpcServer.java:475)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcServer(NameNode.java:857)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:763)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:1014)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:987)
>   at 
> org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1756)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:1332)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster.configureNameService(MiniDFSCluster.java:1101)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:974)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:906)
>   at org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:534)
>   at 
> org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:493)
>   at 
> org.apache.hadoop.TestRefreshCallQueue.setUp(TestRefreshCallQueue.java:67)
>   at 
> org.apache.hadoop.TestRefreshCallQueue.testRefresh(TestRefreshCallQueue.java:115)
>   at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
>   at 
> sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
>   at 
> sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
>   at java.lang.reflect.Method.invoke(Method.java:498)
>   at 
> org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
>   at 
> org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
>   at 
> org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
>   at 
> org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
>   at 
> org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
>   at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
>   at 
> org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
>   at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
>   at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
>   at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
>   at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
>   at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
>   at org.junit.runners.ParentRun

[jira] [Created] (HDFS-15326) TestDataNodeErasureCodingMetrics::testFullBlock fails intermittently

2020-05-02 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-15326:


 Summary: TestDataNodeErasureCodingMetrics::testFullBlock fails 
intermittently
 Key: HDFS-15326
 URL: https://issues.apache.org/jira/browse/HDFS-15326
 Project: Hadoop HDFS
  Issue Type: Test
  Components: datanode, test
Affects Versions: 3.4.0
Reporter: Mingliang Liu


Sample failing build: 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1992/2/testReport/org.apache.hadoop.hdfs.server.datanode/TestDataNodeErasureCodingMetrics/testFullBlock/
Sample failing stack:
{code}
Error Message
Wrongly computed block reconstruction work
Stacktrace
java.lang.AssertionError: Wrongly computed block reconstruction work
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.assertTrue(Assert.java:41)
at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeErasureCodingMetrics.doTest(TestDataNodeErasureCodingMetrics.java:205)
at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeErasureCodingMetrics.testFullBlock(TestDataNodeErasureCodingMetrics.java:97)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
{code}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-15324) TestRefreshCallQueue::testRefresh fails intermittently

2020-05-02 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-15324:


 Summary: TestRefreshCallQueue::testRefresh fails intermittently
 Key: HDFS-15324
 URL: https://issues.apache.org/jira/browse/HDFS-15324
 Project: Hadoop HDFS
  Issue Type: Test
  Components: namenode, test
Affects Versions: 3.4.0
Reporter: Mingliang Liu


This test seems flaky. It failed previously see HDFS-10253 and now it is 
failing intermittently in {{trunk}}. Not sure we should use the same fix to 
Mock as it was used in HDFS-10253.

Sample failing builds:
- 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1992/2/testReport/org.apache.hadoop/TestRefreshCallQueue/testRefresh/
- 
https://builds.apache.org/job/PreCommit-HDFS-Build/29221/testReport/org.apache.hadoop/TestRefreshCallQueue/testRefresh/

Sample failing stack is:
{code}
Error Message
org.apache.hadoop.TestRefreshCallQueue$MockCallQueue could not be constructed.
Stacktrace
java.lang.RuntimeException: 
org.apache.hadoop.TestRefreshCallQueue$MockCallQueue could not be constructed.
at 
org.apache.hadoop.ipc.CallQueueManager.createCallQueueInstance(CallQueueManager.java:193)
at 
org.apache.hadoop.ipc.CallQueueManager.(CallQueueManager.java:83)
at org.apache.hadoop.ipc.Server.(Server.java:3087)
at org.apache.hadoop.ipc.RPC$Server.(RPC.java:1039)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server.(ProtobufRpcEngine.java:427)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine.getServer(ProtobufRpcEngine.java:347)
at org.apache.hadoop.ipc.RPC$Builder.build(RPC.java:848)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.(NameNodeRpcServer.java:475)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcServer(NameNode.java:857)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:763)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:1014)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:987)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1756)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:1332)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.configureNameService(MiniDFSCluster.java:1101)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:974)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:906)
at org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:534)
at 
org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:493)
at 
org.apache.hadoop.TestRefreshCallQueue.setUp(TestRefreshCallQueue.java:67)
at 
org.apache.hadoop.TestRefreshCallQueue.testRefresh(TestRefreshCallQueue.java:115)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384

Re: [Hadoop-3.3 Release update]- branch-3.3 has created

2020-04-24 Thread Mingliang Liu
Brahma,

What about https://issues.apache.org/jira/browse/HADOOP-17007?

Thanks,

On Fri, Apr 24, 2020 at 11:07 AM Brahma Reddy Battula 
wrote:

> Ok. Done. Branch created.
>
> Following blockers are pending, will closely track this.
>
> https://issues.apache.org/jira/browse/HDFS-15287 ( Open: Under discussion
> )
> https://issues.apache.org/jira/browse/YARN-10194 ( Patch Available)
> https://issues.apache.org/jira/browse/HDFS-15286 ( Patch Available)
> https://issues.apache.org/jira/browse/YARN-9898 ( Patch Available)
>
>
> On Fri, Apr 24, 2020 at 7:42 PM Wei-Chiu Chuang
>  wrote:
>
> > +1 we should have the branch ASAP.
> >
> > On Wed, Apr 22, 2020 at 11:07 PM Akira Ajisaka 
> > wrote:
> >
> > > > Since blockers are not closed, I didn't cut the branch because
> > > multiple branches might confuse or sombody might miss to commit.
> > >
> > > The current situation is already confusing. The 3.3.1 version already
> > > exists in JIRA, so some committers wrongly commit non-critical issues
> to
> > > branch-3.3 and set the fix version to 3.3.1.
> > > I think now we should cut branch-3.3.0 and freeze source code except
> the
> > > blockers.
> > >
> > > -Akira
> > >
> > > On Tue, Apr 21, 2020 at 3:05 PM Brahma Reddy Battula <
> bra...@apache.org>
> > > wrote:
> > >
> > >> Sure, I will do that.
> > >>
> > >> Since blockers are not closed, I didn't cut the branch because
> > >> multiple branches might confuse or sombody might miss to commit.Shall
> I
> > >> wait till this weekend to create..?
> > >>
> > >> On Mon, Apr 20, 2020 at 11:57 AM Akira Ajisaka 
> > >> wrote:
> > >>
> > >>> Hi Brahma,
> > >>>
> > >>> Thank you for preparing the release.
> > >>> Could you cut branch-3.3.0? I would like to backport some fixes for
> > >>> 3.3.1 and not for 3.3.0.
> > >>>
> > >>> Thanks and regards,
> > >>> Akira
> > >>>
> > >>> On Fri, Apr 17, 2020 at 11:11 AM Brahma Reddy Battula <
> > bra...@apache.org>
> > >>> wrote:
> > >>>
> >  Hi All,
> > 
> >  we are down to two blockers issues now (YARN-10194 and YARN-9848)
> > which
> >  are in patch available state.Hopefully we can out the RC soon.
> > 
> >  thanks to @Prabhu Joseph 
> > ,@masakate,@akira
> >  and @Wei-Chiu Chuang   and others for helping
> >  resloving the blockers.
> > 
> > 
> > 
> >  On Tue, Apr 14, 2020 at 10:49 PM Brahma Reddy Battula <
> >  bra...@apache.org> wrote:
> > 
> > >
> > > @Prabhu Joseph 
> > > >>> Have committed the YARN blocker YARN-10219 to trunk and
> > > cherry-picked to branch-3.3. Right now, there are two blocker
> Jiras -
> > > YARN-10233 and HADOOP-16982
> > > which i will help to review and commit. Thanks.
> > >
> > > Looks you committed YARN-10219. Noted YARN-10233 and HADOOP-16982
> as
> > a
> > > blockers. (without YARN-10233 we have given so many releases,it's
> > not newly
> > > introduced.).. Thanks
> > >
> > > @Vinod Kumar Vavilapalli  ,@adam Antal,
> > >
> > > I noted YARN-9848 as a blocker as you mentioned above.
> > >
> > > @All,
> > >
> > > Currently following four blockers are pending for 3.3.0 RC.
> > >
> > > HADOOP-16963,YARN-10233,HADOOP-16982 and YARN-9848.
> > >
> > >
> > >
> > > On Tue, Apr 14, 2020 at 8:11 PM Vinod Kumar Vavilapalli <
> > > vino...@apache.org> wrote:
> > >
> > >> Looks like a really bad bug to me.
> > >>
> > >> +1 for revert and +1 for making that a 3.3.0 blocker. I think
> should
> > >> also revert it in a 3.2 maintenance release too.
> > >>
> > >> Thanks
> > >> +Vinod
> > >>
> > >> > On Apr 14, 2020, at 5:03 PM, Adam Antal <
> adam.an...@cloudera.com
> > .INVALID>
> > >> wrote:
> > >> >
> > >> > Hi everyone,
> > >> >
> > >> > Sorry for coming a bit late with this, but there's also one jira
> > >> that can
> > >> > have potential impact on clusters and we should talk about it.
> > >> >
> > >> > Steven Rand found this problem earlier and commented to
> > >> > https://issues.apache.org/jira/browse/YARN-4946.
> > >> > The bug has impact on the RM state store: the RM does not delete
> > >> apps - see
> > >> > more details in his comment here:
> > >> >
> > >>
> >
> https://issues.apache.org/jira/browse/YARN-4946?focusedCommentId=16898599=com.atlassian.jira.plugin.system.issuetabpanels%3Acomment-tabpanel#comment-16898599
> > >> > .
> > >> > (FYI He also created
> > >> https://issues.apache.org/jira/browse/YARN-9848 with
> > >> > the revert task).
> > >> >
> > >> > It might not be an actual blocker, but since there wasn't any
> > >> consensus
> > >> > about a follow up action, I thought we should decide how to
> > proceed
> > >> before
> > >> > release 3.3.0.
> > >> >
> > >> > Regards,
> > >> > Adam
> > >> >
> > >> > On Tue, Apr 14, 2020 at 9:35 AM Prabhu Joseph <
> > >> 

[jira] [Created] (HDFS-15297) TestNNHandlesBlockReportPerStorage::blockReport_02 fails intermittently in trunk

2020-04-23 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-15297:


 Summary: TestNNHandlesBlockReportPerStorage::blockReport_02 fails 
intermittently in trunk
 Key: HDFS-15297
 URL: https://issues.apache.org/jira/browse/HDFS-15297
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: test
Affects Versions: 3.4.0
Reporter: Mingliang Liu


It fails intermittently on {{trunk}} branch. Not sure about other branches. 
Example builds are:
- 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1964/4/testReport/org.apache.hadoop.hdfs.server.datanode/TestNNHandlesBlockReportPerStorage/blockReport_02/
- 

Sample exception stack:
{quote}
java.lang.AssertionError: Wrong number of MissingBlocks is found expected:<2> 
but was:<1>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:834)
at org.junit.Assert.assertEquals(Assert.java:645)
at 
org.apache.hadoop.hdfs.server.datanode.BlockReportTestBase.blockReport_02(BlockReportTestBase.java:336)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:298)
at 
org.junit.internal.runners.statements.FailOnTimeout$CallableStatement.call(FailOnTimeout.java:292)
at java.util.concurrent.FutureTask.run(FutureTask.java:266)
at java.lang.Thread.run(Thread.java:748)
{quote}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-15296) TestBPOfferService::testMissBlocksWhenReregister is flaky

2020-04-23 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-15296:


 Summary: TestBPOfferService::testMissBlocksWhenReregister is flaky
 Key: HDFS-15296
 URL: https://issues.apache.org/jira/browse/HDFS-15296
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: test
Affects Versions: 3.4.0
Reporter: Mingliang Liu


TestBPOfferService.testMissBlocksWhenReregister fails intermittently in 
{{trunk}} branch, not sure about other branches. Example failures are
- 
https://builds.apache.org/job/hadoop-multibranch/job/PR-1964/4/testReport/org.apache.hadoop.hdfs.server.datanode/TestBPOfferService/testMissBlocksWhenReregister/
- 
https://builds.apache.org/job/PreCommit-HDFS-Build/29175/testReport/org.apache.hadoop.hdfs.server.datanode/TestBPOfferService/testMissBlocksWhenReregister/

Sample exception stack is:
{quote}
Stacktrace
java.lang.AssertionError
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hdfs.server.datanode.TestBPOfferService.testMissBlocksWhenReregister(TestBPOfferService.java:350)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:498)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:50)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:47)
at 
org.junit.internal.runners.statements.InvokeMethod.evaluate(InvokeMethod.java:17)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:26)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:325)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:78)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:57)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:290)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:71)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:288)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:58)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:268)
at org.junit.runners.ParentRunner.run(ParentRunner.java:363)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.execute(JUnit4Provider.java:365)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeWithRerun(JUnit4Provider.java:273)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.executeTestSet(JUnit4Provider.java:238)
at 
org.apache.maven.surefire.junit4.JUnit4Provider.invoke(JUnit4Provider.java:159)
at 
org.apache.maven.surefire.booter.ForkedBooter.invokeProviderInSameClassLoader(ForkedBooter.java:384)
at 
org.apache.maven.surefire.booter.ForkedBooter.runSuitesInProcess(ForkedBooter.java:345)
at 
org.apache.maven.surefire.booter.ForkedBooter.execute(ForkedBooter.java:126)
at 
org.apache.maven.surefire.booter.ForkedBooter.main(ForkedBooter.java:418)
{quote}



--
This message was sent by Atlassian Jira
(v8.3.4#803005)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-14786) A new block placement policy tolerating availability zone failure

2019-08-27 Thread Mingliang Liu (Jira)
Mingliang Liu created HDFS-14786:


 Summary: A new block placement policy tolerating availability zone 
failure
 Key: HDFS-14786
 URL: https://issues.apache.org/jira/browse/HDFS-14786
 Project: Hadoop HDFS
  Issue Type: New Feature
  Components: block placement
Reporter: Mingliang Liu


{{NetworkTopology}} assumes "/datacenter/rack/host" 3 layer topology. Default 
block placement policies are rack awareness for better fault tolerance Newer 
block placement policy like {{BlockPlacementPolicyRackFaultTolerant}} tries its 
best to place the replicas to most racks, which further tolerates more racks 
failing. [HADOOP-8470] brought {{NetworkTopologyWithNodeGroup}} to add another 
layer under rack, i.e. "/datacenter/rack/host/nodegroup" 4 layer topology. With 
that, replicas within a rack can be placed in different node groups for better 
isolation.

Existing block placement policies tolerate rack failure since at least two 
racks are chosen in those cases. Chances are all replicas could be placed in 
the same datacenter, though there are multiple data centers in the same cluster 
topology. In other words, fault of higher layers beyond rack is not well 
tolerated.

However, more deployments in public cloud are leveraging multiple available 
zones (AZ) for high-availability since the inter-AZ latency seems affordable in 
many cases. In a single AZ, some cloud providers like AWS support [partitioned 
placement 
groups|https://docs.aws.amazon.com/AWSEC2/latest/UserGuide/placement-groups.html#placement-groups-partition]
 which basically are different racks. A simple network topology mapped to HDFS 
is "/availabilityzone/rack/host" 3 layers.

To achieve high availability tolerating zone failure, this JIRA proposes a new 
data placement policy which tries its best to place replicas in most AZs, most 
racks, and most evenly distributed.

Examples with 3 replicas, we choose racks as following:
# 1AZ: fall back to {{BlockPlacementPolicyRackFaultTolerant}} to most racks
# 2AZ: randomly choose one rack in 1st AZ and randomly choose two racks in the 
other AZ
# 3AZ: randomly choose one rack in one AZ
# 4AZ: randomly choose three AZ and one rack in each AZ
After racks are chosen, hosts are chosen randomly honoring local storage, 
favorite nodes, excluded nodes, storage types etc.

Data may become imbalance if topology is very uneven in AZs. This seems not a 
problem as in public cloud, infrastructure provisioning is more flexible than 
1P.



--
This message was sent by Atlassian Jira
(v8.3.2#803003)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: HADOOP-14163 proposal for new hadoop.apache.org

2018-09-04 Thread Mingliang Liu
bout
> >>> the documentation which is generated by mvn-site (as before)
> >>>
> >>>
> >>> I got multiple valuable feedback and I improved the proposed site
> >>> according to the comments. Allen had some concerns about the used
> >>> technologies (hugo vs. mvn-site) and I answered all the questions why
> >>> I think mvn-site is the best for documentation and hugo is best for
> >>> generating site.
> >>>
> >>>
> >>> I would like to finish this effort/jira: I would like to start a
> >>> discussion about using this proposed version and approach as a new
> >>> site of Apache Hadoop. Please let me know what you think.
> >>>
> >>>
> >>> Thanks a lot,
> >>> Marton
> >>>
> >>> -
> >>> To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
> >>> For additional commands, e-mail: common-dev-h...@hadoop.apache.org
> >>>
> >>
> >>
> >> -
> >> To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
> >> For additional commands, e-mail: common-dev-h...@hadoop.apache.org
> >>
> >
> > -
> > To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
> > For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
>
> -
> To unsubscribe, e-mail: common-dev-unsubscr...@hadoop.apache.org
> For additional commands, e-mail: common-dev-h...@hadoop.apache.org
>
>

-- 
Mingliang Liu


Re: [VOTE] Release Apache Hadoop 2.8.4 (RC0)

2018-05-12 Thread Mingliang Liu
+1 (binding)

1. Download src file, and check md5
2. Build package from src
3. Create a 6 node cluster in Docker containers and start the HDFS/YARN
services (The script is caochong <https://github.com/weiqingy/caochong>).
4. Verify the running version
5. Run simple HDFS/YARN client/admin commands and verify the output
6. Run example programs wordcount and grep
7. Check NN/DN/RM status and logs

Thanks Junping for driving this!

On Tue, May 8, 2018 at 10:41 AM, 俊平堵 <junping...@apache.org> wrote:

> Hi all,
>  I've created the first release candidate (RC0) for Apache Hadoop
> 2.8.4. This is our next maint release to follow up 2.8.3. It includes 77
> important fixes and improvements.
>
> The RC artifacts are available at:
> http://home.apache.org/~junping_du/hadoop-2.8.4-RC0
>
> The RC tag in git is: release-2.8.4-RC0
>
> The maven artifacts are available via repository.apache.org<
> http://repository.apache.org> at:
> https://repository.apache.org/content/repositories/orgapachehadoop-1118
>
> Please try the release and vote; the vote will run for the usual 5
> working days, ending on 5/14/2018 PST time.
>
> Thanks,
>
> Junping
>



-- 
Mingliang Liu


Re: [VOTE] Adopt HDSL as a new Hadoop subproject

2018-03-22 Thread Mingliang Liu
+1 (binding)

Thanks!

On Tue, Mar 20, 2018 at 11:20 AM, Owen O'Malley <owen.omal...@gmail.com>
wrote:

> All,
>
> Following our discussions on the previous thread (Merging branch HDFS-7240
> to trunk), I'd like to propose the following:
>
> * HDSL become a subproject of Hadoop.
> * HDSL will release separately from Hadoop. Hadoop releases will not
> contain HDSL and vice versa.
> * HDSL will get its own jira instance so that the release tags stay
> separate.
> * On trunk (as opposed to release branches) HDSL will be a separate module
> in Hadoop's source tree. This will enable the HDSL to work on their trunk
> and the Hadoop trunk without making releases for every change.
> * Hadoop's trunk will only build HDSL if a non-default profile is enabled.
> * When Hadoop creates a release branch, the RM will delete the HDSL module
> from the branch.
> * HDSL will have their own Yetus checks and won't cause failures in the
> Hadoop patch check.
>
> I think this accomplishes most of the goals of encouraging HDSL development
> while minimizing the potential for disruption of HDFS development.
>
> The vote will run the standard 7 days and requires a lazy 2/3 vote. PMC
> votes are binding, but everyone is encouraged to vote.
>
> +1 (binding)
>
> .. Owen
>



-- 
Mingliang Liu


Re: [DISCUSS] official docker image(s) for hadoop

2017-09-13 Thread Mingliang Liu
> It would be very helpful for testing the RC.
For testing and voting, I have been using docker containers for a while, see 
code at: https://github.com/weiqingy/caochong 


> TL;DR: I propose to create official hadoop images and upload them to the 
> dockerhub
I’m +1 on this idea. The “official” docker image basically means a commitment 
to maintain well documented and broadly tested images, which seems not a burden 
to us.

Ceph has a community docker project https://github.com/ceph/ceph-docker 
 and I think our scope here is similar to 
it.

Mingliang

> On Sep 13, 2017, at 11:39 AM, Yufei Gu  wrote:
> 
> It would be very helpful for testing the RC. To vote a RC, committers and
> PMCs usually spend lots of time to compile, deploy the RC, do several
> sanity tests, then +1 for the RC. The docker image potentially saves the
> compilation and deployment time, and people can do more tests.
> 
> Best,
> 
> Yufei
> 
> On Wed, Sep 13, 2017 at 11:19 AM, Wangda Tan  wrote:
> 
>> +1 to add Hadoop docker image for easier testing / prototyping, it gonna be
>> super helpful!
>> 
>> Thanks,
>> Wangda
>> 
>> On Wed, Sep 13, 2017 at 10:48 AM, Miklos Szegedi <
>> miklos.szeg...@cloudera.com> wrote:
>> 
>>> Marton, thank you for working on this. I think Official Docker images for
>>> Hadoop would be very useful for a lot of reasons. I think that it is
>> better
>>> to have a coordinated effort with production ready base images with
>>> dependent images for prototyping. Does anyone else have an opinion about
>>> this?
>>> 
>>> Thank you,
>>> Miklos
>>> 
>>> On Fri, Sep 8, 2017 at 5:45 AM, Marton, Elek  wrote:
>>> 
 
 TL;DR: I propose to create official hadoop images and upload them to
>> the
 dockerhub.
 
 GOAL/SCOPE: I would like improve the existing documentation with
 easy-to-use docker based recipes to start hadoop clusters with various
 configuration.
 
 The images also could be used to test experimental features. For
>> example
 ozone could be tested easily with these compose file and configuration:
 
 https://gist.github.com/elek/1676a97b98f4ba561c9f51fce2ab2ea6
 
 Or even the configuration could be included in the compose file:
 
 https://github.com/elek/hadoop/blob/docker-2.8.0/example/doc
 ker-compose.yaml
 
 I would like to create separated example compose files for federation,
>>> ha,
 metrics usage, etc. to make it easier to try out and understand the
 features.
 
 CONTEXT: There is an existing Jira https://issues.apache.org/jira
 /browse/HADOOP-13397
 But it’s about a tool to generate production quality docker images
 (multiple types, in a flexible way). If no objections, I will create a
 separated issue to create simplified docker images for rapid
>> prototyping
 and investigating new features. And register the branch to the
>> dockerhub
>>> to
 create the images automatically.
 
 MY BACKGROUND: I am working with docker based hadoop/spark clusters
>> quite
 a while and run them succesfully in different environments (kubernetes,
 docker-swarm, nomad-based scheduling, etc.) My work is available from
>>> here:
 https://github.com/flokkr but they could handle more complex use cases
 (eg. instrumenting java processes with btrace, or read/reload
>>> configuration
 from consul).
 And IMHO in the official hadoop documentation it’s better to suggest
>> to
 use official apache docker images and not external ones (which could be
 changed).
 
 Please let me know if you have any comments.
 
 Marton
 
 -
 To unsubscribe, e-mail: yarn-dev-unsubscr...@hadoop.apache.org
 For additional commands, e-mail: yarn-dev-h...@hadoop.apache.org
 
 
>>> 
>> 



Re: [VOTE] Release Apache Hadoop 2.8.2 (RC0)

2017-09-10 Thread Mingliang Liu
Thanks Junping for doing this!

+1 (non-binding)

- Download the hadoop-2.8.2-src.tar.gz file and checked the md5 value
- Build package using maven (skipping tests) with Java 8
- Spin up a test cluster in Docker containers having 1 master node (NN/RM) and 
3 slave nodes (DN/NM)
- Operate the basic HDFS/YARN operations from command line, both client and 
admin
- Check NN/RM Web UI
- Run distcp to copy files from/to local and HDFS
- Run hadoop mapreduce examples: grep and wordcount
- Check the HDFS service logs

All looked good to me.

Mingliang

> On Sep 10, 2017, at 5:00 PM, Junping Du  wrote:
> 
> Hi folks,
> With fix of HADOOP-14842 get in, I've created our first release candidate 
> (RC0) for Apache Hadoop 2.8.2.
> 
> Apache Hadoop 2.8.2 is the first stable release of Hadoop 2.8 line and 
> will be the latest stable/production release for Apache Hadoop - it includes 
> 305 new fixed issues since 2.8.1 and 63 fixes are marked as blocker/critical 
> issues.
> 
>  More information about the 2.8.2 release plan can be found here: 
> https://cwiki.apache.org/confluence/display/HADOOP/Hadoop+2.8+Release
> 
>  New RC is available at: 
> http://home.apache.org/~junping_du/hadoop-2.8.2-RC0
> 
>  The RC tag in git is: release-2.8.2-RC0, and the latest commit id is: 
> e6597fe3000b06847d2bf55f2bab81770f4b2505
> 
>  The maven artifacts are available via repository.apache.org at: 
> https://repository.apache.org/content/repositories/orgapachehadoop-1062
> 
>  Please try the release and vote; the vote will run for the usual 5 days, 
> ending on 09/15/2017 5pm PST time.
> 
> Thanks,
> 
> Junping
> 


-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [VOTE] Release Apache Hadoop 2.7.4 (RC0)

2017-08-02 Thread Mingliang Liu
Thanks Konstantin!

+1 (non-binding)

I have tried several steps including:
- Download source package and check the checksums
- Build from source using Java 8
- Deploy a test cluster using docker containers w/ 5 DNs and 1 NN
- Run basic HDFS client/admin operations (e.g. read/write files, dfsadmin)
- Run NNThroughputBenchmark (using the new -fs option)
- Run DistCp and example apps (wordcount, sort)

All looked good to me.

> On Jul 29, 2017, at 4:29 PM, Konstantin Shvachko  wrote:
> 
> Hi everybody,
> 
> Here is the next release of Apache Hadoop 2.7 line. The previous stable
> release 2.7.3 was available since 25 August, 2016.
> Release 2.7.4 includes 264 issues fixed after release 2.7.3, which are
> critical bug fixes and major optimizations. See more details in Release
> Note:
> http://home.apache.org/~shv/hadoop-2.7.4-RC0/releasenotes.html
> 
> The RC0 is available at: http://home.apache.org/~shv/hadoop-2.7.4-RC0/
> 
> Please give it a try and vote on this thread. The vote will run for 5 days
> ending 08/04/2017.
> 
> Please note that my up to date public key are available from:
> https://dist.apache.org/repos/dist/release/hadoop/common/KEYS
> Please don't forget to refresh the page if you've been there recently.
> There are other place on Apache sites, which may contain my outdated key.
> 
> Thanks,
> --Konstantin


-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: Corona is here -- And Ozone works !!!!

2017-07-20 Thread Mingliang Liu
Still, it’s good news!

Best,

> On Jul 20, 2017, at 2:44 PM, Anu Engineer  wrote:
> 
> Sorry, it was meant for a wrong alias. My apologies.
> 
> —Anu
> 
> 
> 
> 
> 
> On 7/20/17, 2:40 PM, "Anu Engineer"  wrote:
> 
>> Hi All,
>> 
>> I just deployed a test cluster with Nandakumar and we were able to run 
>> corona from a single node, with 10 thread for 12 mins.
>> 
>> We were able to write 789 MB and were writing 66 keys per second from a 
>> single node.
>> 
>> ***
>> Number of Volumes created: 10
>> Number of Buckets created: 10
>> Number of Keys added: 78984
>> Execution time: 12 minutes
>> 
>> 
>> The fun fact, Ozone just worked :), This is the first time we have written 
>> something like 78K keys into ozone. We are just starting on Corona and we 
>> will expand to run it from multiple nodes.
>> 
>> —Anu
>> 
> 
> -
> To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
> For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org
> 


-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-12064) Reuse object mapper in HDFS

2017-06-28 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-12064:


 Summary: Reuse object mapper in HDFS
 Key: HDFS-12064
 URL: https://issues.apache.org/jira/browse/HDFS-12064
 Project: Hadoop HDFS
  Issue Type: Bug
Reporter: Mingliang Liu
Assignee: Hanisha Koneru
Priority: Minor


Currently there are a few places that are not following the recommended pattern 
of using object mapper - reuse if possible. Actually we can use 
{{ObjectReader}} or {{ObjectWriter}} to replace the object mapper in some 
places: they are straightforward and thread safe.

The benefit is all about performance, so in unit testing code I assume we don't 
have to worry too much.



--
This message was sent by Atlassian JIRA
(v6.4.14#64029)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11903) Cleaning up local storage when closing MiniOzoneCluster

2017-05-30 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11903:


 Summary: Cleaning up local storage when closing MiniOzoneCluster
 Key: HDFS-11903
 URL: https://issues.apache.org/jira/browse/HDFS-11903
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ozone
Reporter: Mingliang Liu
Assignee: Mingliang Liu


Currently we don't delete the local storage when closing {{MiniOzoneCluster}}, 
which will cause creating volume with the same name to fail, even in a new JVM 
process. Let's delete the local storage files: {{/tmp/ozone/metadata.db}} and 
{{/tmp/ozone/user.db}} when closing.

Please note this is not a complete solution. As the {{OzoneMetadataManager}}  
currently is a singleton, we are not able to create two independent 
MiniOzoneCluster in a single JVM. Data maybe shared unexpectedly. Let's address 
that in another JIRA, if needed.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11892) Azure: handle failure gracefully in case of missing account key

2017-05-26 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11892:


 Summary: Azure: handle failure gracefully in case of missing 
account key
 Key: HDFS-11892
 URL: https://issues.apache.org/jira/browse/HDFS-11892
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: fs/azure
Reporter: Mingliang Liu
Assignee: Mingliang Liu


Currently if the {{fs.azure.account.key.youraccount}} is missing, we will get 
error stack like this:

{code}
java.lang.IllegalArgumentException: The String is not a valid Base64-encoded 
string.

at com.microsoft.azure.storage.core.Base64.decode(Base64.java:63)
at 
com.microsoft.azure.storage.StorageCredentialsAccountAndKey.(StorageCredentialsAccountAndKey.java:81)
at 
org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.createStorageAccount(AzureBlobStorageTestAccount.java:464)
at 
org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.createTestAccount(AzureBlobStorageTestAccount.java:501)
at 
org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.create(AzureBlobStorageTestAccount.java:522)
at 
org.apache.hadoop.fs.azure.AzureBlobStorageTestAccount.create(AzureBlobStorageTestAccount.java:451)
at 
org.apache.hadoop.fs.azure.TestNativeAzureFileSystemAuthorization.createTestAccount(TestNativeAzureFileSystemAuthorization.java:50)
at 
org.apache.hadoop.fs.azure.AbstractWasbTestBase.setUp(AbstractWasbTestBase.java:47)
at sun.reflect.NativeMethodAccessorImpl.invoke0(Native Method)
at 
sun.reflect.NativeMethodAccessorImpl.invoke(NativeMethodAccessorImpl.java:62)
at 
sun.reflect.DelegatingMethodAccessorImpl.invoke(DelegatingMethodAccessorImpl.java:43)
at java.lang.reflect.Method.invoke(Method.java:497)
at 
org.junit.runners.model.FrameworkMethod$1.runReflectiveCall(FrameworkMethod.java:47)
at 
org.junit.internal.runners.model.ReflectiveCallable.run(ReflectiveCallable.java:12)
at 
org.junit.runners.model.FrameworkMethod.invokeExplosively(FrameworkMethod.java:44)
at 
org.junit.internal.runners.statements.RunBefores.evaluate(RunBefores.java:24)
at 
org.junit.internal.runners.statements.RunAfters.evaluate(RunAfters.java:27)
at 
org.junit.rules.ExpectedException$ExpectedExceptionStatement.evaluate(ExpectedException.java:168)
at org.junit.rules.RunRules.evaluate(RunRules.java:20)
at org.junit.runners.ParentRunner.runLeaf(ParentRunner.java:271)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:70)
at 
org.junit.runners.BlockJUnit4ClassRunner.runChild(BlockJUnit4ClassRunner.java:50)
at org.junit.runners.ParentRunner$3.run(ParentRunner.java:238)
at org.junit.runners.ParentRunner$1.schedule(ParentRunner.java:63)
at org.junit.runners.ParentRunner.runChildren(ParentRunner.java:236)
at org.junit.runners.ParentRunner.access$000(ParentRunner.java:53)
at org.junit.runners.ParentRunner$2.evaluate(ParentRunner.java:229)
at org.junit.runners.ParentRunner.run(ParentRunner.java:309)
at org.junit.runner.JUnitCore.run(JUnitCore.java:160)
at 
com.intellij.junit4.JUnit4IdeaTestRunner.startRunnerWithArgs(JUnit4IdeaTestRunner.java:68)
at 
com.intellij.rt.execution.junit.IdeaTestRunner$Repeater.startRunnerWithArgs(IdeaTestRunner.java:51)
at 
com.intellij.rt.execution.junit.JUnitStarter.prepareStreamsAndStart(JUnitStarter.java:242)
at 
com.intellij.rt.execution.junit.JUnitStarter.main(JUnitStarter.java:70)
{code}
Actually, this error message is not very helpful. I'd admire you if you can 
immediately find the root cause of this failure.

If the test account is missing, we simply skip the test with meaningful error 
message. Here we should do the same thing: skip the test, and tell the user why 
we skip it.



--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11704) OzoneFileSystem: A Hadoop file system implementation for Ozone

2017-04-25 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11704:


 Summary: OzoneFileSystem: A Hadoop file system implementation for 
Ozone
 Key: HDFS-11704
 URL: https://issues.apache.org/jira/browse/HDFS-11704
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: fs/ozone
Reporter: Mingliang Liu
Assignee: Mingliang Liu






--
This message was sent by Atlassian JIRA
(v6.3.15#6346)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-10891) Über-jira: Test Improvement by adding missing test cases for existing code

2016-12-28 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10891?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-10891.
--
Resolution: Fixed

> Über-jira: Test Improvement by adding missing test cases for existing code
> --
>
> Key: HDFS-10891
> URL: https://issues.apache.org/jira/browse/HDFS-10891
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: test
>    Reporter: Mingliang Liu
>        Assignee: Mingliang Liu
>
> We have covered some test cases in our nightly run and we'd like to backport 
> those test cases to the community code branch. This work includes the efforts 
> but not limited to:
> # Add more *meaningful* test cases for existing code so it's hard to break it 
> as stability and compatibility should always be guarded if possible
> # Refactor existing tests that share common code snippet (e.g. helper 
> methods) and/or logic (e.g. set up a MiniDFSCluster). One approach that is 
> not popular (which it should) is @Parameterized tests
> # Reduce unnecessary overhead in unit tests (e.g. long interval sleep, build 
> MiniDFSCluster multiple times instead of reusing the same one). Some of the 
> cases were addressed by [HDFS-10666].
> This is not a long-term work to cover all future improvement in unit tests. 
> We'll be happy to resolve this after our internal test cases (will file 
> separate JIRAs for that) are mostly addressed/considered.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Reopened] (HDFS-11105) TestRBWBlockInvalidation#testRWRInvalidation fails intermittently

2016-11-16 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu reopened HDFS-11105:
--

> TestRBWBlockInvalidation#testRWRInvalidation fails intermittently
> -
>
> Key: HDFS-11105
> URL: https://issues.apache.org/jira/browse/HDFS-11105
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode, test
>Affects Versions: 2.8.0
>Reporter: Yiqun Lin
>Assignee: Yiqun Lin
> Fix For: 2.8.0, 3.0.0-alpha2
>
> Attachments: HDFS-11105.001.patch, HDFS-11105.002.patch
>
>
> The test {{TestRBWBlockInvalidation#testRWRInvalidation}}  fails 
> intermittently. The stack infos:
> {code}
> org.junit.ComparisonFailure: expected:<[old gs data
> new gs data
> ]> but was:<[]>
>   at org.junit.Assert.assertEquals(Assert.java:115)
>   at org.junit.Assert.assertEquals(Assert.java:144)
>   at 
> org.apache.hadoop.hdfs.server.blockmanagement.TestRBWBlockInvalidation.testRWRInvalidation(TestRBWBlockInvalidation.java:225)
> {code}
> The issue is caused by the blocks reported not completed, similar to 
> HDFS-10499.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-11105) TestRBWBlockInvalidation#testRWRInvalidation fails intermittently

2016-11-16 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11105?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-11105.
--
Resolution: Fixed

> TestRBWBlockInvalidation#testRWRInvalidation fails intermittently
> -
>
> Key: HDFS-11105
> URL: https://issues.apache.org/jira/browse/HDFS-11105
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: namenode, test
>Affects Versions: 2.8.0
>Reporter: Yiqun Lin
>Assignee: Yiqun Lin
> Fix For: 2.8.0, 3.0.0-alpha2
>
> Attachments: HDFS-11105.001.patch, HDFS-11105.002.patch
>
>
> The test {{TestRBWBlockInvalidation#testRWRInvalidation}}  fails 
> intermittently. The stack infos:
> {code}
> org.junit.ComparisonFailure: expected:<[old gs data
> new gs data
> ]> but was:<[]>
>   at org.junit.Assert.assertEquals(Assert.java:115)
>   at org.junit.Assert.assertEquals(Assert.java:144)
>   at 
> org.apache.hadoop.hdfs.server.blockmanagement.TestRBWBlockInvalidation.testRWRInvalidation(TestRBWBlockInvalidation.java:225)
> {code}
> The issue is caused by the blocks reported not completed, similar to 
> HDFS-10499.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-10666) Über-jira: Unit tests should not use fixed sleep interval to wait for conditions

2016-11-11 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10666?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-10666.
--
  Resolution: Fixed
Target Version/s: 2.8.0  (was: 3.0.0-alpha2)

> Über-jira: Unit tests should not use fixed sleep interval to wait for 
> conditions
> 
>
> Key: HDFS-10666
> URL: https://issues.apache.org/jira/browse/HDFS-10666
> Project: Hadoop HDFS
>  Issue Type: Bug
>  Components: test
>        Reporter: Mingliang Liu
>    Assignee: Mingliang Liu
>
> There have been dozens of intermittent failing unit tests because they depend 
> on fixed-interval sleep to wait for conditions to reach before assertion. 
> This umbrella jira is to replace these sleep statements with:
> * {{GenericTestUtils.waitFor()}} to retry the conditions/assertions
> * Trigger internal state change of code to test, e.g. for {{MiniDFSCluster}} 
> we can trigger {BlockReports,HeartBeats,DeletionReports}
> * Fails fast if specific exceptions are caught
> * Coordinate the JUnit thread with activities of the internal threads
> * _ad-hoc fixes_...
> p.s. I don't know how closures in Java 8 comes into play but I'd like to see 
> any effort.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-11089) Accelerate TestDiskBalancerCommand using static shared MiniDFSCluster

2016-11-08 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-11089?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-11089.
--
Resolution: Not A Problem

Thanks [~arpitagarwal] for your suggestion. Closing this JIRA as "Not A 
Problem".

> Accelerate TestDiskBalancerCommand using static shared MiniDFSCluster
> -
>
> Key: HDFS-11089
> URL: https://issues.apache.org/jira/browse/HDFS-11089
> Project: Hadoop HDFS
>  Issue Type: Sub-task
>  Components: test
>Reporter: Xiaobing Zhou
>Assignee: Xiaobing Zhou
>
> It takes 50+ seconds to run the test suite. Similar to HDFS-11079, static 
> shared MiniDFSCluster will be used to accelerate the run.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11095) BlockManagerSafeMode should respect extension period default config value (30s)

2016-11-02 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11095:


 Summary: BlockManagerSafeMode should respect extension period 
default config value (30s)
 Key: HDFS-11095
 URL: https://issues.apache.org/jira/browse/HDFS-11095
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Mingliang Liu
Assignee: Mingliang Liu


{code:title=BlockManagerSafeMode.java}
this.extension = conf.getInt(DFS_NAMENODE_SAFEMODE_EXTENSION_KEY, 0);
{code}
Though the default value (30s) is loaded from {{hdfs-default.xml}}, we should 
also respect this in the code by using 
{{DFSConfigKeys#DFS_NAMENODE_SAFEMODE_EXTENSION_DEFAULT}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11085) Add unit tests for NameNode failing to startup when name dir can not be written

2016-11-01 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11085:


 Summary: Add unit tests for NameNode failing to startup when name 
dir can not be written
 Key: HDFS-11085
 URL: https://issues.apache.org/jira/browse/HDFS-11085
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: namenode, test
Reporter: Mingliang Liu


This can be placed in {{org.apache.hadoop.hdfs.server.namenode.TestStartup}} 
test class.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11083) Add unit test for DFSAdmin -report command

2016-11-01 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11083:


 Summary: Add unit test for DFSAdmin -report command
 Key: HDFS-11083
 URL: https://issues.apache.org/jira/browse/HDFS-11083
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: shell, test
Reporter: Mingliang Liu


{{hdfs dfsadmin -report}} has very useful information about the cluster. There 
are some existing customized tools that depend on this command functionality. 
We should add unit test for it. Specially,
# If one datanode is dead, the report should indicate this
# If one block is corrupt, the "Missing blocks:" field should report this
# etc...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11031) Add additional unit test for DataNode startup behavior when volumes fail

2016-10-18 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11031:


 Summary: Add additional unit test for DataNode startup behavior 
when volumes fail
 Key: HDFS-11031
 URL: https://issues.apache.org/jira/browse/HDFS-11031
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Mingliang Liu


There are several cases to add in {{TestDataNodeVolumeFailure}}:

- DataNode should not start in case of volumes failure
- DataNode should not start in case of lacking data dir read/write permission
- ...



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11030) TestDataNodeVolumeFailure#testVolumeFailure is flaky (though passing)

2016-10-18 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11030:


 Summary: TestDataNodeVolumeFailure#testVolumeFailure is flaky 
(though passing)
 Key: HDFS-11030
 URL: https://issues.apache.org/jira/browse/HDFS-11030
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: datanode, test
Affects Versions: 2.7.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


TestDataNodeVolumeFailure#testVolumeFailure fails a volume and verifies the 
blocks and files are replicated correctly.

To fail a volume, it deletes all the blocks and sets the data dir read only.
{code}
// fail the volume
// delete/make non-writable one of the directories (failed volume)
data_fail = new File(dataDir, "data3");
failedDir = MiniDFSCluster.getFinalizedDir(dataDir, 
cluster.getNamesystem().getBlockPoolId());
if (failedDir.exists() &&
//!FileUtil.fullyDelete(failedDir)
!deteteBlocks(failedDir)
) {
  throw new IOException("Could not delete hdfs directory '" + failedDir + 
"'");
}
data_fail.setReadOnly();
failedDir.setReadOnly();
{code}
However, there are two bugs here:
- The {{failedDir}} directory for finalized blocks is not calculated correctly. 
It should use {{data_fail}} instead of {{dataDir}} as the base directory.
- When deleting block files in {{deteteBlocks(failedDir)}}, it assumes that 
there is no subdirectories in the data dir. This assumption was also noted in 
the comments.
{quote}
// we use only small number of blocks to avoid creating subdirs in the data 
dir..
{quote}
This is not true. On my local cluster and MiniDFSCluster, there will be 
subdir0/subdir0/ two level directories regardless of the number of blocks.

These two bugs made the blocks not deleted.

To fail a volume, it also needs to trigger the DataNode removing the volume and 
send block report to NN. This is basically in the {{triggerFailure()}} method.
{code}
  /**
   * go to each block on the 2nd DataNode until it fails...
   * @param path
   * @param size
   * @throws IOException
   */
  private void triggerFailure(String path, long size) throws IOException {
NamenodeProtocols nn = cluster.getNameNodeRpc();
List locatedBlocks =
  nn.getBlockLocations(path, 0, size).getLocatedBlocks();

for (LocatedBlock lb : locatedBlocks) {
  DatanodeInfo dinfo = lb.getLocations()[1];
  ExtendedBlock b = lb.getBlock();
  try {
accessBlock(dinfo, lb);
  } catch (IOException e) {
System.out.println("Failure triggered, on block: " + b.getBlockId() +  
"; corresponding volume should be removed by now");
break;
  }
}
  }
{code}
Accessing those blocks will not trigger failures if the directory is read-only 
(while the block files are all there). I ran the tests multiple times without 
triggering this failure. We have to write new block files to the data 
directories, or we should have deleted the blocks correctly.

This unit test has been there for years and it seldom fails, just because it's 
never triggered a real volume failure.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11008) Add unit test for parsing "-source" parameter in Balancer tool

2016-10-13 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11008:


 Summary: Add unit test for parsing "-source" parameter in Balancer 
tool
 Key: HDFS-11008
 URL: https://issues.apache.org/jira/browse/HDFS-11008
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Mingliang Liu
Assignee: Mingliang Liu


[HDFS-8826] added an option to specify the source node list so that balancer 
only selects blocks to move from those nodes. This is very efficient in some 
cases.

The unit test of {{-source}} option parsing is missing, for both valid and 
invalid cases. This JIRA is to track the effort of adding unit test for that. 
In fact, we can refactor the existing code by replacing code that manually 
creates {{BalancerParameters}} with new code that explores the 
{{Balancer$Cli#parse()}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11007) AliyunOSSInputStream#read() should update read bytes stat correctly

2016-10-13 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11007:


 Summary: AliyunOSSInputStream#read() should update read bytes stat 
correctly
 Key: HDFS-11007
 URL: https://issues.apache.org/jira/browse/HDFS-11007
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: tools
Reporter: Mingliang Liu
Assignee: Mingliang Liu


{code}
  @Override
  public synchronized int read() throws IOException {
..
if (statistics != null && byteRead >= 0) {
  statistics.incrementBytesRead(1);
}
return byteRead;
  }
{code}

I believe it should be {{statistics.incrementBytesRead(byteRead);}}?



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-11002) Fix broken attr(5) link in ExtendedAttributes.md

2016-10-12 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-11002:


 Summary: Fix broken attr(5) link in ExtendedAttributes.md
 Key: HDFS-11002
 URL: https://issues.apache.org/jira/browse/HDFS-11002
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Affects Versions: 2.7.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


http://www.bestbits.at/acl/man/man5/attr.txt is broken though the website is 
not. I suggest we use the http://man7.org/linux/man-pages/man5/attr.5.html 
which is mainted by maintainer of the Linux 
[man-pages|https://www.kernel.org/doc/man-pages/] project.

The same to http://www.bestbits.at/acl/man/man1/getfattr.txt => 
http://man7.org/linux/man-pages/man1/getfattr.1.html
and http://www.bestbits.at/acl/man/man1/setfattr.txt => 
http://man7.org/linux/man-pages/man1/setfattr.1.html

[~ajisakaa], thoughts? Thanks.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10986) DFSAdmin should show detailed error message if any

2016-10-07 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10986:


 Summary: DFSAdmin should show detailed error message if any
 Key: HDFS-10986
 URL: https://issues.apache.org/jira/browse/HDFS-10986
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: tools
Affects Versions: 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


There are some subcommands in {{DFSAdmin}} that swallow IOException and give 
very limited error message, if any, to the stderr.

{code}
$ hdfs dfsadmin -getDatanodeInfo localhost:9866
Datanode unreachable.
$ hdfs dfsadmin -evictWriters 127.0.0.1:9866
$ echo $?
-1
{code}

We should fix this by providing detailed error message. Actually, the 
{{DFSAdmin#run}} already handles exception carefully. All we need to do is to 
not swallow exceptions without good reason.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10985) o.a.h.ha.TestZKFailoverController#testGracefulFailoverMultipleZKfcs fails intermittently

2016-10-07 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10985:


 Summary: 
o.a.h.ha.TestZKFailoverController#testGracefulFailoverMultipleZKfcs fails 
intermittently
 Key: HDFS-10985
 URL: https://issues.apache.org/jira/browse/HDFS-10985
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ha
Reporter: Mingliang Liu
Assignee: Mingliang Liu


h6. Error Message
{quote}
Unable to become active. Local node did not get an opportunity to do so from 
ZooKeeper, or the local node took too long to transition to active.
{quote}
h6. Stacktrace
{quote}
org.apache.hadoop.ha.ServiceFailedException: Unable to become active. Local 
node did not get an opportunity to do so from ZooKeeper, or the local node took 
too long to transition to active.
at 
org.apache.hadoop.ha.ZKFailoverController.doGracefulFailover(ZKFailoverController.java:690)
at 
org.apache.hadoop.ha.ZKFailoverController.access$400(ZKFailoverController.java:62)
at 
org.apache.hadoop.ha.ZKFailoverController$3.run(ZKFailoverController.java:607)
at 
org.apache.hadoop.ha.ZKFailoverController$3.run(ZKFailoverController.java:604)
at java.security.AccessController.doPrivileged(Native Method)
at javax.security.auth.Subject.doAs(Subject.java:422)
at 
org.apache.hadoop.security.UserGroupInformation.doAs(UserGroupInformation.java:1795)
at 
org.apache.hadoop.ha.ZKFailoverController.gracefulFailoverToYou(ZKFailoverController.java:604)
at 
org.apache.hadoop.ha.ZKFCRpcServer.gracefulFailover(ZKFCRpcServer.java:94)
at 
org.apache.hadoop.ha.TestZKFailoverController.testGracefulFailoverMultipleZKfcs(TestZKFailoverController.java:590)
{quote}

See recent failing build:
# 
https://builds.apache.org/job/PreCommit-HADOOP-Build/10705/testReport/org.apache.hadoop.ha/TestZKFailoverController/testGracefulFailoverMultipleZKfcs/
# to add more

I think we can use {{GenericTestUtils.waitFor()}} to retry the assertions.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10893) Refactor TestDFSShell by setting up MiniDFSCluser once for all commands test

2016-09-22 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10893:


 Summary: Refactor TestDFSShell by setting up MiniDFSCluser once 
for all commands test
 Key: HDFS-10893
 URL: https://issues.apache.org/jira/browse/HDFS-10893
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Minor


This is a minor change. It seems that setting up MiniDFSCluser once for all 
commands test will reduce the total time.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10892) Add unit tests for HDFS command 'dfs -tail'

2016-09-22 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10892:


 Summary: Add unit tests for HDFS command 'dfs -tail'
 Key: HDFS-10892
 URL: https://issues.apache.org/jira/browse/HDFS-10892
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs
Reporter: Mingliang Liu
Assignee: Mingliang Liu


I did not find unit test for HDFS command 'dfs -tail' in {{trunk}} code. I 
think it still merits to have one though the command itself has been used there 
for years.

It also happens to 'dfs -test'. Perhaps we can consolidate the patches here.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10891) Über-jira: Test Improvement by adding missing test cases for existing code

2016-09-22 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10891:


 Summary: Über-jira: Test Improvement by adding missing test cases 
for existing code
 Key: HDFS-10891
 URL: https://issues.apache.org/jira/browse/HDFS-10891
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: test
Reporter: Mingliang Liu
Assignee: Mingliang Liu


We have covered some test cases in our nightly run and we'd like to backport 
those test cases to the community code branch. This work includes the efforts 
but not limited to:
# Add more *meaningful* test cases for existing code so it's hard to break it 
as stability and compatibility should always be guarded if possible
# Refactor existing tests that share common code snippet (e.g. helper methods) 
and/or logic (e.g. set up a MiniDFSCluster). One approach that is not popular 
(which it should) is @Parameterized tests
# Reduce unnecessary overhead in unit tests (e.g. long interval sleep, build 
MiniDFSCluster multiple times instead of reusing the same one)

This is not a long-term work to cover all future improvement in unit tests. 
We'll be happy to resolve this after our internal test cases (will file 
separate JIRAs for that) are mostly addressed/considered.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10807) Doc about upgrading to a version of HDFS with snapshots may be confusing

2016-08-26 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10807:


 Summary: Doc about upgrading to a version of HDFS with snapshots 
may be confusing
 Key: HDFS-10807
 URL: https://issues.apache.org/jira/browse/HDFS-10807
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: documentation
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Minor


{code}
diff --git a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsSnapshots.md 
b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsSnapshots.md
index 94a37cd..d856e8c 100644
--- a/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsSnapshots.md
+++ b/hadoop-hdfs-project/hadoop-hdfs/src/site/markdown/HdfsSnapshots.md
@@ -113,7 +113,7 @@ Upgrading to a version of HDFS with snapshots
 
 The HDFS snapshot feature introduces a new reserved path name used to
 interact with snapshots: `.snapshot`. When upgrading from an
-older version of HDFS, existing paths named `.snapshot` need
+older version of HDFS which does not support snapshots, existing paths named 
`.snapshot` need
 to first be renamed or deleted to avoid conflicting with the reserved path.
 See the upgrade section in
 [the HDFS user guide](HdfsUserGuide.html#Upgrade_and_Rollback)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-10677) Über-jira: Enhancements to NNThroughputBenchmark tool

2016-08-15 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10677?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-10677.
--
   Resolution: Fixed
Fix Version/s: 2.8.0

> Über-jira: Enhancements to NNThroughputBenchmark tool
> -
>
> Key: HDFS-10677
> URL: https://issues.apache.org/jira/browse/HDFS-10677
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: benchmarks, tools
>    Reporter: Mingliang Liu
>        Assignee: Mingliang Liu
> Fix For: 2.8.0
>
>




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10747) o.a.h.hdfs.tools.DebugAdmin usage message is misleading

2016-08-10 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10747:


 Summary: o.a.h.hdfs.tools.DebugAdmin usage message is misleading
 Key: HDFS-10747
 URL: https://issues.apache.org/jira/browse/HDFS-10747
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs-client
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Minor


[HDFS-6917] added a helpful hdfs debug command to validate blocks and call 
recoverlease. The usage doc is kinda misleading, as following:

{code}
$ hdfs debug verify
creating a new configuration
verify [-meta ] [-block ]
  Verify HDFS metadata and block files.  If a block file is specified, we
  will verify that the checksums in the metadata file match the block
  file.
{code}
Actually the {{-meta }} is necessary. {{[]}} is for optional 
arguments, if we follow the 
[convention|http://pubs.opengroup.org/onlinepubs/9699919799].

{code}
$ hdfs debug recoverLease
creating a new configuration
recoverLease [-path ] [-retries ]
  Recover the lease on the specified path.  The path must reside on an
  HDFS filesystem.  The default number of retries is 1.
{code}
{{-path }} is also the same case.





--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10726) org.apache.hadoop.http.TestHttpServerLifecycle times out

2016-08-05 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10726:


 Summary: org.apache.hadoop.http.TestHttpServerLifecycle times out
 Key: HDFS-10726
 URL: https://issues.apache.org/jira/browse/HDFS-10726
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 3.0.0-alpha2
Reporter: Mingliang Liu


{{org.apache.hadoop.http.TestHttpServerLifecycle}} times out in a few of r 
recent builds, e.g.
# 
https://issues.apache.org/jira/browse/HADOOP-13470?focusedCommentId=15408969=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15408969
# 
https://issues.apache.org/jira/browse/HADOOP-13466?focusedCommentId=15408342=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15408342
# 
https://issues.apache.org/jira/browse/HADOOP-13444?focusedCommentId=15400162=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel#comment-15400162

Can not reproduced locally so the test failure is intermittent.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10725) Caller context should always be constructed by a builder

2016-08-05 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10725:


 Summary: Caller context should always be constructed by a builder
 Key: HDFS-10725
 URL: https://issues.apache.org/jira/browse/HDFS-10725
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ipc
Affects Versions: 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Trivial


Currently {{CallerContext}} is constructed by a builder. In this pattern, the 
constructor should be private so that caller context should always be 
constructed by a builder.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10724) Document the caller context config keys

2016-08-04 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10724:


 Summary: Document the caller context config keys
 Key: HDFS-10724
 URL: https://issues.apache.org/jira/browse/HDFS-10724
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: ipc, namenode
Affects Versions: 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Minor


The three config keys {{hadoop.caller.context.enabled}}, 
{{hadoop.caller.context.max.size}} and 
{{hadoop.caller.context.signature.max.size}} are not documented in the 
{{core-default.xml}} file. This jira is to address this.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: [VOTE] Release Apache Hadoop 2.7.3 RC0

2016-07-25 Thread Mingliang Liu
Thanks Vinod.

+1 (non-binding)

* Downloaded and built from source (with Java 8)
* Checked LICENSE and NOTICE files
* Verified MD5 signatures
* Installed pseudo-distributed instance (Mac OS X 10.11)
* Ran through HDFS and mapreduce tests
* Operate HDFS from command line
* Run tools like distcp/NNThroughputBenchmark

L

> On Jul 22, 2016, at 7:15 PM, Vinod Kumar Vavilapalli  
> wrote:
> 
> Hi all,
> 
> I've created a release candidate RC0 for Apache Hadoop 2.7.3.
> 
> As discussed before, this is the next maintenance release to follow up 2.7.2.
> 
> The RC is available for validation at: 
> http://home.apache.org/~vinodkv/hadoop-2.7.3-RC0/ 
> 
> 
> The RC tag in git is: release-2.7.3-RC0
> 
> The maven artifacts are available via repository.apache.org 
>  at 
> https://repository.apache.org/content/repositories/orgapachehadoop-1040/ 
> 
> 
> The release-notes are inside the tar-balls at location 
> hadoop-common-project/hadoop-common/src/main/docs/releasenotes.html. I hosted 
> this at http://home.apache.org/~vinodkv/hadoop-2.7.3-RC0/releasenotes.html 
>  for 
> your quick perusal.
> 
> As you may have noted, a very long fix-cycle for the License & Notice issues 
> (HADOOP-12893) caused 2.7.3 (along with every other Hadoop release) to slip 
> by quite a bit. This release's related discussion thread is linked below: [1].
> 
> Please try the release and vote; the vote will run for the usual 5 days.
> 
> Thanks,
> Vinod
> 
> [1]: 2.7.3 release plan: 
> https://www.mail-archive.com/hdfs-dev%40hadoop.apache.org/msg24439.html 
> 


-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10678) Documenting NNThroughputBenchmark tool

2016-07-21 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10678:


 Summary: Documenting NNThroughputBenchmark tool
 Key: HDFS-10678
 URL: https://issues.apache.org/jira/browse/HDFS-10678
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Mingliang Liu
Assignee: Mingliang Liu


The best (only) documentation for the NNThroughputBenchmark currently exists as 
a JavaDoc on the NNThroughputBenchmark class. This is less than useful, 
especially since we no longer generate javadocs for HDFS as part of the build 
process. I suggest we extract it into a separate markdown doc, or merge it with 
other benchmarking materials (if any?) about HDFS.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10677) Enhancements to NNThroughputBenchmark

2016-07-21 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10677:


 Summary: Enhancements to NNThroughputBenchmark
 Key: HDFS-10677
 URL: https://issues.apache.org/jira/browse/HDFS-10677
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: benchmarks
Reporter: Mingliang Liu
Assignee: Mingliang Liu






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10668) Fix intermittently failing UT TestDataNodeMXBean#testDataNodeMXBeanBlockCount

2016-07-21 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10668:


 Summary: Fix intermittently failing UT 
TestDataNodeMXBean#testDataNodeMXBeanBlockCount
 Key: HDFS-10668
 URL: https://issues.apache.org/jira/browse/HDFS-10668
 Project: Hadoop HDFS
  Issue Type: Sub-task
Reporter: Mingliang Liu
Assignee: Mingliang Liu


h6.Error Message
{code}
After delete one file expected:<4> but was:<5>
{code}

h6. Stacktrace
{code}
java.lang.AssertionError: After delete one file expected:<4> but was:<5>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeMXBean.testDataNodeMXBeanBlockCount(TestDataNodeMXBean.java:124)
{code}

Sample failing Jenkins pre-commit built, see 
[here|https://builds.apache.org/job/PreCommit-HDFS-Build/16094/testReport/org.apache.hadoop.hdfs.server.datanode/TestDataNodeMXBean/testDataNodeMXBeanBlockCount/].



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10666) Über-jira: Unit tests should not depend on nondeterministic behavior using fixed sleep interval

2016-07-21 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10666:


 Summary: Über-jira: Unit tests should not depend on 
nondeterministic behavior using fixed sleep interval
 Key: HDFS-10666
 URL: https://issues.apache.org/jira/browse/HDFS-10666
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0-alpha2
Reporter: Mingliang Liu
Assignee: Mingliang Liu


There have been dozens of intermittent failing unit tests because they depend 
on fixed-interval sleep to wait for conditions to reach before assertion. This 
umbrella jira is to replace these sleep statements with:
* {{GenericTestUtils.waitFor()}} to retry the conditions/assertions
* Trigger internal state change of {{MiniDFSCluster}}, e.g. 
{{trigger\{BlockReports,HeartBeats,DeletionReports\}}}
* fails fast if specific exceptions are caught
* _ad-hoc fixes_ (TBD)

p.s. I don't know how closures in Java 8 comes into play but I'd like to see 
any effort.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10644) Fix intermittently failing TestDFSUpgradeWithHA#testRollbackWithJournalNodes

2016-07-16 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10644:


 Summary: Fix intermittently failing 
TestDFSUpgradeWithHA#testRollbackWithJournalNodes
 Key: HDFS-10644
 URL: https://issues.apache.org/jira/browse/HDFS-10644
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0-alpha2
Reporter: Mingliang Liu


See [example stack trace | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/9996/testReport/org.apache.hadoop.hdfs.server.namenode.ha/TestDFSUpgradeWithHA/testRollbackWithJournalNodes/]
 along with stand output/error.

h6. Stacktrace
{code}
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestDFSUpgradeWithHA.testRollbackWithJournalNodes(TestDFSUpgradeWithHA.java:687)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10642) Fix intermittently failing UT TestLazyPersistReplicaRecovery#testDnRestartWithSavedReplicas

2016-07-16 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10642:


 Summary: Fix intermittently failing UT 
TestLazyPersistReplicaRecovery#testDnRestartWithSavedReplicas
 Key: HDFS-10642
 URL: https://issues.apache.org/jira/browse/HDFS-10642
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: datanode, test
Affects Versions: 3.0.0-alpha2
Reporter: Mingliang Liu


See [example stack trace | 
https://builds.apache.org/job/PreCommit-HADOOP-Build/10001/testReport/org.apache.hadoop.hdfs.server.datanode.fsdataset.impl/TestLazyPersistReplicaRecovery/testDnRestartWithSavedReplicas/].

h6. Error Message
{code}
 Expected: is 
 but: was 
{code}
h6. Stacktrace
{code}
java.lang.AssertionError: 
Expected: is 
 but: was 
at org.hamcrest.MatcherAssert.assertThat(MatcherAssert.java:20)
at org.junit.Assert.assertThat(Assert.java:865)
at org.junit.Assert.assertThat(Assert.java:832)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.LazyPersistTestCase.ensureFileReplicasOnStorageType(LazyPersistTestCase.java:141)
at 
org.apache.hadoop.hdfs.server.datanode.fsdataset.impl.TestLazyPersistReplicaRecovery.testDnRestartWithSavedReplicas(TestLazyPersistReplicaRecovery.java:53)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10641) Fix intermittently failing TestBlockManager#testBlockReportQueueing

2016-07-16 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10641:


 Summary: Fix intermittently failing 
TestBlockManager#testBlockReportQueueing
 Key: HDFS-10641
 URL: https://issues.apache.org/jira/browse/HDFS-10641
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: hdfs, test
Reporter: Mingliang Liu


See example failing [stack 
trace|https://builds.apache.org/job/PreCommit-HADOOP-Build/9996/testReport/org.apache.hadoop.hdfs.server.blockmanagement/TestBlockManager/testBlockReportQueueing/].

h6. Stacktrace
{code}
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hdfs.server.blockmanagement.TestBlockManager.testBlockReportQueueing(TestBlockManager.java:1074)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Reopened] (HDFS-9277) IOException "Unable to load OAuth2 connection factory." in TestWebHDFSOAuth2.listStatusReturnsAsExpected

2016-07-16 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9277?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu reopened HDFS-9277:
-

Thanks [~jojochuang] for reporting this. In our internal nightly test, this UT 
still fails intermittently. Reopen it for investigation.

> IOException "Unable to load OAuth2 connection factory." in 
> TestWebHDFSOAuth2.listStatusReturnsAsExpected
> 
>
> Key: HDFS-9277
> URL: https://issues.apache.org/jira/browse/HDFS-9277
> Project: Hadoop HDFS
>  Issue Type: Bug
>Reporter: Wei-Chiu Chuang
> Attachments: HDFS-9277.001.patch
>
>
> This test is failing consistently in Hadoop-hdfs-trunk and 
> Hadoop-hdfs-trunk-Java8 since September 22.
> REGRESSION:  
> org.apache.hadoop.hdfs.web.TestWebHDFSOAuth2.listStatusReturnsAsExpected
> Error Message:
> Unable to load OAuth2 connection factory.
> Stack Trace:
> java.io.IOException: Unable to load OAuth2 connection factory.
>   at java.io.FileInputStream.open(Native Method)
>   at java.io.FileInputStream.(FileInputStream.java:146)
>   at 
> org.apache.hadoop.security.ssl.ReloadingX509TrustManager.loadTrustManager(ReloadingX509TrustManager.java:164)
>   at 
> org.apache.hadoop.security.ssl.ReloadingX509TrustManager.(ReloadingX509TrustManager.java:81)
>   at 
> org.apache.hadoop.security.ssl.FileBasedKeyStoresFactory.init(FileBasedKeyStoresFactory.java:215)
>   at org.apache.hadoop.security.ssl.SSLFactory.init(SSLFactory.java:131)
>   at 
> org.apache.hadoop.hdfs.web.URLConnectionFactory.newSslConnConfigurator(URLConnectionFactory.java:135)
>   at 
> org.apache.hadoop.hdfs.web.URLConnectionFactory.newOAuth2URLConnectionFactory(URLConnectionFactory.java:110)
>   at 
> org.apache.hadoop.hdfs.web.WebHdfsFileSystem.initialize(WebHdfsFileSystem.java:158)
>   at 
> org.apache.hadoop.hdfs.web.TestWebHDFSOAuth2.listStatusReturnsAsExpected(TestWebHDFSOAuth2.java:147)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10622) o.a.h.security.TestGroupsCaching.testBackgroundRefreshCounters seems flaky

2016-07-13 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10622:


 Summary: 
o.a.h.security.TestGroupsCaching.testBackgroundRefreshCounters seems flaky
 Key: HDFS-10622
 URL: https://issues.apache.org/jira/browse/HDFS-10622
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: security, test
Affects Versions: 2.8.0
Reporter: Mingliang Liu


h5. Error Message

expected:<1> but was:<0>

h5. Stacktrace

java.lang.AssertionError: expected:<1> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at 
org.apache.hadoop.security.TestGroupsCaching.testBackgroundRefreshCounters(TestGroupsCaching.java:638)



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10556) DistCpOptions should be validated automatically

2016-06-21 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10556:


 Summary: DistCpOptions should be validated automatically
 Key: HDFS-10556
 URL: https://issues.apache.org/jira/browse/HDFS-10556
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


{{DistCpOptions}} can be set from command-line or may be set manually. In 
[HDFS-10397], we refactored the validation to make it simpler and more 
efficient. However, the newly added {{validate()}} method may not be 
automatically revoked. This is the major concern for existing downstreams that 
create the {{DistCpOptions}} manually instead of by parser, and have 
conflicting options.

This jira is to make the validation happen automatically. A simple fix is to 
use the approach that validates in individual setters. This is a fix for 
{{branch-2}}. As a long-term fix, in [HDFS-10533], we're making the 
{{DistCpOptions}} immutable so that it will be hard, if not impossible, to use 
it wrongly by downstream applications. However, that code will only go to 
{{trunk}} branch as it breaks backwards-compatibility.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10533) Make DistCpOptions class immutable

2016-06-15 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10533:


 Summary: Make DistCpOptions class immutable
 Key: HDFS-10533
 URL: https://issues.apache.org/jira/browse/HDFS-10533
 Project: Hadoop HDFS
  Issue Type: Improvement
  Components: distcp
Affects Versions: 3.0.0-alpha1
Reporter: Mingliang Liu
Assignee: Mingliang Liu


Currently the {{DistCpOptions}} class encapsulates all DistCp options, which 
may be set from command-line (via the {{OptionsParser}}) or may be set manually 
(eg construct an instance and call setters). As there are multiple option 
fields and more (e.g. [HDFS-9868], [HDFS-10314]) to add, validating them can be 
cumbersome. Ideally, the {{DistCpOptions}} object should be immutable. The 
benefits are:
# {{DistCpOptions}} is simple and easier to use and share, plus it scales well
# validation is automatic, e.g. manually constructed {{DistCpOptions}} gets 
validated before usage
# validation error message is well-defined which does not depend on the order 
of setters

This jira is to track the effort of making the {{DistCpOptions}} immutable by 
using a Builder pattern for creation.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



Re: cp and mv

2016-05-20 Thread Mingliang Liu
Kun,

I think you need to be ware of the difference between client and server side 
logic. Perhaps you’re more interested in the client side in this case. The 
commands are generally running in the shell, and org.apache.hadoop.fs.shell 
package is a good place to start. Specially, have a look at 
CommandWithDestination.java.

Ciao,

L

On May 20, 2016, at 12:05 PM, Kun Ren 
> wrote:

Hi Genius,

Currently I debugged the cp and mv operations, for example:
(1) ./bin/hdfs dfs -cp input/a.xml input/b.xml
(2)./bin/hdfs dfs -mv input/a.xml input/b.xml

My understanding is that for operation cp, it will create a new file b.xml,
and will copy the content of a.xml to b.xml; For mv operations, it will
create b.xml and copy the content of a.xml to b.xml, and delete the a.xml.

However, when I debug the code, i found that both operations will finally
go through the create() method in NameNodeRpcServer.java, but I didn't see
any calls to copy and delete function,  could you please point out to me
where I can debug and see the full logic of the cp and mv operations.
Thanks a lot.



[jira] [Created] (HDFS-10395) GlobalStorageStatistics should check null FileSystem scheme to avoid NPE

2016-05-12 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10395:


 Summary: GlobalStorageStatistics should check null FileSystem 
scheme to avoid NPE
 Key: HDFS-10395
 URL: https://issues.apache.org/jira/browse/HDFS-10395
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: fs
Affects Versions: 2.8.0
Reporter: Mingliang Liu






--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Resolved] (HDFS-10386) DataTransferProtocol#writeBlock missing some javadocs

2016-05-10 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-10386?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-10386.
--
Resolution: Duplicate

This is a duplicate of [HDFS-10387]. Re-open if I'm wrong. Thanks.

> DataTransferProtocol#writeBlock missing some javadocs
> -
>
> Key: HDFS-10386
> URL: https://issues.apache.org/jira/browse/HDFS-10386
> Project: Hadoop HDFS
>  Issue Type: Improvement
>Reporter: Yongjun Zhang
>Assignee: Yongjun Zhang
>
> DataTransferProtocol#writeBlock's javadocs has the following parameters 
> missing:
> {code}
>   final DataChecksum requestedChecksum,
>   final CachingStrategy cachingStrategy,
> {code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10383) Safely close resources in DFSTestUtil

2016-05-09 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10383:


 Summary: Safely close resources in DFSTestUtil
 Key: HDFS-10383
 URL: https://issues.apache.org/jira/browse/HDFS-10383
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Minor


There are a few of methods in {{DFSTestUtil}} that do not close the resource 
safely, or elegantly. We can use the try-with-resource statement to address 
this problem.

Specially, as {{DFSTestUtil}} is popularly used in test, we need to preserve 
any exceptions thrown during the processing of the resource while still 
guaranteeing it's closed finally. Take for example,the current implementation 
of {{DFSTestUtil#createFile()}} closes the FSDataOutputStream in the 
{{finally}} block, and when closing if the internal {{DFSOutputStream#close()}} 
throws any exception, which it often does, the exception thrown during the 
processing will be lost. See this [test 
failure|https://builds.apache.org/job/PreCommit-HADOOP-Build/9320/testReport/org.apache.hadoop.hdfs/TestAsyncDFSRename/testAggressiveConcurrentAsyncRenameWithOverwrite/],
 and we have to guess what was the root cause.

Using try-with-resource, we can close the resources safely, and the exceptions 
thrown both in processing and closing will be available (closing exception will 
be suppressed).




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)

-
To unsubscribe, e-mail: hdfs-dev-unsubscr...@hadoop.apache.org
For additional commands, e-mail: hdfs-dev-h...@hadoop.apache.org



[jira] [Created] (HDFS-10335) Mover$Processor#chooseTarget() always chooses the first matching target datanode

2016-04-26 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10335:


 Summary: Mover$Processor#chooseTarget() always chooses the first 
matching target datanode
 Key: HDFS-10335
 URL: https://issues.apache.org/jira/browse/HDFS-10335
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: balancer & mover
Affects Versions: 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Critical


Currently the 
{{org.apache.hadoop.hdfs.server.mover.Mover$Processor#chooseTarget()}} always 
chooses the first matching target datanode from the candidate list. This may 
make the mover schedule a lot of task to a few of the datanodes (first several 
datanodes of the candidate list). The overall performance will suffer 
significantly from this because of the saturated network/disk usage. Specially, 
if the {{dfs.datanode.balance.max.concurrent.moves}} is set, the scheduled move 
task will be queued on a few of the datanodes, regardless of other available 
storage resources. We need an algorithm which can distribute the move tasks 
approximately even across all the candidate target datanodes (storages).

Thanks [~szetszwo] for offline discussion.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-10306) SafeModeMonitor should not leave safe mode if name system is starting active service

2016-04-18 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10306:


 Summary: SafeModeMonitor should not leave safe mode if name system 
is starting active service
 Key: HDFS-10306
 URL: https://issues.apache.org/jira/browse/HDFS-10306
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: namenode
Reporter: Mingliang Liu
Assignee: Brahma Reddy Battula
 Fix For: 2.9.0


Scenario:
===
write some blocks
wait till roll edits happen
Stop SNN
Delete some blocks in ANN, wait till the blocks are deleted in DN also.
restart the SNN and Wait till block reports come from datanode to SNN
Kill ANN then make SNN to Active.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-10293) StripedFileTestUtil#readAll flaky

2016-04-14 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10293:


 Summary: StripedFileTestUtil#readAll flaky
 Key: HDFS-10293
 URL: https://issues.apache.org/jira/browse/HDFS-10293
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: erasure-coding, test
Affects Versions: 3.0.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


The flaky test helper method cause several UT test failing intermittently. For 
example, the 
{{TestDFSStripedOutputStreamWithFailure#testAddBlockWhenNoSufficientParityNumOfNodes}}
 timed out in a recent run (see 
[exception|https://builds.apache.org/job/PreCommit-HDFS-Build/15158/testReport/org.apache.hadoop.hdfs/TestDFSStripedOutputStreamWithFailure/testAddBlockWhenNoSufficientParityNumOfNodes/]),
 which can be easily reproduced locally.

Debugging at the code, chances are that the helper method is stuck in an 
infinite loop. We need a fix to make the test robust.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-10284) o.a.h.hdfs.server.blockmanagement.TestBlockManagerSafeMode.testCheckSafeMode fails intermittently

2016-04-13 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10284:


 Summary: 
o.a.h.hdfs.server.blockmanagement.TestBlockManagerSafeMode.testCheckSafeMode 
fails intermittently
 Key: HDFS-10284
 URL: https://issues.apache.org/jira/browse/HDFS-10284
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 2.9.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


*Stacktrace*

{code}
org.mockito.exceptions.misusing.UnfinishedStubbingException: 
Unfinished stubbing detected here:
-> at 
org.apache.hadoop.hdfs.server.blockmanagement.TestBlockManagerSafeMode.testCheckSafeMode(TestBlockManagerSafeMode.java:169)

E.g. thenReturn() may be missing.
Examples of correct stubbing:
when(mock.isOk()).thenReturn(true);
when(mock.isOk()).thenThrow(exception);
doThrow(exception).when(mock).someVoidMethod();
Hints:
 1. missing thenReturn()
 2. although stubbed methods may return mocks, you cannot inline mock creation 
(mock()) call inside a thenReturn method (see issue 53)

at 
org.apache.hadoop.hdfs.server.blockmanagement.TestBlockManagerSafeMode.testCheckSafeMode(TestBlockManagerSafeMode.java:169)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-10283) o.a.h.hdfs.server.namenode.TestFSImageWithSnapshot#testSaveLoadImageWithAppending fails intermittently

2016-04-13 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10283:


 Summary: 
o.a.h.hdfs.server.namenode.TestFSImageWithSnapshot#testSaveLoadImageWithAppending
 fails intermittently
 Key: HDFS-10283
 URL: https://issues.apache.org/jira/browse/HDFS-10283
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


The test fails with exception as following: 

{code}
java.io.IOException: Failed to replace a bad datanode on the existing pipeline 
due to no more good datanodes being available to try. (Nodes: 
current=[DatanodeInfoWithStorage[127.0.0.1:47227,DS-dd109c14-79e5-4380-ac5e-4434cd7e25b5,DISK],
 
DatanodeInfoWithStorage[127.0.0.1:56949,DS-6c0be75e-a78c-41b9-bfd0-7ee0cdefaa0e,DISK]],
 
original=[DatanodeInfoWithStorage[127.0.0.1:47227,DS-dd109c14-79e5-4380-ac5e-4434cd7e25b5,DISK],
 
DatanodeInfoWithStorage[127.0.0.1:56949,DS-6c0be75e-a78c-41b9-bfd0-7ee0cdefaa0e,DISK]]).
 The current failed datanode replacement policy is DEFAULT, and a client may 
configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' 
in its configuration.
at 
org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1162)
at 
org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1232)
at 
org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1423)
at 
org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1338)
at 
org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1321)
at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:599)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-10281) o.a.h.hdfs.server.namenode.ha.TestPendingCorruptDnMessages fails intermittently

2016-04-12 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10281:


 Summary: 
o.a.h.hdfs.server.namenode.ha.TestPendingCorruptDnMessages fails intermittently
 Key: HDFS-10281
 URL: https://issues.apache.org/jira/browse/HDFS-10281
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


In our daily UT test, we found the 
{{TestPendingCorruptDnMessages#testChangedStorageId}} failed intermittently, 
see following information:

*Error Message*

expected:<1> but was:<0>

*Stacktrace*
{code}
java.lang.AssertionError: expected:<1> but was:<0>
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:555)
at org.junit.Assert.assertEquals(Assert.java:542)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestPendingCorruptDnMessages.getRegisteredDatanodeUid(TestPendingCorruptDnMessages.java:124)
at 
org.apache.hadoop.hdfs.server.namenode.ha.TestPendingCorruptDnMessages.testChangedStorageId(TestPendingCorruptDnMessages.java:103)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-10201) Implement undo log in parity datanode for hflush operations

2016-03-23 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-10201:


 Summary: Implement undo log in parity datanode for hflush 
operations
 Key: HDFS-10201
 URL: https://issues.apache.org/jira/browse/HDFS-10201
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: erasure-coding
Reporter: Mingliang Liu
Assignee: Mingliang Liu


According to the current design doc for hflush support in erasure coding (see 
[HDFS-7661]), the parity datanode (DN) needs an undo log for flush operations. 
After hflush/hsync, the last cell will be overwritten when 1) the current strip 
is full, 2) the file is closed, 3) or the hflush/hsync is called again for the 
current non-full stripe. To serve new reader client and to tolerate failures 
between successful hflush/hsync and overwrite operation, the parity DN should 
preserve the old cell in the undo log before overwriting it.

As parities correspond to BG length and parity data of different BG length may 
have the same block length, the undo log should also save the respective block 
group (BG) length information for the flushed data.

This jira is to track the effort of designing and implementing an undo log in 
parity DN to support hflush/hsync operations.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9867) Missing block exception should carry locatedBlocks information

2016-02-26 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9867:
---

 Summary: Missing block exception should carry locatedBlocks 
information
 Key: HDFS-9867
 URL: https://issues.apache.org/jira/browse/HDFS-9867
 Project: Hadoop HDFS
  Issue Type: Sub-task
  Components: hdfs-client
Reporter: Mingliang Liu
Assignee: Mingliang Liu
Priority: Minor


When more than {{parityBlkNum}} internal blocks are missing, {{StripeReader}} 
throws IOException. Sample error message like:
{quote}
java.io.IOException: 5 missing blocks, the stripe is: Offset=44695552, 
length=65536, fetchedChunksNum=0, missingChunksNum=5
{quote}

According to our recent experience, it'd be useful for debugging and diagnosing 
to dump the current block group information.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Resolved] (HDFS-9846) Make DelegationTokenFetcher a Tool

2016-02-24 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9846?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-9846.
-
Resolution: Won't Fix

> Make DelegationTokenFetcher a Tool
> --
>
> Key: HDFS-9846
> URL: https://issues.apache.org/jira/browse/HDFS-9846
> Project: Hadoop HDFS
>  Issue Type: Improvement
>  Components: tools
>    Reporter: Mingliang Liu
>        Assignee: Mingliang Liu
>
> Currently the {{org.apache.hadoop.hdfs.tools.DelegationTokenFetcher}} is not 
> implementing the {{Tool}} interface, while it should.
> This jira is to track the effort of refactoring the code to implement the 
> {{Tool}} interface. The main benefits are unified generic option parsing, 
> modifying the configurations, and conjunction with {{ToolRunner}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9846) Make DelegationTokenFetcher a Tool

2016-02-22 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9846:
---

 Summary: Make DelegationTokenFetcher a Tool
 Key: HDFS-9846
 URL: https://issues.apache.org/jira/browse/HDFS-9846
 Project: Hadoop HDFS
  Issue Type: Improvement
Reporter: Mingliang Liu
Assignee: Mingliang Liu


Currently the {{org.apache.hadoop.hdfs.tools.DelegationTokenFetcher}} is not 
implementing the {{Tool}} interface, while it should.

This jira is to track the effort of refactoring the code to implement the 
{{Tool}} interface. The main benefits are unified generic option parsing, 
modifying the configurations, and conjunction with {{ToolRunner}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9765) TestBlockScanner#testVolumeIteratorWithCaching fails intermittently

2016-02-04 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9765:
---

 Summary: TestBlockScanner#testVolumeIteratorWithCaching fails 
intermittently
 Key: HDFS-9765
 URL: https://issues.apache.org/jira/browse/HDFS-9765
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Reporter: Mingliang Liu


It got timed out exception, with following stack:
{code}
java.lang.Exception: test timed out after 6 milliseconds
at java.lang.Thread.sleep(Native Method)
at 
org.apache.hadoop.hdfs.DFSOutputStream.completeFile(DFSOutputStream.java:812)
at 
org.apache.hadoop.hdfs.DFSOutputStream.closeImpl(DFSOutputStream.java:776)
at 
org.apache.hadoop.hdfs.DFSOutputStream.close(DFSOutputStream.java:747)
at 
org.apache.hadoop.fs.FSDataOutputStream$PositionCache.close(FSDataOutputStream.java:72)
at 
org.apache.hadoop.fs.FSDataOutputStream.close(FSDataOutputStream.java:101)
at org.apache.hadoop.hdfs.DFSTestUtil.createFile(DFSTestUtil.java:427)
at org.apache.hadoop.hdfs.DFSTestUtil.createFile(DFSTestUtil.java:376)
at org.apache.hadoop.hdfs.DFSTestUtil.createFile(DFSTestUtil.java:369)
at org.apache.hadoop.hdfs.DFSTestUtil.createFile(DFSTestUtil.java:362)
at 
org.apache.hadoop.hdfs.server.datanode.TestBlockScanner$TestContext.createFiles(TestBlockScanner.java:129)
at 
org.apache.hadoop.hdfs.server.datanode.TestBlockScanner.testVolumeIteratorImpl(TestBlockScanner.java:159)
at 
org.apache.hadoop.hdfs.server.datanode.TestBlockScanner.testVolumeIteratorWithCaching(TestBlockScanner.java:250)
{code}

See recent builds:
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14390/testReport/org.apache.hadoop.hdfs.server.datanode/TestBlockScanner/testVolumeIteratorWithCaching/
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14346/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs-jdk1.8.0_66.txt
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14392/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs-jdk1.8.0_66.txt
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14393/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs-jdk1.8.0_66.txt



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9767) TestFileAppend# fails intermittently

2016-02-04 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9767:
---

 Summary: TestFileAppend# fails intermittently
 Key: HDFS-9767
 URL: https://issues.apache.org/jira/browse/HDFS-9767
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0
Reporter: Mingliang Liu


*Stacktrace*:
{code}
java.io.IOException: Failed to replace a bad datanode on the existing pipeline 
due to no more good datanodes being available to try. (Nodes: 
current=[DatanodeInfoWithStorage[127.0.0.1:52139,DS-1db5fb50-ea3a-4ae2-8b37-ebda1a947c34,DISK],
 
DatanodeInfoWithStorage[127.0.0.1:49736,DS-2929c4a8-389a-4ff3-92be-1539018d17d9,DISK]],
 
original=[DatanodeInfoWithStorage[127.0.0.1:52139,DS-1db5fb50-ea3a-4ae2-8b37-ebda1a947c34,DISK],
 
DatanodeInfoWithStorage[127.0.0.1:49736,DS-2929c4a8-389a-4ff3-92be-1539018d17d9,DISK]]).
 The current failed datanode replacement policy is DEFAULT, and a client may 
configure this via 'dfs.client.block.write.replace-datanode-on-failure.policy' 
in its configuration.
at 
org.apache.hadoop.hdfs.DataStreamer.findNewDatanode(DataStreamer.java:1165)
at 
org.apache.hadoop.hdfs.DataStreamer.addDatanode2ExistingPipeline(DataStreamer.java:1235)
at 
org.apache.hadoop.hdfs.DataStreamer.handleDatanodeReplacement(DataStreamer.java:1426)
at 
org.apache.hadoop.hdfs.DataStreamer.setupPipelineInternal(DataStreamer.java:1341)
at 
org.apache.hadoop.hdfs.DataStreamer.setupPipelineForAppendOrRecovery(DataStreamer.java:1324)
at org.apache.hadoop.hdfs.DataStreamer.run(DataStreamer.java:598)
Standard Output
{code}

See recent builds:
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14352/testReport/org.apache.hadoop.hdfs/TestFileAppend/testMultipleAppends/
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14392/testReport/org.apache.hadoop.hdfs/TestFileAppend/testMultipleAppends/
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14315/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs-jdk1.8.0_66.txt



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9766) TestDataNodeMetrics#testDataNodeTimeSpend fails intermittently

2016-02-04 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9766:
---

 Summary: TestDataNodeMetrics#testDataNodeTimeSpend fails 
intermittently
 Key: HDFS-9766
 URL: https://issues.apache.org/jira/browse/HDFS-9766
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0
Reporter: Mingliang Liu


*Stacktrace*
{code}
java.lang.AssertionError: null
at org.junit.Assert.fail(Assert.java:86)
at org.junit.Assert.assertTrue(Assert.java:41)
at org.junit.Assert.assertTrue(Assert.java:52)
at 
org.apache.hadoop.hdfs.server.datanode.TestDataNodeMetrics.testDataNodeTimeSpend(TestDataNodeMetrics.java:289)
{code}

See recent builds:
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14393/testReport/org.apache.hadoop.hdfs.server.datanode/TestDataNodeMetrics/testDataNodeTimeSpend/
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14317/artifact/patchprocess/patch-unit-hadoop-hdfs-project_hadoop-hdfs-jdk1.8.0_66.txt




--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9716) o.a.h.hdfs.TestRecoverStripedFile fails intermittently in trunk

2016-01-27 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9716:
---

 Summary: o.a.h.hdfs.TestRecoverStripedFile fails intermittently in 
trunk
 Key: HDFS-9716
 URL: https://issues.apache.org/jira/browse/HDFS-9716
 Project: Hadoop HDFS
  Issue Type: Bug
Affects Versions: 2.8.0
Reporter: Mingliang Liu


See recent builds:
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14269/testReport/org.apache.hadoop.hdfs/TestRecoverStripedFile/testRecoverThreeDataBlocks1/
* 
https://builds.apache.org/job/PreCommit-HADOOP-Build/8477/testReport/org.apache.hadoop.hdfs/TestRecoverStripedFile/testRecoverThreeDataBlocks/



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9691) o.a.h.hdfs.server.blockmanagement.TestBlockManagerSafeMode.testCheckSafeMode fails intermittently

2016-01-23 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9691:
---

 Summary: 
o.a.h.hdfs.server.blockmanagement.TestBlockManagerSafeMode.testCheckSafeMode 
fails intermittently
 Key: HDFS-9691
 URL: https://issues.apache.org/jira/browse/HDFS-9691
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0, 2.8.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


It's a flaky test method and can rarely re-produce locally. We can see this 
happened in recent build, e.g. 
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14225/testReport/org.apache.hadoop.hdfs.server.blockmanagement/TestBlockManagerSafeMode/testCheckSafeMode/
* 
https://builds.apache.org/job/PreCommit-HDFS-Build/14139/testReport/org.apache.hadoop.hdfs.server.blockmanagement/TestBlockManagerSafeMode/testCheckSafeMode/

{code}
Error Message

expected: but was:
Stacktrace

java.lang.AssertionError: expected: but was:
at org.junit.Assert.fail(Assert.java:88)
at org.junit.Assert.failNotEquals(Assert.java:743)
at org.junit.Assert.assertEquals(Assert.java:118)
at org.junit.Assert.assertEquals(Assert.java:144)
at 
org.apache.hadoop.hdfs.server.blockmanagement.TestBlockManagerSafeMode.testCheckSafeMode(TestBlockManagerSafeMode.java:165)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9689) Test o.a.h.hdfs.TestRenameWhileOpen fails intermittently

2016-01-22 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9689:
---

 Summary: Test o.a.h.hdfs.TestRenameWhileOpen fails intermittently 
 Key: HDFS-9689
 URL: https://issues.apache.org/jira/browse/HDFS-9689
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0
Reporter: Mingliang Liu
Assignee: Mingliang Liu


The test fails in recent builds, e.g.
https://builds.apache.org/job/PreCommit-HDFS-Build/14063/testReport/org.apache.hadoop.hdfs/TestRenameWhileOpen/
and
https://builds.apache.org/job/PreCommit-HDFS-Build/14212/testReport/org.apache.hadoop.hdfs/TestRenameWhileOpen/testWhileOpenRenameToNonExistentDirectory/

The *Error Message* is like:
{code}
Problem binding to [localhost:60690] java.net.BindException: Address already in 
use; For more details see:  http://wiki.apache.org/hadoop/BindException
{code}
and *Stacktrace* is:
{code}

java.net.BindException: Problem binding to [localhost:60690] 
java.net.BindException: Address already in use; For more details see:  
http://wiki.apache.org/hadoop/BindException
at sun.nio.ch.Net.bind0(Native Method)
at sun.nio.ch.Net.bind(Net.java:463)
at sun.nio.ch.Net.bind(Net.java:455)
at 
sun.nio.ch.ServerSocketChannelImpl.bind(ServerSocketChannelImpl.java:223)
at sun.nio.ch.ServerSocketAdaptor.bind(ServerSocketAdaptor.java:74)
at org.apache.hadoop.ipc.Server.bind(Server.java:469)
at org.apache.hadoop.ipc.Server$Listener.(Server.java:695)
at org.apache.hadoop.ipc.Server.(Server.java:2464)
at org.apache.hadoop.ipc.RPC$Server.(RPC.java:958)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine$Server.(ProtobufRpcEngine.java:535)
at 
org.apache.hadoop.ipc.ProtobufRpcEngine.getServer(ProtobufRpcEngine.java:510)
at org.apache.hadoop.ipc.RPC$Builder.build(RPC.java:800)
at 
org.apache.hadoop.hdfs.server.namenode.NameNodeRpcServer.(NameNodeRpcServer.java:392)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createRpcServer(NameNode.java:743)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.initialize(NameNode.java:685)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:884)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.(NameNode.java:863)
at 
org.apache.hadoop.hdfs.server.namenode.NameNode.createNameNode(NameNode.java:1581)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNode(MiniDFSCluster.java:1247)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.configureNameService(MiniDFSCluster.java:1016)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.createNameNodesAndSetConf(MiniDFSCluster.java:891)
at 
org.apache.hadoop.hdfs.MiniDFSCluster.initMiniDFSCluster(MiniDFSCluster.java:823)
at org.apache.hadoop.hdfs.MiniDFSCluster.(MiniDFSCluster.java:482)
at 
org.apache.hadoop.hdfs.MiniDFSCluster$Builder.build(MiniDFSCluster.java:441)
at 
org.apache.hadoop.hdfs.TestRenameWhileOpen.testWhileOpenRenameToNonExistentDirectory(TestRenameWhileOpen.java:332)
{code}



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


[jira] [Created] (HDFS-9672) o.a.h.hdfs.TestLeaseRecovery2 fails intermittently

2016-01-20 Thread Mingliang Liu (JIRA)
Mingliang Liu created HDFS-9672:
---

 Summary: o.a.h.hdfs.TestLeaseRecovery2 fails intermittently
 Key: HDFS-9672
 URL: https://issues.apache.org/jira/browse/HDFS-9672
 Project: Hadoop HDFS
  Issue Type: Bug
  Components: test
Affects Versions: 3.0.0
Reporter: Mingliang Liu


It fails in recent builds, see:
https://builds.apache.org/job/PreCommit-HDFS-Build/14177/testReport/org.apache.hadoop.hdfs/
https://builds.apache.org/job/PreCommit-HDFS-uild/14147/testReport/org.apache.hadoop.hdfs/

Failing test methods include:
* 
org.apache.hadoop.hdfs.TestLeaseRecovery2.testHardLeaseRecoveryWithRenameAfterNameNodeRestart
* org.apache.hadoop.hdfs.TestLeaseRecovery2.testLeaseRecoverByAnotherUser
* org.apache.hadoop.hdfs.TestLeaseRecovery2.testHardLeaseRecovery
* 
org.apache.hadoop.hdfs.TestLeaseRecovery2.org.apache.hadoop.hdfs.TestLeaseRecovery2



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


Re: [VOTE] Release Apache Hadoop 2.7.2 RC1

2015-12-22 Thread Mingliang Liu
+1 (non-binding)

1. Download the src tar file and validate the integrity
2. Build and configure the local cluster
3. Operate the HDFS from command line interface
4. Run several example MR jobs
5. Check logs

Thanks.

L

> On Dec 16, 2015, at 6:49 PM, Vinod Kumar Vavilapalli  
> wrote:
> 
> Hi all,
> 
> I've created a release candidate RC1 for Apache Hadoop 2.7.2.
> 
> As discussed before, this is the next maintenance release to follow up 2.7.1.
> 
> The RC is available for validation at: 
> http://people.apache.org/~vinodkv/hadoop-2.7.2-RC1/ 
> 
> 
> The RC tag in git is: release-2.7.2-RC1
> 
> The maven artifacts are available via repository.apache.org 
>  at 
> https://repository.apache.org/content/repositories/orgapachehadoop-1026/ 
> 
> 
> The release-notes are inside the tar-balls at location 
> hadoop-common-project/hadoop-common/src/main/docs/releasenotes.html. I hosted 
> this at http://people.apache.org/~vinodkv/hadoop-2.7.2-RC1/releasenotes.html 
> for quick perusal.
> 
> As you may have noted,
> - The RC0 related voting thread got halted due to some critical issues. It 
> took a while again for getting all those blockers out of the way. See the 
> previous voting thread [3] for details.
> - Before RC0, an unusually long 2.6.3 release caused 2.7.2 to slip by quite a 
> bit. This release's related discussion threads are linked below: [1] and [2].
> 
> Please try the release and vote; the vote will run for the usual 5 days.
> 
> Thanks,
> Vinod
> 
> [1]: 2.7.2 release plan: http://markmail.org/message/oozq3gvd4nhzsaes 
> 
> [2]: Planning Apache Hadoop 2.7.2 
> http://markmail.org/message/iktqss2qdeykgpqk 
> 
> [3]: [VOTE] Release Apache Hadoop 2.7.2 RC0: 
> http://markmail.org/message/5txhvr2qdiqglrwc
> 



Re: [VOTE] Release Apache Hadoop 2.6.3 RC0

2015-12-16 Thread Mingliang Liu
+1 (non-binding)

1. Download the pre-built tar and check the validity
2. Configure and start a pseudo-distributed cluster
3. Run example grep MapReduce job locally
4. Operate the HDFS copying files from/to local directory
5. Check execution logs

All good. Thanks.

L

> On Dec 11, 2015, at 4:16 PM, Junping Du  wrote:
> 
> 
> Hi all developers in hadoop community,
>   I've created a release candidate RC0 for Apache Hadoop 2.6.3 (the next 
> maintenance release to follow up 2.6.2.) according to email thread of release 
> plan 2.6.3 [1]. Sorry for this RC coming a bit late as several blocker issues 
> were getting committed until yesterday. Below is the details:
> 
> The RC is available for validation at:
> *http://people.apache.org/~junping_du/hadoop-2.6.3-RC0/
> *
> 
> The RC tag in git is: release-2.6.3-RC0
> 
> The maven artifacts are staged via repository.apache.org at:
> *https://repository.apache.org/content/repositories/orgapachehadoop-1025/?
> *
> 
> You can find my public key at:
> http://svn.apache.org/repos/asf/hadoop/common/dist/KEYS
> 
> Please try the release and vote. The vote will run for the usual 5 days.
> 
> Thanks and happy weekend!
> 
> 
> Cheers,
> 
> Junping
> 
> 
> [1]: 2.6.3 release plan: http://markmail.org/thread/nc2jogbgni37vu6y
> 



[jira] [Resolved] (HDFS-9506) Move invalidate blocks to ReplicationManager

2015-12-04 Thread Mingliang Liu (JIRA)

 [ 
https://issues.apache.org/jira/browse/HDFS-9506?page=com.atlassian.jira.plugin.system.issuetabpanels:all-tabpanel
 ]

Mingliang Liu resolved HDFS-9506.
-
Resolution: Duplicate

As we keep updating the patch for [HDFS-9442], this issue is addressed there. 
Closing this jira as _Duplicate_.

> Move invalidate blocks to ReplicationManager
> 
>
> Key: HDFS-9506
> URL: https://issues.apache.org/jira/browse/HDFS-9506
> Project: Hadoop HDFS
>  Issue Type: Task
>  Components: namenode
>    Reporter: Mingliang Liu
>        Assignee: Mingliang Liu
>
> The [HDFS-9442] moves basic replication mechanism from {{BlockManager}} to 
> newly added {{ReplicationManager}}. After that we can move more replication 
> related logic to {{ReplicationManager}} as well, e.g. _invalidate blocks_ and 
> _corrupt replicas_. The goal is, again, cleaner code logic, well-organized 
> source files, and easier lock separating work in future.
> This jira is to track the effort of moving {{InvalidateBlocks}} to 
> {{ReplicationManager}}.



--
This message was sent by Atlassian JIRA
(v6.3.4#6332)


  1   2   >