Yes, Just create some directory with some contents in it within target directory. And set permission to 600. Then can run either 'mvn clean' or 'git clean'
-Vinay On Tue, Mar 17, 2015 at 9:13 PM, Sean Busbey <bus...@cloudera.com> wrote: > Is the simulation just removing the executable bit on the directory? I'd > like to get something I can reproduce locally. > > On Tue, Mar 17, 2015 at 2:29 AM, Vinayakumar B <vinayakum...@apache.org> > wrote: > > > I have simulated the problem in my env and verified that, both 'git clean > > -xdf' and 'mvn clean' will not remove the directory. > > mvn fails where as git simply ignores (not even display any warning) the > > problem. > > > > > > > > Regards, > > Vinay > > > > On Tue, Mar 17, 2015 at 2:32 AM, Sean Busbey <bus...@cloudera.com> > wrote: > > > > > Can someone point me to an example build that is broken? > > > > > > On Mon, Mar 16, 2015 at 3:52 PM, Sean Busbey <bus...@cloudera.com> > > wrote: > > > > > > > I'm on it. HADOOP-11721 > > > > > > > > On Mon, Mar 16, 2015 at 3:44 PM, Haohui Mai <whe...@apache.org> > wrote: > > > > > > > >> +1 for git clean. > > > >> > > > >> Colin, can you please get it in ASAP? Currently due to the jenkins > > > >> issues, we cannot close the 2.7 blockers. > > > >> > > > >> Thanks, > > > >> Haohui > > > >> > > > >> > > > >> > > > >> On Mon, Mar 16, 2015 at 11:54 AM, Colin P. McCabe < > cmcc...@apache.org > > > > > > >> wrote: > > > >> > If all it takes is someone creating a test that makes a directory > > > >> > without -x, this is going to happen over and over. > > > >> > > > > >> > Let's just fix the problem at the root by running "git clean > -fqdx" > > in > > > >> > our jenkins scripts. If there's no objections I will add this in > > and > > > >> > un-break the builds. > > > >> > > > > >> > best, > > > >> > Colin > > > >> > > > > >> > On Fri, Mar 13, 2015 at 1:48 PM, Lei Xu <l...@cloudera.com> wrote: > > > >> >> I filed HDFS-7917 to change the way to simulate disk failures. > > > >> >> > > > >> >> But I think we still need infrastructure folks to help with > jenkins > > > >> >> scripts to clean the dirs left today. > > > >> >> > > > >> >> On Fri, Mar 13, 2015 at 1:38 PM, Mai Haohui <ricet...@gmail.com> > > > >> wrote: > > > >> >>> Any updates on this issues? It seems that all HDFS jenkins > builds > > > are > > > >> >>> still failing. > > > >> >>> > > > >> >>> Regards, > > > >> >>> Haohui > > > >> >>> > > > >> >>> On Thu, Mar 12, 2015 at 12:53 AM, Vinayakumar B < > > > >> vinayakum...@apache.org> wrote: > > > >> >>>> I think the problem started from here. > > > >> >>>> > > > >> >>>> > > > >> > > > > > > https://builds.apache.org/job/PreCommit-HDFS-Build/9828/testReport/junit/org.apache.hadoop.hdfs.server.datanode/TestDataNodeVolumeFailure/testUnderReplicationAfterVolFailure/ > > > >> >>>> > > > >> >>>> As Chris mentioned TestDataNodeVolumeFailure is changing the > > > >> permission. > > > >> >>>> But in this patch, ReplicationMonitor got NPE and it got > > terminate > > > >> signal, > > > >> >>>> due to which MiniDFSCluster.shutdown() throwing Exception. > > > >> >>>> > > > >> >>>> But, TestDataNodeVolumeFailure#teardown() is restoring those > > > >> permission > > > >> >>>> after shutting down cluster. So in this case IMO, permissions > > were > > > >> never > > > >> >>>> restored. > > > >> >>>> > > > >> >>>> > > > >> >>>> @After > > > >> >>>> public void tearDown() throws Exception { > > > >> >>>> if(data_fail != null) { > > > >> >>>> FileUtil.setWritable(data_fail, true); > > > >> >>>> } > > > >> >>>> if(failedDir != null) { > > > >> >>>> FileUtil.setWritable(failedDir, true); > > > >> >>>> } > > > >> >>>> if(cluster != null) { > > > >> >>>> cluster.shutdown(); > > > >> >>>> } > > > >> >>>> for (int i = 0; i < 3; i++) { > > > >> >>>> FileUtil.setExecutable(new File(dataDir, "data"+(2*i+1)), > > > >> true); > > > >> >>>> FileUtil.setExecutable(new File(dataDir, "data"+(2*i+2)), > > > >> true); > > > >> >>>> } > > > >> >>>> } > > > >> >>>> > > > >> >>>> > > > >> >>>> Regards, > > > >> >>>> Vinay > > > >> >>>> > > > >> >>>> On Thu, Mar 12, 2015 at 12:35 PM, Vinayakumar B < > > > >> vinayakum...@apache.org> > > > >> >>>> wrote: > > > >> >>>> > > > >> >>>>> When I see the history of these kind of builds, All these are > > > >> failed on > > > >> >>>>> node H9. > > > >> >>>>> > > > >> >>>>> I think some or the other uncommitted patch would have created > > the > > > >> problem > > > >> >>>>> and left it there. > > > >> >>>>> > > > >> >>>>> > > > >> >>>>> Regards, > > > >> >>>>> Vinay > > > >> >>>>> > > > >> >>>>> On Thu, Mar 12, 2015 at 6:16 AM, Sean Busbey < > > bus...@cloudera.com > > > > > > > >> wrote: > > > >> >>>>> > > > >> >>>>>> You could rely on a destructive git clean call instead of > maven > > > to > > > >> do the > > > >> >>>>>> directory removal. > > > >> >>>>>> > > > >> >>>>>> -- > > > >> >>>>>> Sean > > > >> >>>>>> On Mar 11, 2015 4:11 PM, "Colin McCabe" < > > cmcc...@alumni.cmu.edu> > > > >> wrote: > > > >> >>>>>> > > > >> >>>>>> > Is there a maven plugin or setting we can use to simply > > remove > > > >> >>>>>> > directories that have no executable permissions on them? > > > >> Clearly we > > > >> >>>>>> > have the permission to do this from a technical point of > view > > > >> (since > > > >> >>>>>> > we created the directories as the jenkins user), it's > simply > > > >> that the > > > >> >>>>>> > code refuses to do it. > > > >> >>>>>> > > > > >> >>>>>> > Otherwise I guess we can just fix those tests... > > > >> >>>>>> > > > > >> >>>>>> > Colin > > > >> >>>>>> > > > > >> >>>>>> > On Tue, Mar 10, 2015 at 2:43 PM, Lei Xu <l...@cloudera.com> > > > >> wrote: > > > >> >>>>>> > > Thanks a lot for looking into HDFS-7722, Chris. > > > >> >>>>>> > > > > > >> >>>>>> > > In HDFS-7722: > > > >> >>>>>> > > TestDataNodeVolumeFailureXXX tests reset data dir > > permissions > > > >> in > > > >> >>>>>> > TearDown(). > > > >> >>>>>> > > TestDataNodeHotSwapVolumes reset permissions in a finally > > > >> clause. > > > >> >>>>>> > > > > > >> >>>>>> > > Also I ran mvn test several times on my machine and all > > tests > > > >> passed. > > > >> >>>>>> > > > > > >> >>>>>> > > However, since in DiskChecker#checkDirAccess(): > > > >> >>>>>> > > > > > >> >>>>>> > > private static void checkDirAccess(File dir) throws > > > >> >>>>>> DiskErrorException { > > > >> >>>>>> > > if (!dir.isDirectory()) { > > > >> >>>>>> > > throw new DiskErrorException("Not a directory: " > > > >> >>>>>> > > + dir.toString()); > > > >> >>>>>> > > } > > > >> >>>>>> > > > > > >> >>>>>> > > checkAccessByFileMethods(dir); > > > >> >>>>>> > > } > > > >> >>>>>> > > > > > >> >>>>>> > > One potentially safer alternative is replacing data dir > > with > > > a > > > >> regular > > > >> >>>>>> > > file to stimulate disk failures. > > > >> >>>>>> > > > > > >> >>>>>> > > On Tue, Mar 10, 2015 at 2:19 PM, Chris Nauroth < > > > >> >>>>>> cnaur...@hortonworks.com> > > > >> >>>>>> > wrote: > > > >> >>>>>> > >> TestDataNodeHotSwapVolumes, TestDataNodeVolumeFailure, > > > >> >>>>>> > >> TestDataNodeVolumeFailureReporting, and > > > >> >>>>>> > >> TestDataNodeVolumeFailureToleration all remove > executable > > > >> permissions > > > >> >>>>>> > from > > > >> >>>>>> > >> directories like the one Colin mentioned to simulate > disk > > > >> failures at > > > >> >>>>>> > data > > > >> >>>>>> > >> nodes. I reviewed the code for all of those, and they > all > > > >> appear to > > > >> >>>>>> be > > > >> >>>>>> > >> doing the necessary work to restore executable > permissions > > > at > > > >> the > > > >> >>>>>> end of > > > >> >>>>>> > >> the test. The only recent uncommitted patch I¹ve seen > > that > > > >> makes > > > >> >>>>>> > changes > > > >> >>>>>> > >> in these test suites is HDFS-7722. That patch still > looks > > > >> fine > > > >> >>>>>> > though. I > > > >> >>>>>> > >> don¹t know if there are other uncommitted patches that > > > >> changed these > > > >> >>>>>> > test > > > >> >>>>>> > >> suites. > > > >> >>>>>> > >> > > > >> >>>>>> > >> I suppose it¹s also possible that the JUnit process > > > >> unexpectedly died > > > >> >>>>>> > >> after removing executable permissions but before > restoring > > > >> them. > > > >> >>>>>> That > > > >> >>>>>> > >> always would have been a weakness of these test suites, > > > >> regardless of > > > >> >>>>>> > any > > > >> >>>>>> > >> recent changes. > > > >> >>>>>> > >> > > > >> >>>>>> > >> Chris Nauroth > > > >> >>>>>> > >> Hortonworks > > > >> >>>>>> > >> http://hortonworks.com/ > > > >> >>>>>> > >> > > > >> >>>>>> > >> > > > >> >>>>>> > >> > > > >> >>>>>> > >> > > > >> >>>>>> > >> > > > >> >>>>>> > >> > > > >> >>>>>> > >> On 3/10/15, 1:47 PM, "Aaron T. Myers" <a...@cloudera.com > > > > > >> wrote: > > > >> >>>>>> > >> > > > >> >>>>>> > >>>Hey Colin, > > > >> >>>>>> > >>> > > > >> >>>>>> > >>>I asked Andrew Bayer, who works with Apache Infra, > what's > > > >> going on > > > >> >>>>>> with > > > >> >>>>>> > >>>these boxes. He took a look and concluded that some > perms > > > are > > > >> being > > > >> >>>>>> set > > > >> >>>>>> > in > > > >> >>>>>> > >>>those directories by our unit tests which are precluding > > > >> those files > > > >> >>>>>> > from > > > >> >>>>>> > >>>getting deleted. He's going to clean up the boxes for > us, > > > but > > > >> we > > > >> >>>>>> should > > > >> >>>>>> > >>>expect this to keep happening until we can fix the test > in > > > >> question > > > >> >>>>>> to > > > >> >>>>>> > >>>properly clean up after itself. > > > >> >>>>>> > >>> > > > >> >>>>>> > >>>To help narrow down which commit it was that started > this, > > > >> Andrew > > > >> >>>>>> sent > > > >> >>>>>> > me > > > >> >>>>>> > >>>this info: > > > >> >>>>>> > >>> > > > >> >>>>>> > >>>"/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS- > > > >> >>>>>> > > > > >> >>>>>> > > > >> > > > > >>>Build/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data3/ > > > >> >>>>>> > has > > > >> >>>>>> > >>>500 perms, so I'm guessing that's the problem. Been that > > way > > > >> since > > > >> >>>>>> 9:32 > > > >> >>>>>> > >>>UTC > > > >> >>>>>> > >>>on March 5th." > > > >> >>>>>> > >>> > > > >> >>>>>> > >>>-- > > > >> >>>>>> > >>>Aaron T. Myers > > > >> >>>>>> > >>>Software Engineer, Cloudera > > > >> >>>>>> > >>> > > > >> >>>>>> > >>>On Tue, Mar 10, 2015 at 1:24 PM, Colin P. McCabe < > > > >> cmcc...@apache.org > > > >> >>>>>> > > > > >> >>>>>> > >>>wrote: > > > >> >>>>>> > >>> > > > >> >>>>>> > >>>> Hi all, > > > >> >>>>>> > >>>> > > > >> >>>>>> > >>>> A very quick (and not thorough) survey shows that I > > can't > > > >> find any > > > >> >>>>>> > >>>> jenkins jobs that succeeded from the last 24 hours. > > Most > > > >> of them > > > >> >>>>>> seem > > > >> >>>>>> > >>>> to be failing with some variant of this message: > > > >> >>>>>> > >>>> > > > >> >>>>>> > >>>> [ERROR] Failed to execute goal > > > >> >>>>>> > >>>> org.apache.maven.plugins:maven-clean-plugin:2.5:clean > > > >> >>>>>> (default-clean) > > > >> >>>>>> > >>>> on project hadoop-hdfs: Failed to clean project: > Failed > > to > > > >> delete > > > >> >>>>>> > >>>> > > > >> >>>>>> > >>>> > > > >> >>>>>> > > > > >> >>>>>> > > > > >> >>>>>> > > > >> > > > > > > >>>>/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/hadoop-hdfs-pr > > > >> >>>>>> > >>>>oject/hadoop-hdfs/target/test/data/dfs/data/data3 > > > >> >>>>>> > >>>> -> [Help 1] > > > >> >>>>>> > >>>> > > > >> >>>>>> > >>>> Any ideas how this happened? Bad disk, unit test > > setting > > > >> wrong > > > >> >>>>>> > >>>> permissions? > > > >> >>>>>> > >>>> > > > >> >>>>>> > >>>> Colin > > > >> >>>>>> > >>>> > > > >> >>>>>> > >> > > > >> >>>>>> > > > > > >> >>>>>> > > > > > >> >>>>>> > > > > > >> >>>>>> > > -- > > > >> >>>>>> > > Lei (Eddy) Xu > > > >> >>>>>> > > Software Engineer, Cloudera > > > >> >>>>>> > > > > >> >>>>>> > > > >> >>>>> > > > >> >>>>> > > > >> >> > > > >> >> > > > >> >> > > > >> >> -- > > > >> >> Lei (Eddy) Xu > > > >> >> Software Engineer, Cloudera > > > >> > > > > > > > > > > > > > > > > -- > > > > Sean > > > > > > > > > > > > > > > > -- > > > Sean > > > > > > > > > -- > Sean >