Seems like all builds of Precommit-HDFS-Build failing with below problem. FATAL: Command "git clean -fdx" returned status code 1: stdout: stderr: hudson.plugins.git.GitException <http://stacktrace.jenkins-ci.org/search?query=hudson.plugins.git.GitException>: Command "git clean -fdx" returned status code 1: stdout: stderr:
Can someone remove "git clean -fdx" from build configurations of Precommit-HDFS-Build ? Regards, Vinay On Tue, Mar 17, 2015 at 12:59 PM, Vinayakumar B <vinayakum...@apache.org> wrote: > I have simulated the problem in my env and verified that, both 'git clean > -xdf' and 'mvn clean' will not remove the directory. > mvn fails where as git simply ignores (not even display any warning) the > problem. > > > > Regards, > Vinay > > On Tue, Mar 17, 2015 at 2:32 AM, Sean Busbey <bus...@cloudera.com> wrote: > >> Can someone point me to an example build that is broken? >> >> On Mon, Mar 16, 2015 at 3:52 PM, Sean Busbey <bus...@cloudera.com> wrote: >> >> > I'm on it. HADOOP-11721 >> > >> > On Mon, Mar 16, 2015 at 3:44 PM, Haohui Mai <whe...@apache.org> wrote: >> > >> >> +1 for git clean. >> >> >> >> Colin, can you please get it in ASAP? Currently due to the jenkins >> >> issues, we cannot close the 2.7 blockers. >> >> >> >> Thanks, >> >> Haohui >> >> >> >> >> >> >> >> On Mon, Mar 16, 2015 at 11:54 AM, Colin P. McCabe <cmcc...@apache.org> >> >> wrote: >> >> > If all it takes is someone creating a test that makes a directory >> >> > without -x, this is going to happen over and over. >> >> > >> >> > Let's just fix the problem at the root by running "git clean -fqdx" >> in >> >> > our jenkins scripts. If there's no objections I will add this in and >> >> > un-break the builds. >> >> > >> >> > best, >> >> > Colin >> >> > >> >> > On Fri, Mar 13, 2015 at 1:48 PM, Lei Xu <l...@cloudera.com> wrote: >> >> >> I filed HDFS-7917 to change the way to simulate disk failures. >> >> >> >> >> >> But I think we still need infrastructure folks to help with jenkins >> >> >> scripts to clean the dirs left today. >> >> >> >> >> >> On Fri, Mar 13, 2015 at 1:38 PM, Mai Haohui <ricet...@gmail.com> >> >> wrote: >> >> >>> Any updates on this issues? It seems that all HDFS jenkins builds >> are >> >> >>> still failing. >> >> >>> >> >> >>> Regards, >> >> >>> Haohui >> >> >>> >> >> >>> On Thu, Mar 12, 2015 at 12:53 AM, Vinayakumar B < >> >> vinayakum...@apache.org> wrote: >> >> >>>> I think the problem started from here. >> >> >>>> >> >> >>>> >> >> >> https://builds.apache.org/job/PreCommit-HDFS-Build/9828/testReport/junit/org.apache.hadoop.hdfs.server.datanode/TestDataNodeVolumeFailure/testUnderReplicationAfterVolFailure/ >> >> >>>> >> >> >>>> As Chris mentioned TestDataNodeVolumeFailure is changing the >> >> permission. >> >> >>>> But in this patch, ReplicationMonitor got NPE and it got terminate >> >> signal, >> >> >>>> due to which MiniDFSCluster.shutdown() throwing Exception. >> >> >>>> >> >> >>>> But, TestDataNodeVolumeFailure#teardown() is restoring those >> >> permission >> >> >>>> after shutting down cluster. So in this case IMO, permissions were >> >> never >> >> >>>> restored. >> >> >>>> >> >> >>>> >> >> >>>> @After >> >> >>>> public void tearDown() throws Exception { >> >> >>>> if(data_fail != null) { >> >> >>>> FileUtil.setWritable(data_fail, true); >> >> >>>> } >> >> >>>> if(failedDir != null) { >> >> >>>> FileUtil.setWritable(failedDir, true); >> >> >>>> } >> >> >>>> if(cluster != null) { >> >> >>>> cluster.shutdown(); >> >> >>>> } >> >> >>>> for (int i = 0; i < 3; i++) { >> >> >>>> FileUtil.setExecutable(new File(dataDir, "data"+(2*i+1)), >> >> true); >> >> >>>> FileUtil.setExecutable(new File(dataDir, "data"+(2*i+2)), >> >> true); >> >> >>>> } >> >> >>>> } >> >> >>>> >> >> >>>> >> >> >>>> Regards, >> >> >>>> Vinay >> >> >>>> >> >> >>>> On Thu, Mar 12, 2015 at 12:35 PM, Vinayakumar B < >> >> vinayakum...@apache.org> >> >> >>>> wrote: >> >> >>>> >> >> >>>>> When I see the history of these kind of builds, All these are >> >> failed on >> >> >>>>> node H9. >> >> >>>>> >> >> >>>>> I think some or the other uncommitted patch would have created >> the >> >> problem >> >> >>>>> and left it there. >> >> >>>>> >> >> >>>>> >> >> >>>>> Regards, >> >> >>>>> Vinay >> >> >>>>> >> >> >>>>> On Thu, Mar 12, 2015 at 6:16 AM, Sean Busbey < >> bus...@cloudera.com> >> >> wrote: >> >> >>>>> >> >> >>>>>> You could rely on a destructive git clean call instead of maven >> to >> >> do the >> >> >>>>>> directory removal. >> >> >>>>>> >> >> >>>>>> -- >> >> >>>>>> Sean >> >> >>>>>> On Mar 11, 2015 4:11 PM, "Colin McCabe" <cmcc...@alumni.cmu.edu >> > >> >> wrote: >> >> >>>>>> >> >> >>>>>> > Is there a maven plugin or setting we can use to simply remove >> >> >>>>>> > directories that have no executable permissions on them? >> >> Clearly we >> >> >>>>>> > have the permission to do this from a technical point of view >> >> (since >> >> >>>>>> > we created the directories as the jenkins user), it's simply >> >> that the >> >> >>>>>> > code refuses to do it. >> >> >>>>>> > >> >> >>>>>> > Otherwise I guess we can just fix those tests... >> >> >>>>>> > >> >> >>>>>> > Colin >> >> >>>>>> > >> >> >>>>>> > On Tue, Mar 10, 2015 at 2:43 PM, Lei Xu <l...@cloudera.com> >> >> wrote: >> >> >>>>>> > > Thanks a lot for looking into HDFS-7722, Chris. >> >> >>>>>> > > >> >> >>>>>> > > In HDFS-7722: >> >> >>>>>> > > TestDataNodeVolumeFailureXXX tests reset data dir >> permissions >> >> in >> >> >>>>>> > TearDown(). >> >> >>>>>> > > TestDataNodeHotSwapVolumes reset permissions in a finally >> >> clause. >> >> >>>>>> > > >> >> >>>>>> > > Also I ran mvn test several times on my machine and all >> tests >> >> passed. >> >> >>>>>> > > >> >> >>>>>> > > However, since in DiskChecker#checkDirAccess(): >> >> >>>>>> > > >> >> >>>>>> > > private static void checkDirAccess(File dir) throws >> >> >>>>>> DiskErrorException { >> >> >>>>>> > > if (!dir.isDirectory()) { >> >> >>>>>> > > throw new DiskErrorException("Not a directory: " >> >> >>>>>> > > + dir.toString()); >> >> >>>>>> > > } >> >> >>>>>> > > >> >> >>>>>> > > checkAccessByFileMethods(dir); >> >> >>>>>> > > } >> >> >>>>>> > > >> >> >>>>>> > > One potentially safer alternative is replacing data dir >> with a >> >> regular >> >> >>>>>> > > file to stimulate disk failures. >> >> >>>>>> > > >> >> >>>>>> > > On Tue, Mar 10, 2015 at 2:19 PM, Chris Nauroth < >> >> >>>>>> cnaur...@hortonworks.com> >> >> >>>>>> > wrote: >> >> >>>>>> > >> TestDataNodeHotSwapVolumes, TestDataNodeVolumeFailure, >> >> >>>>>> > >> TestDataNodeVolumeFailureReporting, and >> >> >>>>>> > >> TestDataNodeVolumeFailureToleration all remove executable >> >> permissions >> >> >>>>>> > from >> >> >>>>>> > >> directories like the one Colin mentioned to simulate disk >> >> failures at >> >> >>>>>> > data >> >> >>>>>> > >> nodes. I reviewed the code for all of those, and they all >> >> appear to >> >> >>>>>> be >> >> >>>>>> > >> doing the necessary work to restore executable permissions >> at >> >> the >> >> >>>>>> end of >> >> >>>>>> > >> the test. The only recent uncommitted patch I¹ve seen that >> >> makes >> >> >>>>>> > changes >> >> >>>>>> > >> in these test suites is HDFS-7722. That patch still looks >> >> fine >> >> >>>>>> > though. I >> >> >>>>>> > >> don¹t know if there are other uncommitted patches that >> >> changed these >> >> >>>>>> > test >> >> >>>>>> > >> suites. >> >> >>>>>> > >> >> >> >>>>>> > >> I suppose it¹s also possible that the JUnit process >> >> unexpectedly died >> >> >>>>>> > >> after removing executable permissions but before restoring >> >> them. >> >> >>>>>> That >> >> >>>>>> > >> always would have been a weakness of these test suites, >> >> regardless of >> >> >>>>>> > any >> >> >>>>>> > >> recent changes. >> >> >>>>>> > >> >> >> >>>>>> > >> Chris Nauroth >> >> >>>>>> > >> Hortonworks >> >> >>>>>> > >> http://hortonworks.com/ >> >> >>>>>> > >> >> >> >>>>>> > >> >> >> >>>>>> > >> >> >> >>>>>> > >> >> >> >>>>>> > >> >> >> >>>>>> > >> >> >> >>>>>> > >> On 3/10/15, 1:47 PM, "Aaron T. Myers" <a...@cloudera.com> >> >> wrote: >> >> >>>>>> > >> >> >> >>>>>> > >>>Hey Colin, >> >> >>>>>> > >>> >> >> >>>>>> > >>>I asked Andrew Bayer, who works with Apache Infra, what's >> >> going on >> >> >>>>>> with >> >> >>>>>> > >>>these boxes. He took a look and concluded that some perms >> are >> >> being >> >> >>>>>> set >> >> >>>>>> > in >> >> >>>>>> > >>>those directories by our unit tests which are precluding >> >> those files >> >> >>>>>> > from >> >> >>>>>> > >>>getting deleted. He's going to clean up the boxes for us, >> but >> >> we >> >> >>>>>> should >> >> >>>>>> > >>>expect this to keep happening until we can fix the test in >> >> question >> >> >>>>>> to >> >> >>>>>> > >>>properly clean up after itself. >> >> >>>>>> > >>> >> >> >>>>>> > >>>To help narrow down which commit it was that started this, >> >> Andrew >> >> >>>>>> sent >> >> >>>>>> > me >> >> >>>>>> > >>>this info: >> >> >>>>>> > >>> >> >> >>>>>> > >>>"/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS- >> >> >>>>>> > >> >> >>>>>> >> >> >> >>>Build/hadoop-hdfs-project/hadoop-hdfs/target/test/data/dfs/data/data3/ >> >> >>>>>> > has >> >> >>>>>> > >>>500 perms, so I'm guessing that's the problem. Been that >> way >> >> since >> >> >>>>>> 9:32 >> >> >>>>>> > >>>UTC >> >> >>>>>> > >>>on March 5th." >> >> >>>>>> > >>> >> >> >>>>>> > >>>-- >> >> >>>>>> > >>>Aaron T. Myers >> >> >>>>>> > >>>Software Engineer, Cloudera >> >> >>>>>> > >>> >> >> >>>>>> > >>>On Tue, Mar 10, 2015 at 1:24 PM, Colin P. McCabe < >> >> cmcc...@apache.org >> >> >>>>>> > >> >> >>>>>> > >>>wrote: >> >> >>>>>> > >>> >> >> >>>>>> > >>>> Hi all, >> >> >>>>>> > >>>> >> >> >>>>>> > >>>> A very quick (and not thorough) survey shows that I can't >> >> find any >> >> >>>>>> > >>>> jenkins jobs that succeeded from the last 24 hours. Most >> >> of them >> >> >>>>>> seem >> >> >>>>>> > >>>> to be failing with some variant of this message: >> >> >>>>>> > >>>> >> >> >>>>>> > >>>> [ERROR] Failed to execute goal >> >> >>>>>> > >>>> org.apache.maven.plugins:maven-clean-plugin:2.5:clean >> >> >>>>>> (default-clean) >> >> >>>>>> > >>>> on project hadoop-hdfs: Failed to clean project: Failed >> to >> >> delete >> >> >>>>>> > >>>> >> >> >>>>>> > >>>> >> >> >>>>>> > >> >> >>>>>> > >> >> >>>>>> >> >> >> >>>>/home/jenkins/jenkins-slave/workspace/PreCommit-HDFS-Build/hadoop-hdfs-pr >> >> >>>>>> > >>>>oject/hadoop-hdfs/target/test/data/dfs/data/data3 >> >> >>>>>> > >>>> -> [Help 1] >> >> >>>>>> > >>>> >> >> >>>>>> > >>>> Any ideas how this happened? Bad disk, unit test setting >> >> wrong >> >> >>>>>> > >>>> permissions? >> >> >>>>>> > >>>> >> >> >>>>>> > >>>> Colin >> >> >>>>>> > >>>> >> >> >>>>>> > >> >> >> >>>>>> > > >> >> >>>>>> > > >> >> >>>>>> > > >> >> >>>>>> > > -- >> >> >>>>>> > > Lei (Eddy) Xu >> >> >>>>>> > > Software Engineer, Cloudera >> >> >>>>>> > >> >> >>>>>> >> >> >>>>> >> >> >>>>> >> >> >> >> >> >> >> >> >> >> >> >> -- >> >> >> Lei (Eddy) Xu >> >> >> Software Engineer, Cloudera >> >> >> > >> > >> > >> > -- >> > Sean >> > >> >> >> >> -- >> Sean >> > >