I agree that getting the release process restarted is of utmost importance to the project. To help make that happen I'm happy to volunteer to be a release manager for the next release. This will be the first release post-split, so there will undoubtedly be some issues to work out. I think the focus should be on getting an alpha release out, so I suggest we create a new 0.21 branch from trunk, then spend time fixing blockers (which will be a superset of the existing 0.21 blockers).
Cheers, Tom On Wed, Mar 24, 2010 at 1:27 PM, Brian Bockelman <bbock...@cse.unl.edu> wrote: > Hey Allen, > > Your post provoked a few thoughts: > 1) Hadoop is a large, but relatively immature project (as in, there's still a > lot of major features coming down the pipe). If we wait to release on > features, especially when there are critical bugs, we end up with a large > number of patches between releases. This ends up encouraging custom patch > sets and custom distributions. > 2) The barrier for patch acceptance is high, especially for opportunistic > developers. This is a good thing for code quality, but for getting patches > in a timely manner. This means that there are a lot of 'mostly good' patches > out there in JIRA which have not landed. This again encourages folks to > develop their own custom patch sets. > 3) We make only bugfixes for past minor releases, meaning the stable Apache > release is perpetually behind in features, even features that are not core. > > Not sure how to best fix these things. One possibility: > a) Have a stable/unstable series (0.19.x is unstable, 0.20.x is stable, > 0.21.x is unstable). For the unstable releases, lower the bar for code > acceptance for less-risky patches. > b) Combined with a a time-based release for bugfixes (and non-dangerous > features?) in order to keep the feature releases "fresh". > > (a) aims to tackle problems (1) and (2). (b) aims to tackle (3). > > This might not work for everything. If I had a goal, it would be to decrease > the number of active distributions from 3 to 2 - otherwise you end up > spending far too much time consensus building. > > Just a thought from an outside, relatively content observer, > > Brian > > On Mar 24, 2010, at 1:38 PM, Allen Wittenauer wrote: > >> On 3/15/10 9:06 AM, "Owen O'Malley" <o...@yahoo-inc.com> wrote: >>> From our 21 experience, it looks like our old release strategy is >>> failing. >> >> Maybe this is a dumb question but... Are we sure it isn't the community >> failing? >> >> From where I stand, the major committers (PMC?) have essentially forked >> Hadoop into three competing source trees. No one appears to be dedicated to >> helping the community release because the focus is on their own tree. Worse >> yet, two of these trees are publicly available with both sides pushing their >> own tree as vastly superior (against each other and against the official >> Apache branded one). >> >> What are the next steps in getting this resolved? Is >> Hadoop-as-we-know-it essentially dead? What is going to prevent the fiasco >> that is 0.21 from impacting 0.22? >> >> For me personally, I'm more amused than upset that 0.21 hasn't been >> released. But I'm less happy that there appears to be a focus on feature >> additions rather than getting some of the 0.21 blockers settled (I'm >> assuming here that most of the 0.21 blockers apply to 0.22 as well). >> >> I don't think retroactively declaring 0.20 as 1.0 is going to make the >> situation any better. [In fact, I believe it will make it worse, since it >> gives an external impression that 0.20 is somehow stable at all levels. We >> all know this isn't true.] > >