I want to throw in the following discussion topic.

We are clearly having integration issues when putting together
bleeding edge Hadoop based stacks. I don't want to argue about the
roots of the phenomena not about how it can be fixed. The purpose of
this discussion is different. 

There is an ongoing discussion happening at
  http://s.apache.org/hadoop-stablization

and http://is.gd/HX5JA2

I think this is the time for community to start doing something and
improving the situation, helping our users in the passing. I am sure
the effort outlined below would beneficial for individuals as well as
commercial vendors alike.

Now, think Ubuntu - arguably the _most_ successful community driven
Linux distribution in the world. There are two trains of the releases
happening all the time: STS and LTS. The former have about 6 months of
support in the form of updates and is used as a technology preview for
the next LTS release. The latter is good for 3 years and normally is a
way more stable then STS releases.

Currently, BigTop is focusing on the latest (and apparently not
greatest) of its component releases. While sometimes it gives certain
guarantees that A of X.y version will work with B of W.v version, it
isn't much of reconciliation to the users and vendors who simply want
to have a stable reference implementation of a Hadoop stack that has
been defined and supported by the community, with clear path for the
updates and testing put out in the open.

Because of the nature of the differences between Hadoop 1 and Hadoop 2
we can even try to run two different LTS for both kinds of stack.

I believe there's not much more to say about it, except that this is,
in my opinion, a good way to establish our project as de-facto go-to
place for community driven Hadoop based stacks and a focal point for
the integration in the ASF storage and analytics projects.

Let's the discussion begin.
  Cos

Attachment: signature.asc
Description: Digital signature

Reply via email to