On Dec 6, 2010, at 10:40 AM, Chris Douglas wrote:

This question is backwards. If the assertion is that a part of the
framework's development should be arrested, that claim requires a
discussion and vote. The PMC should not have to weigh in on allowing
code to change. -C


Agreed.

Arresting development on SequenceFile is preposterous. There are several petabytes of data sitting on it all over for several reasons, including legacy. Stopping development on it is unreasonable. Apache Hadoop is volunteer driven, volunteers should be allowed to contribute as they see fit.

+1

Arun

On Mon, Dec 6, 2010 at 9:16 AM, Owen O'Malley <[email protected]> wrote:

On Dec 1, 2010, at 11:11 AM, Owen O'Malley wrote:

All,
We really need some guidance on the general direction for the project. Please comment and/or vote. If no one cares, then I'll probably commit it to
Yahoo's internal branch.

-- Owen

The question is how the Hadoop project wants to move forward.

It was motivated by Doug's veto of HADOOP-6685, which was based on his personal decisions about how the project should go forward and not on
anything that had been decided by the PMC.

These decisions are much more important to MapReduce, which is a
framework, than HDFS which is a client/server model.

1. Should Hadoop include a user-facing library of useful code?

There has been a suggestion that user-facing library code, such as
SequenceFile, TFile, DistCp, etc. should be deprecated and that Hadoop should allow third party projects like Avro to supply the user- facing library code that makes Hadoop usable. I think it is critical that we keep those components as part of Hadoop and extend them as the framework evolves. Users depend heavily on SequenceFile for storing their data in Hadoop and
they should not  be deprecated as Doug has suggested.

2. Should MapReduce support non-Writables through the pipeline out of the
box?

There has also been a discussion about whether we should support
non-Writables natively. There is already library code in Avro that lets users use Avro types in a custom MapReduce API. A general MapReduce API that encompasses all of the serialization frameworks and does not lock users into
a particular one is much more powerful.

Furthermore, making it convenient for the users, by including the plugins in the default configuration and class path, will enable the use of Avro,
Thrift and ProtoBuf objects by people who would rather not focus on
serialization. Avro and Writables should not be the only first class
serializations that Hadoop supports by default.

3. Should a framework dependency on ProtoBuf be allowed?

Doug has added several framework dependences on Avro. The question is whether it is acceptable to use the ProtoBuf library in the framework. Avro is good for uses where there are a lot of objects of the same type. ProtoBuf is better for small number of objects. The question is whether Avro, JSON, and XML should be the only serialization libraries that are acceptable to
use in the framework.



Reply via email to