Thanks to Matt for his philosophical take. I agree.

The intent is to set a common goal, so that we work toward getting v2 in a
usable state as a community. Part of that is making choices to get it done
on time, which we have already seen on this thread: setting out more
clearly what we mean by “DSv2” and what we think we can get done on time.

I don’t mean to say that we should commit to a plan that *requires* a delay
to the next release (which describes the goal better than 3.0 does). But we
should commit to making sure the goal is met, acknowledging that this is
one of the most important efforts for many people that work in this
community.

I think it would help to clarify what this commitment means, at least to me:

   1. What it means: the community will seriously consider delaying the
   next release if this isn’t done by our initial deadline.
   2. What it does not mean: delaying the release no matter what happens.

In that event that this feature isn’t done on time, it would be up to the
community to decide what to do. But in the mean time, I think it is healthy
to set a goal and work toward it. (I am not making a distinction between
PMC and community here.)

I think this commitment is a good idea for the same reason why we set other
goals: to hold ourselves accountable. When one sets a New Years resolution
to drop 10 pounds, it isn’t that the hope or intent wasn’t there before. It
is about having a (self-imposed) constraint that helps you make hard
choices: cake now or meet my goal?

Spark 3.0 has many other major features as well, delaying the release has
significant cost and we should try our best to not let it happen.”

I agree with Wenchen here. No one wants to actually delay the release. We
just want to push ourselves to make some tough decisions, using that delay
as a motivating factor.

The fact that some entity other than the PMC thinks that Spark 3.0 should
contain certain new features or that it will be costly to them if 3.0 does
not contain those features is not dispositive.

I’m not quite sure what you mean here. While I am representing my employer,
I am bringing up this topic as a member of the community, to suggest a
direction for the community to take, and I fully accept that the decision
is up to the community. I think it is reasonable to candidly state how this
matters; that context informs the discussion.

On Fri, Feb 22, 2019 at 1:55 PM Mark Hamstra <m...@clearstorydata.com>
wrote:

> To your other message: I already see a number of PMC members here. Who's
>> the other entity?
>>
>
> I'll answer indirectly since pointing fingers isn't really my intent. In
> the absence of a PMC vote, I react negatively to individuals making new
> declarative policy statements or statements to the effect that Spark
> 3.0 will (or will not) include these features..., or that it will be too
> costly to do something. Maybe these are innocent shorthand that leave off a
> clarifying "in my opinion" or "according to the current state of JIRA" or
> some such.
>
> My points are simply that nobody other than the PMC has an authoritative
> say on such matters, and if we are at a point where the community needs
> some definitive guidance, then we need PMC involvement and a vote. That's
> not intended to preclude or terminate community discussion, because that
> is, indeed, lovely to see.
>
> On Fri, Feb 22, 2019 at 12:04 PM Sean Owen <sro...@apache.org> wrote:
>
>> To your other message: I already see a number of PMC members here. Who's
>> the other entity? The PMC is the thing that says a thing is a release,
>> sure, but this discussion is properly a community one. And here we are,
>> this is lovely to see.
>>
>> (May I remind everyone to casually, sometime, browse the large list of
>> other JIRAs targeted for Spark 3? it's much more than DSv2!)
>>
>> I can't speak to specific decisions here, but, I see:
>>
>> Spark 3 doesn't have a release date. Notionally it's 6 months after Spark
>> 2.4 (Nov 2018). It'd be reasonable to plan for a little more time. Can we
>> throw out... June 2019, and I update the website? It can slip but that
>> gives a concrete timeframe around which to plan. What can comfortably get
>> in by June 2019?
>>
>> Agreement that "DSv2" is going into Spark 3, for some definition of DSv2
>> that's probably roughly Matt's list.
>>
>> Changes that can't go into a minor release (API changes, etc) must by
>> definition go into Spark 3.0. Agree those first and do those now. Delay
>> Spark 3 until they're done and prioritize accordingly.
>> Changes that can go into a minor release can go into 3.1, if needed.
>> This has been in discussion long enough that I think whatever design(s)
>> are on the table for DSv2 now are as close as one is going to get. The
>> perfect is the enemy of the good.
>>
>> Aside from throwing out a date, I probably just restated what everyone
>> said. But I was 'summoned' :)
>>
>> On Fri, Feb 22, 2019 at 12:40 PM Mark Hamstra <m...@clearstorydata.com>
>> wrote:
>>
>>> However, as other people mentioned, Spark 3.0 has many other major
>>>> features as well
>>>>
>>>
>>> I fundamentally disagree. First, Spark 3.0 has nothing until the PMC
>>> says it has something, and we have made no commitment along the lines that
>>> "Spark 3.0.0 will not be released unless it contains new features x, y and
>>> z." Second, major-version releases are not about adding new features.
>>> Major-version releases are about making changes to the public API that we
>>> cannot make in feature or bug-fix releases. If that is all that is
>>> accomplished in a particular major release, that's fine -- in fact, we
>>> quite intentionally did not target new features in the Spark 2.0.0 release.
>>> The fact that some entity other than the PMC thinks that Spark 3.0 should
>>> contain certain new features or that it will be costly to them if 3.0 does
>>> not contain those features is not dispositive. If there are public API
>>> changes that should occur in a timely fashion and there is also a list of
>>> new features that some users or contributors want to see in 3.0 but that
>>> look likely to not be ready in a timely fashion, then the PMC should fully
>>> consider releasing 3.0 without all those new features. There is no reason
>>> that they can't come in with 3.1.0.
>>>
>>

-- 
Ryan Blue
Software Engineer
Netflix

Reply via email to