The doc looks good to me.

Ryan, the role of the shepherd is to make sure that someone
knowledgeable with Spark processes is involved: this person can advise
on technical and procedural considerations for people outside the
community. Also, if no one is willing to be a shepherd, the proposed
idea is probably not going to receive much traction in the first
place.

Tim

On Thu, Feb 16, 2017 at 9:17 AM, Cody Koeninger <c...@koeninger.org> wrote:
> Reynold, thanks, LGTM.
>
> Sean, great concerns.  I agree that behavior is largely cultural and
> writing down a process won't necessarily solve any problems one way or
> the other.  But one outwardly visible change I'm hoping for out of
> this a way for people who have a stake in Spark, but can't follow
> jiras closely, to go to the Spark website, see the list of proposed
> major changes, contribute discussion on issues that are relevant to
> their needs, and see a clear direction once a vote has passed.  We
> don't have that now.
>
> Ryan, realistically speaking any PMC member can and will stop any
> changes they don't like anyway, so might as well be up front about the
> reality of the situation.
>
> On Thu, Feb 16, 2017 at 10:43 AM, Sean Owen <so...@cloudera.com> wrote:
>> The text seems fine to me. Really, this is not describing a fundamentally
>> new process, which is good. We've always had JIRAs, we've always been able
>> to call a VOTE for a big question. This just writes down a sensible set of
>> guidelines for putting those two together when a major change is proposed. I
>> look forward to turning some big JIRAs into a request for a SPIP.
>>
>> My only hesitation is that this seems to be perceived by some as a new or
>> different thing, that is supposed to solve some problems that aren't
>> otherwise solvable. I see mentioned problems like: clear process for
>> managing work, public communication, more committers, some sort of binding
>> outcome and deadline.
>>
>> If SPIP is supposed to be a way to make people design in public and a way to
>> force attention to a particular change, then, this doesn't do that by
>> itself. Therefore I don't want to let a detailed discussion of SPIP detract
>> from the discussion about doing what SPIP implies. It's just a process
>> document.
>>
>> Still, a fine step IMHO.
>>
>> On Thu, Feb 16, 2017 at 4:22 PM Reynold Xin <r...@databricks.com> wrote:
>>>
>>> Updated. Any feedback from other community members?
>>>
>>>
>>> On Wed, Feb 15, 2017 at 2:53 AM, Cody Koeninger <c...@koeninger.org>
>>> wrote:
>>>>
>>>> Thanks for doing that.
>>>>
>>>> Given that there are at least 4 different Apache voting processes,
>>>> "typical Apache vote process" isn't meaningful to me.
>>>>
>>>> I think the intention is that in order to pass, it needs at least 3 +1
>>>> votes from PMC members *and no -1 votes from PMC members*.  But the 
>>>> document
>>>> doesn't explicitly say that second part.
>>>>
>>>> There's also no mention of the duration a vote should remain open.
>>>> There's a mention of a month for finding a shepherd, but that's different.
>>>>
>>>> Other than that, LGTM.
>>>>
>>>> On Mon, Feb 13, 2017 at 9:02 AM, Reynold Xin <r...@databricks.com> wrote:
>>>>>
>>>>> Here's a new draft that incorporated most of the feedback:
>>>>> https://docs.google.com/document/d/1-Zdi_W-wtuxS9hTK0P9qb2x-nRanvXmnZ7SUi4qMljg/edit#
>>>>>
>>>>> I added a specific role for SPIP Author and another one for SPIP
>>>>> Shepherd.
>>>>>
>>>>> On Sat, Feb 11, 2017 at 6:13 PM, Xiao Li <gatorsm...@gmail.com> wrote:
>>>>>>
>>>>>> During the summit, I also had a lot of discussions over similar topics
>>>>>> with multiple Committers and active users. I heard many fantastic ideas. 
>>>>>> I
>>>>>> believe Spark improvement proposals are good channels to collect the
>>>>>> requirements/designs.
>>>>>>
>>>>>>
>>>>>> IMO, we also need to consider the priority when working on these items.
>>>>>> Even if the proposal is accepted, it does not mean it will be implemented
>>>>>> and merged immediately. It is not a FIFO queue.
>>>>>>
>>>>>>
>>>>>> Even if some PRs are merged, sometimes, we still have to revert them
>>>>>> back, if the design and implementation are not reviewed carefully. We 
>>>>>> have
>>>>>> to ensure our quality. Spark is not an application software. It is an
>>>>>> infrastructure software that is being used by many many companies. We 
>>>>>> have
>>>>>> to be very careful in the design and implementation, especially
>>>>>> adding/changing the external APIs.
>>>>>>
>>>>>>
>>>>>> When I developed the Mainframe infrastructure/middleware software in
>>>>>> the past 6 years, I were involved in the discussions with 
>>>>>> external/internal
>>>>>> customers. The to-do feature list was always above 100. Sometimes, the
>>>>>> customers are feeling frustrated when we are unable to deliver them on 
>>>>>> time
>>>>>> due to the resource limits and others. Even if they paid us billions, we
>>>>>> still need to do it phase by phase or sometimes they have to accept the
>>>>>> workarounds. That is the reality everyone has to face, I think.
>>>>>>
>>>>>>
>>>>>> Thanks,
>>>>>>
>>>>>>
>>>>>> Xiao Li
>>>>>>>
>>>>>>>
>>
>
> ---------------------------------------------------------------------
> To unsubscribe e-mail: dev-unsubscr...@spark.apache.org
>

---------------------------------------------------------------------
To unsubscribe e-mail: dev-unsubscr...@spark.apache.org

Reply via email to