As for GH for code review, I find that it works very well for nits. It’s also 
great for doc changes, given how GH allows you suggest changes to files 
in-place and automatically create PRs for those changes. That lowers the 
barrier for those tiny contributions.

For anything relatively substantial, I vastly prefer to summarise my feedback 
(and see others’ feedback summarised) in JIRA comments - an opinion I and other 
contributors have shared in one or two similar threads over the years.


> On 24 Jan 2020, at 12:21, Aleksey Yeshchenko <alek...@apple.com.INVALID> 
> wrote:
> 
> The person introducing flakiness to a test will almost always have run it 
> locally and on CI first with success. It’s usually later when they first 
> start failing, and it’s often tricky to attribute to a particular 
> commit/person.
> 
> So long as we have these - and we’ve had flaky tests for as long as C* has 
> existed - the problem will persist, and gating PRs on clean runs won’t 
> achieve anything other than dealing with folks who straight up ignore the 
> spirit of the policy and knowingly commit code with test breakage that can be 
> attributed to their change. I’m not aware of such committers in this 
> community, however.
> 
>> On 24 Jan 2020, at 09:01, Benedict Elliott Smith <bened...@apache.org> wrote:
>> 
>>> I find it only useful for nits, or for coaching-level comments that I would 
>>> never want propagated to Jira.
>> 
>> Actually, I'll go one step further. GitHub encourages comments that are too 
>> trivial, poisoning the well for third parties trying to find useful 
>> information.  If the comment wouldn't be made in Jira, it probably shouldn't 
>> be made. 
>> 
>> 
>> 
>> On 24/01/2020, 08:56, "Benedict Elliott Smith" <bened...@apache.org> wrote:
>> 
>>   The common factor is flaky tests, not people.  You get a clean run, you 
>> commit.  Turns out, a test was flaky.  This reduces trust in CI, so people 
>> commit without looking as closely at results.  Gating on clean tests doesn't 
>> help, as you run until you're clean.  Rinse and repeat.  Breakages 
>> accumulate.  
>> 
>>   This is what happens leading up to every release - nobody commits knowing 
>> there's a breakage.  We have a problem with bad tests, not bad people or 
>> process.
>> 
>>   FWIW, I no longer like the GitHub workflow.  I find it only useful for 
>> nits, or for coaching-level comments that I would never want propagated to 
>> Jira.  I find a strong patch submission of any size is better managed with 
>> human-curated Jira comments, and provides a better permanent record.  When 
>> skimming a discussion, Jira is more informative than GitHub.  Even with the 
>> GitHub UX, the context hinders rather than helps.
>> 
>>   As to propagating to Jira: has anyone here ever read them?  I haven't as 
>> they're impenetrable; ugly and almost entirely noise.  If anything, I would 
>> prefer that we discourage GitHub for review as a project, not move towards 
>> it.
>> 
>>   This is without getting into the problem of multiple branch PRs.  Until 
>> this is _provably_ painless, we cannot introduce a workflow that requires it 
>> and blocks commit on it.  Working with multiple branches is difficult enough 
>> already, surely?
>> 
>> 
>> 
>>   On 24/01/2020, 03:16, "Jeff Jirsa" <jji...@gmail.com> wrote:
>> 
>>       100% agree
>> 
>>       François and team wrote a doc on testing and gating commits
>>       Blake wrote a doc on testing and gating commits
>>       Every release there’s a thread on testing and gating commits
>> 
>>       People are the common factor every time. Nobody wants to avoid merging 
>> their patch because someone broke a test elsewhere. 
>> 
>>       We can’t block merging technically with the repo as it exists now so 
>> it’s always going to come down to people and peer pressure, until or unless 
>> someone starts reverting commits that break tests
>> 
>>       (Of course, someone could write a tool that automatically reverts new 
>> commits as long as tests fail....)
>> 
>>       On Jan 23, 2020, at 5:54 PM, Joshua McKenzie <jmcken...@apache.org> 
>> wrote:
>>> 
>>> 
>>>> 
>>>> 
>>>> I am reacting to what I currently see
>>>> happening in the project; tests fail as the norm and this is kinda seen as
>>>> expected, even though it goes against the policies as I understand it.
>>> 
>>> After over half a decade seeing us all continue to struggle with this
>>> problem, I've come around to the school of "apply pain" (I mean that as
>>> light-hearted as you can take it) when there's a failure to incent fixing;
>>> specifically in this case, the only idea I can think of is preventing merge
>>> w/any failing tests on a PR. We go through this cycle as we approach each
>>> major release: we have the gatekeeper of "we're not going to cut a release
>>> with failing tests obviously", and we clean them up. After the release, the
>>> pressure is off, we exhale, relax, and flaky test failures (and others)
>>> start to creep back in.
>>> 
>>> If the status quo is the world we want to live in, that's totally fine and
>>> no judgement intended - we can build tooling around test failure history
>>> and known flaky tests etc to optimize engineer workflows around that
>>> expectation. But what I keep seeing on threads like this (and have always
>>> heard brought up in conversation) is that our collective *moral* stance is
>>> that we should have green test boards and not merge code if it introduces
>>> failing tests.
>>> 
>>> Not looking to prescribe or recommend anything, just hoping that
>>> observation above might be of interest or value to the conversation.
>>> 
>>>> On Thu, Jan 23, 2020 at 4:17 PM Michael Shuler <mich...@pbandjelly.org>
>>>> wrote:
>>>> 
>>>>> On 1/23/20 3:53 PM, David Capwell wrote:
>>>>> 
>>>>> 2) Nightly build email to dev@?
>>>> 
>>>> Nope. builds@c.a.o is where these go.
>>>> https://lists.apache.org/list.html?bui...@cassandra.apache.org
>>>> 
>>>> Michael
>>>> 
>>>> ---------------------------------------------------------------------
>>>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
>>>> For additional commands, e-mail: dev-h...@cassandra.apache.org
>>>> 
>>>> 
>> 
>>       ---------------------------------------------------------------------
>>       To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
>>       For additional commands, e-mail: dev-h...@cassandra.apache.org
>> 
>> 
>> 
>> 
>> 
>>   ---------------------------------------------------------------------
>>   To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
>>   For additional commands, e-mail: dev-h...@cassandra.apache.org
>> 
>> 
>> 
>> 
>> 
>> ---------------------------------------------------------------------
>> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
>> For additional commands, e-mail: dev-h...@cassandra.apache.org
>> 
> 
> 
> ---------------------------------------------------------------------
> To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
> For additional commands, e-mail: dev-h...@cassandra.apache.org
> 


---------------------------------------------------------------------
To unsubscribe, e-mail: dev-unsubscr...@cassandra.apache.org
For additional commands, e-mail: dev-h...@cassandra.apache.org

Reply via email to