+1 to what Owen said, I don't think disabling *StressNewTest* is a good idea.
On Fri, 28 Feb 2020 at 18:35, Owen Nichols <onich...@pivotal.io> wrote: > -1 to making StressNew not required > > +1 to eliminating the current loophole — StressNew should never give a > free pass. > > Any time your PR is having trouble passing StressNew, please bring it up > on the dev list. We can review on a case-by-case basis and decide whether > to try increasing the timeout, changing the repeat count, refactoring the > PR, or as an absolute last resort requesting authorization for an override > (for example, a change to spotless rules might touch a huge number of files > but carry no risk). > > One bug we should fix is that StressNew sometimes counts more files > touched than really were, especially if you had many commits or merges or > rebases on your PR branch. Possible workarounds there include squashing > and/or creating a new PR and/or splitting into multiple PRs. I’ve spent > some time trying to reproduce why files are mis-counted, with no success, > but perhaps someone cleverer with git could provide a fix. > > Another issue is that StressNew is only in the PR pipeline, not the main > develop pipeline. This feels like an asymmetry where PRs must pass a > “higher” standard. We should consider adding some form of StressNew to the > main pipeline as well (maybe compare to the previous SHA that passed?). > > The original motivation for the 25-file limit was an attempt to limit how > long StressNew might run for. Since concourse already applies a timeout, > that check is unnecessary. However, a compromise solution might be to use > the number of files changed to dial back the number of repeats, e.g. stay > with 50 repeats if fewer than 25 files changed, otherwise compute 1250 / > <#-files-changed> and do only that many repeats (e.g. if 50 files changed, > run all tests 25x instead of 50x). > > While StressNew is intended to catch new flaky tests, it can also catch > poorly-designed tests that fail just by running twice in the same VM. This > may be a sign that the test does not clean up properly and could be > polluting other tests in unexpected ways? It might be useful to run a > “StressNew” with repeats=2 over a much broader scope, maybe even all tests, > at least once in a while? > > > > On Feb 28, 2020, at 9:51 AM, Mark Hanson <mhan...@pivotal.io> wrote: > > > > Hi All, > > > > Proposal: Force StressNewTest to fail a change with 25 or more files > rather than pass it without running it. > > > > Currently, the StressNewTest job in the pipeline will just pass a job > that has more than 25 files changed. It will be marked as green with no > work done. There are reasons, relating to run time being too long to be > tracked by concourse, so we just let it through as a pass. I think this is > a bad signal. I think that this should automatically be a failure in the > short term. As a result, I also think it should not be required. It is a > bad signal, and I think that by making it a fail, this will at least not > give us a false sense of security. I understand that this opens the flood > gates so to speak, but I don’t think as a community it is a big problem > because we can still say that you should not merge if the StressNewTest > fails because of your test. > > > > I personally find the false sense of security more troubling than > anything. Hence the reason, I would like this to be > > > > Let me know what you think.. > > > > Thanks, > > Mark > > -- Ju@N