This is really good feedback, Nick. I agree, we need them to be reliable
enough to not be a source of constant false positives prior to putting them
into the checklist.
On Thu, Oct 4, 2018 at 15:34 Nick Allen <n...@nickallen.org> wrote:

> I think we still have an issue of reliability.  I can never reliably get
> them all to pass.  I have no idea which failures are real.  Am I the only
> one that experiences this?
>
> We need a reliable pass/fail on these before we talk about adding them to
> the checklist.  For example, I just tried to run them on METRON-1771.  I
> don't think we have a problem with these changes, but I have not been able
> to get one run to fully pass.  See the attached output of those runs.
>
>
>
> On Wed, Oct 3, 2018 at 7:36 AM Shane Ardell <shane.m.ard...@gmail.com>
> wrote:
>
>> I ran them locally a handful of times just now, and on average they took
>> approximately 15 minutes to complete.
>>
>> On Tue, Oct 2, 2018, 18:22 Michael Miklavcic <michael.miklav...@gmail.com
>> >
>> wrote:
>>
>> > @Shane Just how much time are we talking about, on average? I don't
>> think
>> > many in the community have had much exposure to running the e2e tests in
>> > their current form. It might still be worth it in the short term.
>> >
>> > On Tue, Oct 2, 2018 at 10:20 AM Shane Ardell <shane.m.ard...@gmail.com>
>> > wrote:
>> >
>> > > The protractor-flake package should catch and re-run false failures,
>> so
>> > > people shouldn't get failing tests when they are done running. I just
>> > meant
>> > > that we often re-run flaky tests with protractor-flake, so it can
>> take a
>> > > while to run and could increase the build time considerably.
>> > >
>> > > On Tue, Oct 2, 2018, 18:00 Casey Stella <ceste...@gmail.com> wrote:
>> > >
>> > > > Are the tests so brittle that, even with flaky, people will run upon
>> > > false
>> > > > failures as part of contributing a PR?  If so, do we have a list of
>> the
>> > > > brittle ones (and the things that would disambiguate a true failure
>> > from
>> > > a
>> > > > false failure) that we can add to the documentation?
>> > > >
>> > > > On Tue, Oct 2, 2018 at 11:58 AM Shane Ardell <
>> shane.m.ard...@gmail.com
>> > >
>> > > > wrote:
>> > > >
>> > > > > I also would like to eventually have these tests automated. There
>> > are a
>> > > > > couple hurdles to setting up our e2e tests to run with our build.
>> I
>> > > think
>> > > > > the biggest hurdle is setting up a dedicated server with data for
>> the
>> > > e2e
>> > > > > tests to use. I would assume this requires funding, engineering
>> > > support,
>> > > > > obfuscated data, etc. I also think we should migrate our e2e
>> tests to
>> > > > > Cypress first because Protractor lacks debugging tools that would
>> > make
>> > > > our
>> > > > > life much easier if, for example, we had a failure in our CI build
>> > but
>> > > > > could not reproduce locally. In addition, our current Protractor
>> > tests
>> > > > are
>> > > > > brittle and extremely slow.
>> > > > >
>> > > > > All that said, it seems we agree that we could add another PR
>> > checklist
>> > > > > item in the meantime. Clarifying those e2e test instructions
>> should
>> > be
>> > > > part
>> > > > > of that task.
>> > > > >
>> > > > > On Mon, Oct 1, 2018 at 2:36 PM Casey Stella <ceste...@gmail.com>
>> > > wrote:
>> > > > >
>> > > > > > I'd also like to make sure that clear instructions are provided
>> (or
>> > > > > linked
>> > > > > > to) about how to run them.  Also, we need to make sure the
>> > > instructions
>> > > > > are
>> > > > > > rock-solid for running them.
>> > > > > > Looking at
>> > > > > >
>> > > > > >
>> > > > >
>> > > >
>> > >
>> >
>> https://github.com/apache/metron/tree/master/metron-interface/metron-alerts#e2e-tests
>> > > > > > ,
>> > > > > > would someone who doesn't have much or any knowledge of the UI
>> be
>> > > able
>> > > > to
>> > > > > > run that without assistance?
>> > > > > >
>> > > > > > For instance, we use full-dev, do we need to stop data from
>> being
>> > > > played
>> > > > > > into full-dev for the tests to work?
>> > > > > >
>> > > > > > Casey
>> > > > > >
>> > > > > > On Mon, Oct 1, 2018 at 8:29 AM Casey Stella <ceste...@gmail.com
>> >
>> > > > wrote:
>> > > > > >
>> > > > > > > I'm not super keen on expanding the steps to contribute,
>> > especially
>> > > > in
>> > > > > an
>> > > > > > > avenue that should be automated.
>> > > > > > > That being said, I think that until we get to the point of
>> > > automating
>> > > > > the
>> > > > > > > e2e tests, it's sensible to add them to the checklist.
>> > > > > > > So, I would support it, but I would also urge us to move
>> forward
>> > > the
>> > > > > > > efforts of running these tests as part of the CI build.
>> > > > > > >
>> > > > > > > What is the current gap there?
>> > > > > > >
>> > > > > > > Casey
>> > > > > > >
>> > > > > > > On Mon, Oct 1, 2018 at 7:41 AM Shane Ardell <
>> > > > shane.m.ard...@gmail.com>
>> > > > > > > wrote:
>> > > > > > >
>> > > > > > >> Hello everyone,
>> > > > > > >>
>> > > > > > >> In another discussion thread from July, I briefly mentioned
>> the
>> > > idea
>> > > > > of
>> > > > > > >> adding a step to the pull request checklist asking
>> contributors
>> > to
>> > > > run
>> > > > > > the
>> > > > > > >> UI end-to-end tests. Since we aren't running e2e tests as
>> part
>> > of
>> > > > the
>> > > > > CI
>> > > > > > >> build, it's easy for contributors to unintentionally break
>> these
>> > > > > tests.
>> > > > > > >> Reminding contributors to run these tests will hopefully help
>> > > catch
>> > > > > > >> situations like this before opening a pull request.
>> > > > > > >>
>> > > > > > >> Does this make sense to everyone?
>> > > > > > >>
>> > > > > > >> Regards,
>> > > > > > >> Shane
>> > > > > > >>
>> > > > > > >
>> > > > > >
>> > > > >
>> > > >
>> > >
>> >
>>
>

Reply via email to