On Tue, 06 Mar 2018 15:27:06 +0000 Stephen Houston <smhousto...@gmail.com> said:

> We have developers leaving or severely cutting back their work, and this
> includes developers who carry a large work load.  Now we see Stefan has
> lost faith and interest in QA/CI and is going to step back from that... I
> think at some point we need to agree to stop arguing the merits of getting
> better structure and better organization and agree that SOMETHING needs to
> be done and start putting together a plan.  So I think indefini put
> together a phab ticket https://phab.enlightenment.org/T6740 we should
> really work there to put together a plan to help.

not sure why you said the above... that exact ticket is entirely unrelated to
what stefan is talking about. in fact they disagree.

> On Tue, Mar 6, 2018 at 8:46 AM Carsten Haitzler <ras...@rasterman.com>
> wrote:
> 
> > On Tue, 6 Mar 2018 13:43:09 +0100 Stefan Schmidt <ste...@osg.samsung.com>
> > said:
> >
> > > Hello.
> > >
> > >
> > > On 03/06/2018 12:34 PM, Carsten Haitzler (The Rasterman) wrote:
> > > > On Tue, 6 Mar 2018 09:54:50 +0100 Stefan Schmidt <
> > ste...@osg.samsung.com>
> > > > said:
> > > >
> > > >> Hello.
> > > >>
> > > >>
> > > >> On 03/06/2018 07:44 AM, Carsten Haitzler (The Rasterman) wrote:
> > > >>> tbh platforms is an issue. windows is kind of big as setting up a
> > > >>> cross-build environment is a fair bit of work. setting up a windows
> > vm
> > > >>> costs money to test (need licenses, or if you can scrounge one off
> > another
> > > >>> pc you have/had). osx is worse in that you have to go buy hardware
> > to run
> > > >>> osx to test.
> > > >>>
> > > >>> i think making it a requirement that every commit work on every
> > platform
> > > >>> is not viable UNLESS:
> > > >>>
> > > >>> 1. a windows cross-compile env is available to developers (e.g. ssh
> > into a
> > > >>> vm set up for this with clear documentation and scripts/tools to do
> > the
> > > >>> builds. i have one set up for me at home).
> > > >>> 2. a vm with remote display that developers can use to run/test
> > changes
> > > >>> easily.
> > > >>> 3. an actual osx box that developers can ssh into and compile and
> > runa nd
> > > >>> test and remotely view/interact with like the windows vm.
> > > >>> 4. same for freebsd etc.
> > > >> We do not have this and I am pretty sure we never will (I can only
> > hope the
> > > >> future proofs me wrong). Maybe we should be more honest and state
> > that any
> > > > then i can't see how we could make it a requirement for committing
> > that they
> > > > build/run there.
> > >
> > > Having them running a build is very different from having full remote
> > shell
> > > access. E.g. osx build slaves available on TravisCi do not have any
> > option to
> > > get shell access.
> > >
> > > Detecting a problem is the first step. A build test for these *before*
> > > entering master would do this. It would still need fixing, but that is
> > the
> > > same we have right now. The difference really is where we detect the
> > break
> > > and if we have it sitting in master or not.
> >
> > ok - i wasn't thinking github's travis but more an actual osx machine
> > plugged
> > in somewhere we can remotely manage and use. :) i know travis will be very
> > limited.
> >
> > > > when people report issues with specific builds/platforms then
> > > > address them as needed.
> > >
> > > I can report the build being broken on osx, on aarch64, with the debug
> > > profile enabled and also make check with cxx enabled right now. That are
> > 4
> > > things broken in master just what i see today.
> >
> > arm64 - known. and this is a bit of a surprise from luajit. it's
> > technically
> > been broken like this forever. it wasn't some change we made. somehow
> > luajit
> > started breaking at some point. the "we only allow you to use 27 bits out
> > of 64
> > of a pointer" thing.
> >
> > so this is not something CI would magically stop as changes to the
> > dependencies
> > started making things crash there. at least i have personally compiled
> > using a
> > luajit compiled myself on arm64 and it worked. this was a while back.
> >
> > cxx bindings. indeed these are a problem to the point where i just disable
> > them
> > in my build. i am not sure if this is good or bad though. it's kind of
> > tired
> > with the eo and bindings work and as things change multiple languages need
> > to
> > adapt.
> >
> > as for osx - i have no info as to what is going on there or why. :( i don't
> > know even where the build breaks/logs etc. are. i do know
> > build.enlightenment.org and that basically seems ok.
> >
> > > >> platform we support (besides Linux on x86_64) has only been tested at
> > some
> > > >> point in the past.
> > > > i actually cross-build for windows maybe every few weeks or so, and
> > freebsd
> > > > maybe similarly too. i build on rpi3 (32bit) too regularly enough.
> > > >
> > > > we haven't tested on every linux distro either... so should we only
> > claim a
> > > > few distributions? i don't think we're being dishonest really.
> > releases do
> > > > get a lot more testing to see they work across os's. master may not
> > get as
> > > > much until a release cycle.
> > >
> > > Yeah, and as I had the pleasure of handling these releases I can tell you
> > > that dealing with them so late in the cycle is a nightmare. Having a
> > smoother
> > > release cycle is actually one of my main motivations to have this early
> > > detection and avoidance of breaks.
> >
> > indeed you are right - catching things early is far better. i really was
> > more
> > about us being honest about the releases working on those platforms or
> > not. :)
> > that's all. you are right there and i totally agree with that.
> >
> > > >>> if a platform is on EASILY accessible and able to be EASILY worked
> > with,
> > > >>> then making this a requirement to pass a build/test on that platform
> > is
> > > >>> silly.
> > > >>>
> > > >>> developers have to be able to do incremental builds. not a "wait 10
> > mins
> > > >>> for the next build cycle to happen then wait another 10 for the log
> > to
> > > >>> see if it worked this time". that's fine for reports. it's not fine
> > for
> > > >>> actual development.
> > > >>>
> > > >>> if this infra existed and worked well, THEN i think it might be sane
> > to
> > > >>> start adding "it has to pass a build test". then deciding on what
> > > >>> platforms have to be supported is another step. this has to be
> > pain-free
> > > >>> or as close as possible to that.
> > > >> Good luck to finding somehow setting this all up and keep it working.
> > > >> Definitely not me. :-) If I look back how badly the idea of having a
> > > >> windows vm, a mac and a arm device hooked up to Jenkins turned out. I
> > > >> simply gave up on that part.
> > > > well then i just can't see us ever making it a requirement they build
> > across
> > > > these os's on every single commit if it can't be automated and made
> > > > available to developers to actually look into and see what is working
> > or
> > > > not and why. :(
> > > >
> > > >>> not to mention... the test suites need to actually be reliable. i
> > just
> > > >>> found one of the ecore_file tests was grabbing a page from sf.net
> > ... and
> > > >>> now sf.net is refusing to servie it anymore thus test suites keep
> > failing.
> > > >>> tests that are fragile like this should not be some gatekeeper as to
> > if
> > > >>> code goes in or not.
> > > >>>
> > > >>> if a test suite is to be a gatekeeper it has to be done right. that
> > means
> > > >>> it has to work reliably on the build host. run very quickly. things
> > like
> > > >>> testing network fetches has to not rely on anything outside of that
> > > >>> vm/box/chroot etc. etc. ... and we don't have that situation. this
> > > >>> probably needs to be fixed first and foremost. not to mention just
> > > >>> dealing with check and our tests to find what went wrong is a
> > nightmare.
> > > >>> finding the test that goes wrong in a sea of output ... is bad.
> > > >>>
> > > >>> so my take iis this: first work on the steps needed to get the final
> > > >>> outcome. better test infra. easier to write tests. easier to run and
> > find
> > > >>> just the test that failed and run it by itself easily etc. it should
> > be as
> > > >>> simple as:
> > > >>>
> > > >>> make check
> > > >>> ...
> > > >>> FAIL: src/tests/some_test_binary
> > > >>>
> > > >>> and to test it i just copy & paste that binary/line and nothing more
> > and i
> > > >>> get exactly that one test that failed. i don't have to set env vars,
> > read
> > > >>> src code to find the test name and so on. ... it currently is not
> > this
> > > >>> simple by far. ;(
> > > >> Yes, our tests are not as reliable as they should be.
> > > >> Yes, they would need to run in an controlled env.
> > > >> Yes, we might need so look at alternatives to libcheck.
> > > > i'm just saying these need to not randomly reject commits from devs
> > when the
> > > > issue has nothing to do with what the dev is doing. it can't become an
> > > > automated requirement unless its reliably correct. :(
> > > >
> > > >> But even with me agreeing to the three things above the core question
> > stays
> > > >> still open.
> > > >>
> > > >> Is this developer community willing to accept a working test suite as
> > a
> > > >> gatekeeper? I don't think this is the case.
> > > > i think it's best to make it an expectation that devs run make check
> > and
> > > > compile against efl before a push
> > >
> > > That is an expectation that is failing in my reality for many years now.
> > > Getting people to run make check or even distcheck is a fight against
> > > windmills. Before I normally can do the first beta release from master
> > for a
> > > new cycle I need to do onion bug fixing by peeling off one bug after
> > another
> > > to finally get some tarballs produced.
> >
> > i know that at least i run make check a few times per week and i am doing
> > multiple builds from scratch per day including building against efl.
> >
> > > > before even considering making it the
> > > > gatekeeper. we aren't even there yet with enough tooling, let alone
> > talking
> > > > of an automated gatekeeper. just a reliable, easy to use and complete
> > test
> > > > suite would be a big step forward. i think it's putting the cart
> > before the
> > > > horse to talk automated gatekeepers + jenkins ... etc. without getting
> > > > these first things right.
> > >
> > > The fun fact here is that most bogus reports actually come from another
> > bugs
> > > that have been introduced before and are now shadowing new ones.
> > >
> > > We also have bogus reports due to our homebrewn infrastructure and
> > Jenkins
> > > misconfigurations.
> >
> > that's not too great. :(
> >
> > > >> My personal motivation to work on QA and CI has gone down to zero
> > over the
> > > >> years. It just feels like a Sisyphus task to look at master again and
> > again
> > > >> why it is broken now. Dig through the commits, bisect them, point
> > fingers
> > > >> and constantly poke people top get it fixed. All long after the
> > problem
> > > >> have entered master.
> > > > what you want is gerrit. and i've seen how that works. i'm not a fan.
> > it
> > > > ends up either:
> > > >
> > > > 1. it's ignored and patches sit in review for weeks, months or years
> > and
> > > > vanish or
> > > > 2. it's gamed because everything has to go through it, it's minimized
> > to try
> > > > and remove the impact. people just vote things up without real review
> > etc.
> > > >
> > > > if you automated the voting to a build check instead of humans, you'd
> > need
> > > > to find some way to do that and have a test build bot vote and do it
> > FAST.
> > > > that means you need enough infra for a build per commit and it has to
> > be
> > > > totally reliable. the build env and the tests. is that really possible?
> > > > then if you don't do them in strict order you end up with conflicts,
> > and if
> > > > some are rejected from a push with multiple commits, you have dependent
> > > > commits...
> > > >
> > > > i don't see how this ends up better. it ends up worse i think.
> > > >
> > > >> I willing to admit that the approach I used to reach my goals might
> > have
> > > >> been flawed and simply failed. Someone else might want to pick up the
> > > >> slack on it.
> > > > i really don't think we have a lot of break issues given size and
> > > > complexity. not build breaks anyway.
> > >
> > > See the list above just for the four I see today.
> >
> > well i guess i disabled cxx bindings so i stop seeing those, but i do see
> > things buildiong cross-compile for windows and on freebsd as well as arm32
> > regularly and don't run into build issues there.
> >
> > indeed your examples are for systems i don't have builds or build infra
> > for.
> >
> > > >  right now if jenkins detects a build break... how does
> > > > a developer know? it can take hours before it detects something.
> > should they
> > > > sit hitting reload on the browser for the next hour hoping one of the
> > builds
> > > > going contains their commit? we have a broken feedback cycle. jenkins
> > should
> > > > mail the mailing list with "commit X broke build: log etc. link here".
> > if a
> > > > build breaks, jenkins should go back commits until it finds a working
> > one
> > >
> > > Oh, now we even expect automated git bisecting from it? :-)
> >
> > i was even thinking a simple "roll back 1 at a time brute force until
> > found" :)
> > nothing as fancy as a bisect! :) assuming
> >
> > > >  then
> > > > report the commit that broke. at least devs would get a notification.
> > >
> > > These mails are getting sent to the e-b0rk mailing list as I have been
> > asked
> > > to not sent them to the main development list.
> >
> > i totally missed this list. i guess that's why i don't know. :) i should
> > subscribe. i imagine the lack of knowing about this might be part of the
> > problem.
> >
> > > >  unless i
> > > > they sit staring at build.e.org with reloads or someone tells me, i
> > just
> > > > have no idea a build would have broken.
> > > >
> > > > i think we've talked about this many times before... :)
> > > >
> > >
> > > We did and we both had many of the same arguments before. :-)
> > >
> > > I reached the point where I have no motivation left for doing CI/QA work,
> > > though. Thus I am going to drop it and hope someone else picks it up.
> >
> > :(
> >
> > --
> > ------------- Codito, ergo sum - "I code, therefore I am" --------------
> > Carsten Haitzler - ras...@rasterman.com
> >
> >
> >
> > ------------------------------------------------------------------------------
> > Check out the vibrant tech community on one of the world's most
> > engaging tech sites, Slashdot.org! http://sdm.link/slashdot
> > _______________________________________________
> > enlightenment-devel mailing list
> > enlightenment-devel@lists.sourceforge.net
> > https://lists.sourceforge.net/lists/listinfo/enlightenment-devel
> >
> ------------------------------------------------------------------------------
> Check out the vibrant tech community on one of the world's most
> engaging tech sites, Slashdot.org! http://sdm.link/slashdot
> _______________________________________________
> enlightenment-devel mailing list
> enlightenment-devel@lists.sourceforge.net
> https://lists.sourceforge.net/lists/listinfo/enlightenment-devel
> 


-- 
------------- Codito, ergo sum - "I code, therefore I am" --------------
Carsten Haitzler - ras...@rasterman.com


------------------------------------------------------------------------------
Check out the vibrant tech community on one of the world's most
engaging tech sites, Slashdot.org! http://sdm.link/slashdot
_______________________________________________
enlightenment-devel mailing list
enlightenment-devel@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/enlightenment-devel

Reply via email to