I propose to do this tomorrow Tuesday, starting with an intro from me, using youtube live.
WDYT? Thanks -Vincent > On 30 Aug 2018, at 12:27, Adel Atallah <[email protected]> wrote: > > Just to be clear, when I proposed "having a whole day dedicated on > using these tools", I didn't meant having to have it every week but > only once, so we can properly start improving the tests. It would be > some kind of training. > On my side I don't think I'll be able to have on a week one day > dedicated to tests and one for bug fixing, I won't have time left for > the roadmap as I will only work on the product 50% of the time. > > > On Thu, Aug 30, 2018 at 12:18 PM, Vincent Massol <[email protected]> wrote: >> Hi, >> >> I don’t remember discussing this with you Thomas. Actually I’m not convinced >> to have a fixed day: >> * we already have a fixed BFD and having a second one doesn’t leave much >> flexibility for working on roadmap items when it’s the best >> * test sessions can be short (0.5-1 hours) and it’s easy to do them between >> other tasks >> * it can be boring to spend a full day on them >> >> Now, I agree that not having a fixed day will make it hard to make sure that >> we work 20% on that topic. >> >> So if you prefer we can define a day, knowing that some won’t be able to >> always attend during that day and in this case they should do it on another >> day. What’s important is to have 20% done each week (i.e. enough work done >> on it). >> >> In term of day, if we have to choose one, I’d say Tuesday. That’s the most >> logical to me. >> >> WDYT? What do you prefer? >> >> Thanks >> -Vincent >> >>> On 30 Aug 2018, at 10:38, Thomas Mortagne <[email protected]> wrote: >>> >>> Indeed we discussed this but I don't see it in your mail Vincent. >>> >>> On Thu, Aug 30, 2018 at 10:33 AM, Adel Atallah <[email protected]> >>> wrote: >>>> Hello, >>>> >>>> Maybe we should agree on having a whole day dedicated on using these >>>> tools with a maximum number of developers. >>>> That way we will be able to help each other and maybe it will make the >>>> process easier to carry out in the future. >>>> >>>> WDYT? >>>> >>>> Thanks, >>>> Adel >>>> >>>> >>>> On Wed, Aug 29, 2018 at 11:20 AM, Vincent Massol <[email protected]> >>>> wrote: >>>>> Hi devs (and anyone else interested to improve the tests of XWiki), >>>>> >>>>> History >>>>> ====== >>>>> >>>>> It all started when I analyzed our global TPC and found that it was going >>>>> down globally even though we have the fail-build-on-jacoco-threshold >>>>> strategy. >>>>> >>>>> I sent several email threads: >>>>> >>>>> - Loss of TPC: http://markmail.org/message/hqumkdiz7jm76ya6 >>>>> - TPC evolution: http://markmail.org/message/up2gc2zzbbe4uqgn >>>>> - Improve our TPC strategy: http://markmail.org/message/grphwta63pp5p4l7 >>>>> >>>>> Note: As a consequence of this last thread, I implemented a Jenkins >>>>> Pipeline to send us a mail when the global TPC of an XWiki module goes >>>>> down so that we fix it ASAP. This is still a development in progress. A >>>>> first version is done and running at >>>>> https://ci.xwiki.org/view/Tools/job/Clover/ but I need to debug it and >>>>> fix it (it’s not working ATM). >>>>> >>>>> As a result of the global TPC going down/stagnating, I have proposed to >>>>> have 10.7 focused on Tests + BFD. >>>>> - Initially I proposed to focus on increasing the global TPC by looking >>>>> at the reports from 1) above >>>>> (http://markmail.org/message/qjemnip7hjva2rjd). See the last report at >>>>> https://up1.xwikisas.com/#mJ0loeB6nBrAgYeKA7MGGw (we need to fix the red >>>>> parts). >>>>> - Then with the STAMP mid-term review, a bigger urgency surfaced and I >>>>> asked if we could instead focus on fixing tests as reported by Descartes >>>>> to increase both coverage and mutation score (ie test quality), since >>>>> those are 2 metrics/KPIs measured by STAMP and since XWiki participates >>>>> to STAMP we need to work on them and increase them substantially. See >>>>> http://markmail.org/message/ejmdkf3hx7drkj52 >>>>> >>>>> The results of XWiki 10.7 has been quite poor on test improvements (more >>>>> focus on BFD than tests, lots of devs on holidays, etc). This forces us >>>>> to have a different strategy. >>>>> >>>>> Full Strategy proposal >>>>> ================= >>>>> >>>>> 1) As many XWiki SAS devs as possible (and anyone else from the community >>>>> who’s interested ofc! :)) should spend 1 day per week working on >>>>> improving STAMP metrics >>>>> * Currently the agreement is that Thomas and myself will do this for the >>>>> foreseeable future till we get some good-enough metric progress >>>>> * Some other devs from XWiki SAS will help out for XWiki 10.8 only FTM >>>>> (Marius, Adel if he can, Simon in the future). The idea is to see where >>>>> that could get us by using substantial manpower. >>>>> >>>>> 2) All committers: More generally the global TPC failure is also already >>>>> active and dev need to modify modules that see their global TPC go down. >>>>> >>>>> 3) All committers: Of course, the jacoco strategy is also active at each >>>>> module level. >>>>> >>>>> STAMP tools >>>>> ========== >>>>> >>>>> There are 4 tools developed by STAMP: >>>>> * Descartes: Improves quality of tests by increasing their mutation >>>>> scores. See http://markmail.org/message/bonb5f7f37omnnog and also >>>>> https://massol.myxwiki.org/xwiki/bin/view/Blog/MutationTestingDescartes >>>>> * DSpot: Automatically generate new tests, based on existing tests. See >>>>> https://massol.myxwiki.org/xwiki/bin/view/Blog/TestGenerationDspot >>>>> * CAMP: Takes a Dockerfile and generates mutations of it, then deploys >>>>> and execute tests on the software to see if the mutation works or not. >>>>> Note this is currently not fitting the need of XWiki and thus I’ve been >>>>> developing another tool as an experiment (which may go back in CAMP one >>>>> day), based on TestContainers, see >>>>> https://massol.myxwiki.org/xwiki/bin/view/Blog/EnvironmentTestingExperimentations >>>>> * EvoCrash: Takes a stack trace from production logs and generates a test >>>>> that, when executed, reproduces the crash. See >>>>> https://markmail.org/message/v74g3tsmflquqwra. See also >>>>> https://github.com/SERG-Delft/EvoCrash >>>>> >>>>> Since XWiki is part of the STAMP research project, we need to use those 4 >>>>> tools to increase the KPIs associated with the tools. See below. >>>>> >>>>> Objectives/KPIs/Metrics for STAMP >>>>> =========================== >>>>> >>>>> The STAMP project has defined 9 KPIs that all partners (and thus XWiki) >>>>> need to work on: >>>>> >>>>> 1) K01: Increase test coverage >>>>> * Global increase by reducing by 40% the non-covered code. For XWiki >>>>> since we’re at about 70%, this means reaching about 80% before the end of >>>>> STAMP (ie. before end of 2019) >>>>> * Increase the coverage contributions of each tool developed by STAMP. >>>>> >>>>> Strategy: >>>>> * Primary goal: >>>>> ** Increase coverage by executing Descartes and improving our tests. This >>>>> is http://markmail.org/message/ejmdkf3hx7drkj52 >>>>> ** Don’t do anything with DSpot. I’ll do that part. Note that the goal is >>>>> to write a Jenkins pipeline to automatically execute DSpot from time to >>>>> time and commit the generated tests in a separate test source and have >>>>> our build execute both src/test/java and this new test source. >>>>> ** Don’t do anything with TestContainers FTM since I need to finish a >>>>> first working version. I may need help in the future to implement docker >>>>> images for more configurations (on Oracle, in a cluster, with >>>>> LibreOffice, with an external SOLR server, etc). >>>>> ** For EvoCrash: We’ll count contributions of EvoCrash to coverage in K08. >>>>> * Secondary goal: >>>>> ** Increase our global TPC as mentioned above by fixing the modules in >>>>> red. >>>>> >>>>> 2) K02: Reduce flaky tests. >>>>> * Objective: reduce the number of flaky tests by 20% >>>>> >>>>> Strategy: >>>>> * Record flaky tests in jira >>>>> * Fix the max number of them >>>>> >>>>> 3) K03: Better test quality >>>>> * Objective: increase mutation score by 20% >>>>> >>>>> Strategy: >>>>> * Same strategy as K01. >>>>> >>>>> 4) K04: More configuration-related paths tested >>>>> * Objective: increase the code coverage of configuration-related paths in >>>>> our code by 20% (e.g. DB schema creation, cluster)related code, >>>>> SOLR-related code, LibreOffice-related code, etc). >>>>> >>>>> Strategy: >>>>> * Leave it to FTM. The idea is to measure Clover TPC with the base >>>>> configuration, then execute all other configurations (with >>>>> TestContainers) and regenerate the Clover report to see how much the TPC >>>>> has increased. >>>>> >>>>> 5) K05: Reduce system-specific bugs >>>>> * Objective: 30% improvement >>>>> >>>>> Strategy: >>>>> * Run TestContainers, execute existing tests and find new bugs related to >>>>> configurations. Record them >>>>> >>>>> 6) K06: More configurations/Faster tests >>>>> * Objective: increase the number of automatically tested configurations >>>>> by 50% >>>>> >>>>> Strategy: >>>>> * Increase the # of configurations we test with TestContainers. I’ll do >>>>> that part initially. >>>>> * Reduce time it takes to deploy the software under a given configuration >>>>> vs time it used to take when done manually before STAMP. I’ll do this >>>>> one. I’ve already worked on it in the past year with the dockerization of >>>>> XWiki. >>>>> >>>>> 7) K07: Pending, nothing to do FTM >>>>> >>>>> 8) K08: More crash replicating test cases >>>>> * Objective: increase the number of crash replicating test cases by at >>>>> least 70% >>>>> >>>>> Strategy: >>>>> * For all issues that are still open and that have stack traces and for >>>>> all issues closed but without tests, run EvoCrash on them to try to >>>>> generate a test. >>>>> * Record and count the number of successful EvoCrash-generated test cases. >>>>> * Derive a regression test (which can be very different from the negative >>>>> of the test generated by evocrash!). >>>>> * Measure the new coverage increase >>>>> * Note that I haven’t experimented much with this yet myself. >>>>> >>>>> 9) K09: Pending, nothing to do FTM. >>>>> >>>>> Conclusion >>>>> ========= >>>>> >>>>> Right now, I need your help for the following KPIs: K01, K02, K03, K08. >>>>> >>>>> Since there’s a lot to understand in this email, I’m open to: >>>>> * Organizing a meeting on youtube live to discuss all this >>>>> * Answering any questions on this thread ofc >>>>> * Also feel free to ask on IRC/Matrix. >>>>> >>>>> Here’s an extract from STAMP which has more details about the >>>>> KPIs/metrics: >>>>> https://up1.xwikisas.com/#QJyxqspKXSzuWNOHUuAaEA >>>>> >>>>> Thanks >>>>> -Vincent >>>>> >>>>> >>>>> >>>>> >>>>> >>>>> >>> >>> >>> >>> -- >>> Thomas Mortagne >>

