Hey all. I will soon have some student dev resources and I'm pondering how to best use them. Apart from the hundreds of issues, one thing I was thinking about adding is more tests against reference implementations, and having speed regression tests.
For the reference implementations, we could hard-code the results of algorithms into the tests. That is done for some algorithms, but only very few. It would guard us against "obvious" functionality bugs, which still show up from time to time. For speed regression tests, it has happened that things got slower, in particular with innocent looking things like input validation. I think it would be good to have some tests that ensure that we don't get too much slower. I'm not entirely sure how do to that, though. I know Vlad put some effort into a continuous benchmarking suite, but I think since then there have been several efforts to log speed of implementations in a consistent way, and we might want to look into these. Do you think that these are interesting issues to work on, or do you think there are more pressing ones? We still have a lot to do on the API side, though I'm a bit hesitant to give that to new devs. Cheers, Andy ------------------------------------------------------------------------------ _______________________________________________ Scikit-learn-general mailing list Scikit-learn-general@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/scikit-learn-general