Hey all.

I will soon have some student dev resources and I'm pondering how to 
best use them.
Apart from the hundreds of issues, one thing I was thinking about adding 
is more tests against reference implementations,
and having speed regression tests.

For the reference implementations, we could hard-code the results of 
algorithms into the tests. That is done for some
algorithms, but only very few. It would guard us against "obvious" 
functionality bugs, which still show up from time to time.

For speed regression tests, it has happened that things got slower, in 
particular with innocent looking things like input validation.
I think it would be good to have some tests that ensure that we don't 
get too much slower.
I'm not entirely sure how do to that, though.
I know Vlad put some effort into a continuous benchmarking suite, but I 
think since then there have been several
efforts to log speed of implementations in a consistent way, and we 
might want to look into these.

Do you think that these are interesting issues to work on, or do you 
think there are more pressing ones?

We still have a lot to do on the API side, though I'm a bit hesitant to 
give that to new devs.

Cheers,
Andy

------------------------------------------------------------------------------
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general

Reply via email to