2013/3/21 Ricardo Corral C. <ricardocorr...@gmail.com>:
> Ok, this is a brief description of what I'm interested in.
>
> Recently, I faced a problem of evaluating the quality of a method to
> obtain features from protein structures.
> I adopted the approach given in [1] to measure separability of my
> classes independently of my capacity of make good predictions.
> This is basically a hypothesis testing of whether or not the
> distribution of classes over feature vectors is somewhat random.
> This test is made over the construction of a Relative Neighbourhood
> Graph, which is O(n^3), thus, so prohibitive for practical use.
> There is an efficient method for constructing RNG on the plane
> described in [2] O(n*log(n)), but O(n^2) for a higher d dimension (in
> fact O(n^2*f(d)) with f(d) <= (2*sqrt(d) +2)^d...).
>
> Actually, I have the test implemented, and I'm refining a speedup of
> RNG construction based on the Half-Space Proximal (HSP) graph. This is
> O(n^2log(n)), and there is no dependence of dimension other than time
> consumed in calculating distances.
>
> This is made by doing RNG test over edges in HSP (attached images for
> clarify this).
>
> Could this be of interest for sklearn users? And if so, be considered for 
> GSoC?

This looks interesting but please include the missing references.

--
Olivier
http://twitter.com/ogrisel - http://github.com/ogrisel

------------------------------------------------------------------------------
Everyone hates slow websites. So do we.
Make your web apps faster with AppDynamics
Download AppDynamics Lite for free today:
http://p.sf.net/sfu/appdyn_d2d_mar
_______________________________________________
Scikit-learn-general mailing list
Scikit-learn-general@lists.sourceforge.net
https://lists.sourceforge.net/lists/listinfo/scikit-learn-general

Reply via email to