On Mar 13, 2007, at 23:27 , Peter Drake wrote:
Hmmm -- p. 735 of Russell & Norvig's AI text contains a strong argument
that "nearest-neighbor methods cannot be trusted for high- dimensional
data".
AFAIK that's because when you add more and more dimensions and try to
calculate the distance
between two data points you get less and less different values the higher
the number of
dimensions. You have so many dimensions that have the same distances
(between different sets
of data points I mean) that the few that are different get lost. Or
that's at least that's what
I've been told.
Urban
- --
http://bettong.net
Explanation makes sense. My comment is maybe out of context, sorry about
that but just ask this anyway. In other context? EA search, I had "idea" of
using dynamic distance functions help local optimum problem. Kind of
accepting fact that don't know how distance should be calculated. For fixing
that distance function changes in search time dimension, and weights
different dimensions differently, so algorithm has even possibility to
advance from local optimum in longer search run and so on don't converge
local optimum. I don't know/remember any studies related dynamic distance(in
my case dynamic fitness) function but kind of intrested if there is as it is
kind of simple idea. It is kind of biologically realistic to think that
enviroment changes and values different attributes differently depending
situation.
This allows such behaviour that small change in one dimension can cause big
difference between data-points without knowing anything about data-point
semantics.
t. Harri
_______________________________________________
computer-go mailing list
[email protected]
http://www.computer-go.org/mailman/listinfo/computer-go/