Hi (esp. Ben),
here's my feedback on
https://wiki.opencog.org/w/Measuring_Surprisingness
Let me jump right into the matter
1. maxSub = max{ g(coh(A1)) * g(coh(B1)) * g(coh(C1)) * P(A1 B1 C1) |
A1 is a subset of A, and B1 is a subset of B, and C1 is a subset of C) }
I think is wrong because
1.1. I don't think coh(A1) to coh(C1) should be used, but rather
coh(A) to coh(C). This depends on the exact definition of coh
of course but I would think that the coherence of a set should
be defined by the regularities within its subsets, not its
supersets, it seems to much to ask (I mean, I can't expect the
universe to be regular just cause my computer file system is
in order, etc, the other way around would work better though),
but again it's a question of definition...
1.2. Even if the coherences set to 1 (best case) P(A1 B1 C1) may
still badly underestimate P(A B C) (unless coh(A1) somehow
implies that P(A1)~=P(A) for superset A of A1, which I don't
see why). The example calculation maxSub(CAR, horror)=
P(Bangui horror) = .0004 is consistent with that
observation. I suspect the intended formula is
maxSub = max{ g(coh(A1)) * g(coh(B1)) * g(coh(C1)) * P(A1 B1 C1) *
P(A)/P(A1) * P(B)/P(B1) * P(C)/P(C1) | A1 is a subset of A, and B1 is a
subset of B, and C1 is a subset of C) }
that is rescaling factors P(A)/P(A1), etc, have been put.
Another option I guess would be to define maxSub as the mirror
of maxSuper (i.e. swapping supersets for subsets).
2. minSub = min{ h(coh(A1)) * h(coh(B1)) * h(coh(C1)) * [ P(A1 B1 C1) +
minInd_{A*, B*, C*} ] | A1 is a subset of A, and B1 is a subset of B,
and C1 is a subset of C) }
where A* = A - A1, etc.; and minInd_{A*, B*, C*} indicates minInd
evaluated at A*, B*, C* rather than A, B, C. The point of
The explanation ends with "The point of". But I think I understand,
however I don't think it's correct, it seems minInd_{A*, B*, C*} is
gonna underestimate too much the rest because it dismisses
combinations such as P(A* B1 C1), P(A1 B* C1), etc. Thus IMO better
use the scaling factors as in my correction of maxSub. Actually,
the example
minSub(CAR, horror) = P(Bangui horror) + ( P(CAR-Bangui horror) -
P(CAR-Bangui)P(horror) )
doesn't match the formula so maybe it has a typo, or I
misunderstood, still seems wrong/incomplete though.
Obviously I understand that these are heuristics and so they don't
have to be correct, though I suppose they could be better (kinda the
definition of a heuristic). Other than that I think it makes a lot of
sense.
More general remarks...
3. It seems the intervals (obtained from minInd, maxInd, etc) used to
measure the deviation between expectation and "reality" could be
replaced by second order distributions. This would avoid the
possible problem of having too wide intervals, and would also allow
concepts based on fewer data points to not be too misleading.
Actually I recall being asked by Misgana some code to calculate the
Jensen-Shannon divergence for some interestingness-based
attentional focus experiments, so maybe it has already been
attempted, even rudimentarily.
4. Last point, I just want to be sure I understand why surprisingness
is useful. It seems to me, if we agree that surprisingness is the
difference between expectation and reality, it then is a useful
indicator to gain knowledge (i.e. perfect our model of reality).
Also, surprisingness is subsumed by interestingness, right?
Something interesting isn't necessarily surprising, but something
surprising is (always?) interesting. And something interesting is
something worth focusing on, so surprisingness is a particular
mechanism to detect and deal with interestingness, right? Anyway, I
know it's pretty basic, and I don't expect much disagreement but
natural language is ambiguous (especially to me) so...
Thanks,
Nil
--
You received this message because you are subscribed to the Google Groups
"opencog" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/opencog.
To view this discussion on the web visit
https://groups.google.com/d/msgid/opencog/ab490c01-05dc-56b6-e50f-195649c61c15%40gmail.com.
For more options, visit https://groups.google.com/d/optout.