> > Maximum entropy principle is about the idea that given a number of
> > equivalent prospective models, you should pick the one with the
> > highest entropy.
> I guess EM also tries to do the same, ie tries to maximize the
> likelihood of incomplete pdf match the complete pdf iteratively for
> said parameter.

You misunderstand maximum entropy.

Assume the true pdf p(x), and the modelled pdf q(x). The purpose of
cross-entropy minimization is to find a q(x) so that the KL-divergence
between the two pdfs will be minimal:

q := argmin_q' D(p || q')

Assume that q is a parametric model, with parameter t. Then q(x)=p(x |
t). The purpose of maximum likelihood (ML) is to pick the model (or
parameters) given which the data are likeliest:

t := argmax_t' p(x | t')
q(x) := p(x | t)

Although these two look similar, they are *not* the same! Let's now
imagine that we have a number of models, q1(x), q2(x), q3(x). The
purpose of MaxEnt is to pick the one with highest entropy H(q):

q := argmax_q' H(q)

H(q) is the entropy of a particular pdf q:

H(q) := -Integrate[q(x)log q(x)]dx

So there is very little similarity between MaxEnt and ML. Again, EM is
only a particular approach to ML. There are some connections, but
they're a tad more intricate (Csiszar & Tusnady 1984).

-- 
mag. Aleks Jakulin
http://ai.fri.uni-lj.si/aleks/
Artificial Intelligence Laboratory,
Faculty of Computer and Information Science, University of Ljubljana.




.
.
=================================================================
Instructions for joining and leaving this list, remarks about the
problem of INAPPROPRIATE MESSAGES, and archives are available at:
.                  http://jse.stat.ncsu.edu/                    .
=================================================================

Reply via email to