On 8/15/06, Matt Mahoney <[EMAIL PROTECTED]> wrote:
I realize it is tempting to use lossy text compression as a test for AI
because that is what the human brain does when we read text and recall it in
paraphrased fashion. We remember the ideas and discard details about the
expression of those ideas. A lossy text compressor that did the same thing
would certainly demonstrate AI.
But there are two problems with using lossy compression as a test of AI:
1. The test is subjective.
2. Lossy compression does not imply AI.
Lets assume we solve the subjectivity problem by having human judges
evaluate whether the decompressed output is "close enough" to the input. We
already do this with lossy image, audio and video compression (without much
The second problem remains: ideal lossy compression does not imply passing
the Turing test. For lossless compression, it can be proven that it does.
Let p(s) be the (unknown) probability that s will be the prefix of a text
dialog. Then a machine that can compute p(s) exactly is able to generate
response A to question Q with the distribution p(QA)/p(Q) which is
indistinguishable from human. The same model minimizes the compressed size,
This proof is really not useful. The Turing test is subjective; all
you are saying is that lossy compression is lossy, and lossless
compression is not. A solution to the first problem would also solve
the second problem.
It is necessary to allow lossy compression in order for this
compression test to be useful for AI, because lossless and
uncomprehending compression is already bumping up against the
theoretical limits for text compression.
To unsubscribe, change your address, or temporarily deactivate your subscription,
please go to http://v2.listbox.com/member/[EMAIL PROTECTED]