No Stefan, we know humans are really good predictors, the actual test is to 
make a really good predictor for text so that your compression is extremely 
high and so that it usually has a high probability for the next letter or bit 
to predict that it should predict during decompression in a lossless generator 
and hence the error it stores is very small ex. 10MB only. 
Compression/decompression is done as a chain of predictions. If you can 
compress the 100MB enwiki8 to 10MB and still regenerate the 100MB back 
losslessly perfectly then this means your prediction accuracy is extremely 
high. Now if you take probabilities for the next letter to predict during 
decompression and let it randomly choose any of the top 10 letters or bits, it 
can regenerate not the enwiki8, but related data you'd want too. And that's how 
you make a lossless text predictor able to return us new data we never had 
directly.
------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/T409fc28ec41e6e3a-Meb07a0ed0463c724dd035b55
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to