> Water does not always run downhill, sometimes it runs uphill. Consider an AGI trying to discover world facts from textual inference and finding Dawkin's book "The River that Runs Uphill" (or things about the moon's color and finding the phrase "once in a blue moon" everywhere or the SF story "The Moon is Green").
The problem seems to be that it's AI-complete to pick out the exceptions, the figures of speech, the metaphorical uses, and so forth -- so you could use textual inference if only you had an already working AGI :-) Bottom line: kids grow up in a confusing, ambiguous world (and there are plenty of funny stories obout how they get things wrong as they learn) -- yet they generally seem to be able to organize their experiences into usable knowledge corpora as they go. The human mind is autogenous. What's the trick? ------------------------------------------- agi Archives: http://www.listbox.com/member/archive/303/=now RSS Feed: http://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b Powered by Listbox: http://www.listbox.com
