> Water does not always run downhill, sometimes it runs uphill.

Consider an AGI trying to discover world facts from textual inference and 
finding Dawkin's book "The River that Runs Uphill" (or things about the 
moon's color and finding the phrase "once in a blue moon" everywhere or the 
SF story "The Moon is Green").

The problem seems to be that it's AI-complete to pick out the exceptions, the 
figures of speech, the metaphorical uses, and so forth -- so you could use 
textual inference if only you had an already working AGI :-)

Bottom line: kids grow up in a confusing, ambiguous world (and there are 
plenty of funny stories obout how they get things wrong as they learn) -- yet 
they generally seem to be able to organize their experiences into usable 
knowledge corpora as they go. The human mind is autogenous. What's the trick?

-------------------------------------------
agi
Archives: http://www.listbox.com/member/archive/303/=now
RSS Feed: http://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
http://www.listbox.com/member/?member_id=8660244&id_secret=95818715-a78a9b
Powered by Listbox: http://www.listbox.com

Reply via email to