Steve:View #2 (mine, stated from your approximate viewpoint) is that simple 
programs (like Dr. Eliza) have in the past and will in the future do things 
that people aren't good at. This includes tasks that encroach on 
"intelligence", e.g. modeling complex phonema and refining designs.

Steve,

In principle, I'm all for the idea that I think you (and perhaps Bryan) have 
expressed of a "GI Assistant" - some program that could be of general 
assistance to humans dealing with similar problems across many domains. A 
diagnostics expert, perhaps, that could help analyse breakdowns in say, the 
human body, a car or any of many other machines, a building or civil structure, 
etc. etc. And it's certainly an idea worth exploring.

 But I have yet to see any evidence that it is any more viable than a proper 
AGI - because, I suspect, it will run up against the same problems of 
generalizing -  e.g. though "breakdowns" may be v. similar in many different 
kinds of machines, technological and natural, they will also each have their 
own special character.

If you are serious about any such project, it might be better to develop it 
first as an intellectual discipline.rather than a program to test its viability 
- perhaps what it really comes down to is a form of systems thinking or science.





-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=114414975-3c8e69
Powered by Listbox: http://www.listbox.com

Reply via email to