Yeah, from what I've read I think you're right. The perception is probably more PR and hype, I wonder if that'll hurt them by raising expectations too high? On the technical side and evaluating their approaches though, it almost to a degree gets to the question of statistical and algorithmic based approaches to NLP versus more pure AI related approaches?
-Chris > On Mar 20, 2014, at 9:12 PM, Mike Archbold <[email protected]> wrote: > > I get the feeling the public perceives Watson as general intelligence, > since they seem to be pursuing multiple application areas. I was just > reading about Deep Mind, and again it is billed as the core of an AI > stack. But, in both cases, they still need considerable programming > per specific domain, if I understand correctly what I read about these > approaches. > >> On 3/20/14, Ben Goertzel <[email protected]> wrote: >> Neither Watson nor Google is currently addressing the core issues of AGI. >> But they are building infrastructure, and supporting algorithms, that could >> provide platforms for them to address these in future... Whether they will >> or not, remains to be seen. Maybe they will wait for some smaller firms to >> make real progress on AGI and then buy these firms.... >> >> Google bought Deep Mind before DM had enough time to make any AGI >> breakthroughs. It remains to be seen whether Google will be a good >> environment for them to conduct truly AGI-focused research... >> >> -- Ben G >> >> >> On Fri, Mar 21, 2014 at 6:24 AM, Chris Nolan >> <[email protected]>wrote: >> >>> Speaking of AI applications, I'm curious what people think of recent >>> directions in NLP and QA systems being explored? Say IBM Watson vs >>> Google? >>> Which do people think has the better potential to expand and is the right >>> direction to go in comercializing AI systems? When I say expand, I mean >>> the >>> potential for creating more generality in current AI systems (with the >>> understanding that it's not true AGI work). >>> >>> >>> >>> On Wednesday, March 19, 2014 2:15 PM, John Rose >>> <[email protected]> >>> wrote: >>> It could move in and take over our functions, our jobs, and perform them >>> with such drastically less resource consumption and see us as beings that >>> have competed with almost every other species and destroyed the very >>> nature >>> of this planet. AGI could compete with us as an invasive species would >>> but >>> in a way to make the planet restore to its previous health before humans >>> were here. >>> >>> I don't see it totally happening that way. I see it more as AGI embeds >>> itself into our systems of civilization, our governments, corporations, >>> and >>> slowly renders us over time as being not that important entities, less >>> and >>> less as spiritual beings, more and more as lower level animals with less >>> rights, sort of like p-zombies, the individual having diminished >>> importance, >>> we become herded as sheeple. Kind of like what is happening now as we >>> acquiesce. >>> >>> I don't say it will or has to go that way. The future isn't >>> predetermined. >>> In fact, AGI has the capability to liberate the individual. >>> >>> There are real things to fear though. It's been relatively calm the past >>> few >>> decades but at any moment you could have some new disease break out, an >>> extreme natural disaster, a global famine, an ice age, fear is good. >>> >>> John >>> >>> -----Original Message----- >>> From: just camel [mailto:[email protected]] >>> Sent: Sunday, March 16, 2014 10:21 AM >>> To: AGI >>> Subject: Re: [agi] Practical Applications for AGI >>> >>> Why would it be competitive? There just is no reason. It's a highly >>> anthropomorphic and contemporary notion of us. Just because we think that >>> a >>> competitive culture is a fruitful thing to have does not make competition >>> universal. In fact there have been civilizations that existed (despite >>> having much less of an abundance) without competition for much longer >>> than >>> us and way more sustainable. >>> >>> The problem is that people like Yudkowsky and Bostrom might have an IQ of >>> over 9000 but they are also spiritually trapped in this belief system of >>> fear and scarcity and thus will come up with 1000 biased and illogical >>> scenarios in which AGI will terminate humanity (instead of pointing out >>> that >>> the by far biggest existential risk for humanity stems from our monetary >>> system and neoliberal capitalism). Yet there are thousands of people with >>> a >>> superior consciousness (less inherent >>> entropy) who are not as indoctrinated by our culture and it makes no >>> sense >>> to believe that any superintelligent agent would adopt the irrational, >>> fearful, destructive, de-evolutionary and unsustainable polluted mindset >>> of >>> the average westener? >>> >>> >>>> On 03/19/2014 01:22 PM, John Rose wrote: >>>> Only prob it might relegate us as biological waste byproducts >>>> unnecessary for its competitive survival. >>> >>> >>> >>> ------------------------------------------- >>> AGI >>> Archives: https://www.listbox.com/member/archive/303/=now >>> RSS Feed: https://www.listbox.com/member/archive/rss/303/248029-3b178a58 >>> Modify Your Subscription: >>> https://www.listbox.com/member/?& >>> Powered by Listbox: http://www.listbox.com >>> >>> >>> >>> >>> >>> ------------------------------------------- >>> AGI >>> Archives: https://www.listbox.com/member/archive/303/=now >>> RSS Feed: >>> https://www.listbox.com/member/archive/rss/303/20347893-f72b365c >>> >>> Modify Your Subscription: https://www.listbox.com/member/?& >>> Powered by Listbox: http://www.listbox.com >>> >>> >>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/212726-deec6279> | >>> Modify<https://www.listbox.com/member/?&>Your Subscription >>> <http://www.listbox.com> >> >> >> >> -- >> Ben Goertzel, PhD >> http://goertzel.org >> >> "In an insane world, the sane man must appear to be insane". -- Capt. James >> T. Kirk >> >> "Emancipate yourself from mental slavery / None but ourselves can free our >> minds" -- Robert Nesta Marley >> >> >> >> ------------------------------------------- >> AGI >> Archives: https://www.listbox.com/member/archive/303/=now >> RSS Feed: https://www.listbox.com/member/archive/rss/303/11943661-d9279dae >> Modify Your Subscription: >> https://www.listbox.com/member/?& >> Powered by Listbox: http://www.listbox.com > > > ------------------------------------------- > AGI > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/20347893-f72b365c > Modify Your Subscription: https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
