On Wed, Aug 27, 2014 at 1:53 PM, Terren Suydam <[email protected]> wrote:
> > On Wed, Aug 27, 2014 at 6:31 AM, Telmo Menezes <[email protected]> > wrote: > >> >> >> On Wed, Aug 27, 2014 at 1:14 AM, Terren Suydam <[email protected]> >> wrote: >> >>> Hi Telmo, >>> >>> I think if it were as simple as you make it seem, relative to what we >>> have today, we'd have engineered systems like that already. >>> >> >> It wasn't my intention to make it look simple. What I claim is that we >> already have a treasure trove of very interesting algorithms. None of them >> is AGI, but what they can do becomes more impressive with more computing >> power and access to data. >> > > I agree that can be made to do impressive things. Watson definitely > impressed me. > > Take google translator. It's far from perfect, but way ahead anything we >> had a decade ago. As far as I can tell, this was achieved with algorithms >> that had been known for a long time, but that now can operate on the >> gigantic dataset and computer farm available to google. >> >> Imagine what a simple minimax search tree could do with immense computing >> power and data access. >> > > The space of possibilities quickly scales beyond the wildest imaginings of > computing power. Chess AIs are already better than humans, because they > more or less implement this approach, and it turns out you "only" need to > computer a few hundred million positions per second to do that. Obviously > that's a toy environment... the possibilities inherent in the real world > are even be enumerable according to some predefined ontology (i.e. that > would be required to specify in a minimax type AI). > Ok, but of course minimax was also a toy example. Several algorithms that already exist could be combined: deep learning, bayesian belief networks, genetic programming and so on. A clever combination of algorithms plus the still ongoing exponential growth in available computational power could soon unleash something impressive. Of course I am just challenging your intuition, mostly because it's a fun topic :) Who knows who's right... Another interesting/scary scenario to think about is the possibility of a self-mutating computer program proliferating under our noses until it's too late (and exploiting the Internet to create a very powerful meta-computer by stealing a few cpu cycles from everyone). > > >> >> >>> You're talking about an AI that arrives at novel solutions, which >>> requires the ability to invent/simulate/act on new models in new domains >>> (AGI). >>> >> >> Evolutionary computation already achieves novelty and invention, to a >> degree. I concur that it is still not AGI. But it could already be a >> threat, given enough computational resources. >> > > AGI is a threat because it's utility function would necessarily be > sufficiently "meta" that it could create novel sub-goals. We would not > necessarily be able to control whether it chose a goal that was compatible > with ours. > > It comes down to how the utility function is defined. For Google Car, the > utility function probably tests actions along the lines of "get from A to B > safely, as quickly as possible". If a Google Car is engineered with > evolutionary methods to generate novel solutions (would be overkill but > bear with me), the novelty generated is contained within the utility > function. It might generate a novel route that conventional map algorithms > wouldn't find, but it would be impossible for it to find a solution like > "helicopter the car past this traffic jam". > What prevents the car from transforming into an helicopter and flying is not the utility function but the set of available actions. I have been playing with evolutionary computation for some time now, and one thing I learned is to not trust my intuition on the real constraints implied by such set of actions. > > >> >> >>> I'm not saying this is impossible, in fact I see this as inevitable on a >>> longer timescale. I'm saying that I doubt that the military is committing >>> any significant resources into that kind of research when easier approaches >>> are much more likely to bear fruit... but I really have no idea what the >>> military is researching, so it's just a hunch. >>> >> >> Why does it matter if it's the military that does this? To a sufficiently >> advanced AI, we are just monkeys throwing rocks at each other. It will >> surely figure out a way to take control of our resources, including >> weaponry. >> >> > > I think the thread started with a focus on killing machines. But your > point is taken. > > >> >>> What I would wager on is that the military is developing drones along >>> the same lines as what Google has achieved with its self-driving cars. >>> Highly competent, autonomous drones that excel in very specific >>> environments. The utility functions involved would be specified explicitly >>> in terms of "hard-coded" representations of stimuli. For AGI they would >>> need to be equipped to invent new models of the world, articulate those >>> models with respect to self and with respect to existing goal structures, >>> simulate them, and act on them. I think we are a long way from those kinds >>> of AIs. The only researcher I see making inroads towards that kind of AI is >>> Steve Grand. >>> >> >> But again, a reasonable fear is that a sufficiently powerful conventional >> AI is already a threat (due to increasing autonomy and data access + our >> possible inability to cover all the loopholes in utility functions). >> >> > The threats involved with AIs are contained within the scope of their > utility functions. As it turns out, the moment you widen the utility > function beyond a very narrow (and specifiable) domain, AI gets much, much > harder. > As above... Telmo. > > Terren > > >> Cheers >> Telmo. >> >> > -- > You received this message because you are subscribed to the Google Groups > "Everything List" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > To post to this group, send email to [email protected]. > Visit this group at http://groups.google.com/group/everything-list. > For more options, visit https://groups.google.com/d/optout. > -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. Visit this group at http://groups.google.com/group/everything-list. For more options, visit https://groups.google.com/d/optout.

