Josh: Any well designed AI system should not have the masturbatory tendencies
to take unjustified risks.
Josh,
Jeez, you guys will not face reality. MOST of the problems we deal with involve
risks (and uncertainty). That's what human intelligence does most of the time -
that's what any adaptive intelligence does & will have to do - deal with
problems involving risks and uncertainty. Even your superduperintelligence will
have to face those too. (Who knows how those pesky, rebellious humans may try
to resist its benevolent decisions?) Let me quote again
Most of the problems that we face in our everyday lives are ill-defined
problems. In contrast, psychologists have focussed mainly on well-defined
problems. Why have they done this? One important reason is because well-defined
problems have a best strategy for their solution. As a result it is usually
easy to identify the errors and deficiencies in the strategies adopted by human
problem-solvers..
Michael Eysenck, Principles of Cognitive Psychology, East Sussex: Psychology
Press 2001
You guys are similarly copping out - dealing with the easy rather than the hard
(risk & uncertainty) problems - the ones with the neat answers, as opposed to
the ones that are open-ended. The AI as opposed to the AGI problems.
Won't somebody actually deal with the problem - how will your AGI system
decide to invest or not to invest $10,000 in a Chinese mutual fund tomorrow?
(You guys are supposed to be in the problem-solving business).
Mark, it seems that you're missing the point. We as humans aren't ABSOLUTELY
CERTAIN of anything. But we are perfectly capable of operating on the fine
line between assumed certainty and uncertainty. We KNOW that molecules are
made of up bonded atoms, but past a certain point, we can't say what a basic
unit of energy is. But we know what a molecule looks like so we can bond atoms
to form them. Much is the same with intelligence. We simply mimic behaviors
and find parallels in code and systems. If we can prove that a turing machine
is universal, or that a code base is universal, then there must be some
configuration of code that is capable of representing every subatomic
interaction occurring in our universe down to the most minute detail, thus
duplicating our experience entirely (regardless of the fact that this would
take several universe lifetimes to do manually). So if this is plausible, it's
perfectly sound to discuss optimization of the stated code. Our brains (and
their emergent trait of our fancy, intelligence) are a much smaller piece of
this (computable) chemical puzzle. Since they were derived through evolution
(the high intelligence, low efficiency topic), there are many inefficient
mechanisms that have evolved for reasons other than exponentially increasing
our brains processing power. Why is this hard to grasp?
In terms of your investment question, it's all a matter of needs. That is a
simple risk assessment. To any intelligent being, the money gained is only an
ends to a means. To an AI interested in furthering it's knowledge, or
bettering mankind (or machine kind), money simply means more energy, power,
resources, etc. Ultimately, if your goal is just to amass money without any
reasoning, your goals system is flawed. Any well designed AI system should not
have the masturbatory tendencies to take unjustified risks. We're talking
about multiple priority levels here. The Nash equilibrium would be sought
after on many levels. The computer is going to give the system it's best shot
and guess.
On 5/17/07, Mike Tintner <[EMAIL PROTECTED]> wrote:
Pei: AI is about "what is the best solution to a problem if
> the system has INSUFFICIENT knowledge and resources".
Just so. I have just spent the last hour thinking about this area, and you
have spoken the line I allotted to you almost perfectly.
Your definition is a CONTRADICTION IN TERMS.
If a system has insufficient knowledge there is NO BEST, no "optimal,"
solution to any decision - that's a fiction.
If a system is uncertain, there is no certain way to deal with it.
If there is no right answer, there really is no right answer.
The whole of science has spent the last hundred years caught in the above
contradiction. Recognizing uncertainty and risk everywhere - and then
trying
to find a "certain," "right", "optimal", way to deal with them. This runs
right through science. "Oh yes, we can see that life is incredibly
problematic and uncertain... but this is the certain, best way to deal
with
it."
So science has developed games theory - arguably the most important theory
of behaviour - and then spent its energies trying to find the right way to
play games. The perfect equilibrium etc. And missed the whole point. There
is no right way to play games - on the few occasions that one is discovered,
(and there are a few occasions and situations), people STOP PLAYING IT -
because it ceases to be a proper game,and it ceases to be a model of real
life.In one sense, it's no wonder Nash went mad..
But scientists and techie's so badly want a right answer, they haven't been
able to admit there isn't one.
"What made [Stephen Jay] Gould unique as a scientist was that he had a
historian's mind and not an engineer's. He liked mess, confusion and
contradiction. Most scientists, in my experience, are the opposite. They are
engineers at heart. They think the world is made up of puzzles, and
somewhere out there is the one correct solution to every puzzle."
Andrew Brown
Hey, this is a fundamentally pluralistic world, not a [behaviourally]
monistic one. Deal with it:
Here's a simple problem for you with insufficient knowledge and resources::
you have $10,000 to invest tomorrow. You're thinking of investing in a
Chinese stockmarket mutual fund, because the market is on the up and you
reckon there could be a lot of money still to be made. (And there really
could). On the other hand, maybe it's a crazy bubble about to burst, and you
could lose a lot of money too. So what do you do tomorrow - buy or do
nothing - invest or keep your money in that savings account?
What's the "best" decision, or the best way of reaching a decision, or the
best way of finding a way of reaching a decision- in the next 24 hours (or,
in the end, in any time period)? [And what do you think, Ben?]
If you don't have the "best" answer, then your whole approach both to
defining and implementing intelligence is fundamentally flawed. A few
hundred million investors will be waiting to hear your reply. They'd love
to
know the best answer. No more need for all these different schools of
investors to argue so furiously, no more need for all these schools just of
investment AI/ computation alone to keep arguing either.Pei's cracked it,
guys. Over here.
You really would do well to think very long and hard about that simple
problem - it will change your life. I hope you will have the courage to
answer the problem.
(BTW MOST of the problems humans face in everyday life can be represented
as
investment problems - it's a basic, not an eccentric problem).
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?&
--
This message has been scanned for viruses and
dangerous content by MailScanner, and is
believed to be clean.
--
Josh Treadwell
[EMAIL PROTECTED]
480-206-3776
------------------------------------------------------------------------------
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?&
------------------------------------------------------------------------------
No virus found in this incoming message.
Checked by AVG Free Edition.
Version: 7.5.467 / Virus Database: 269.7.1/807 - Release Date: 16/05/2007
18:05
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=231415&user_secret=fabd7936