Well, as one ignoramus speaking to another (hopefully the smart ones on the
list will correct us) I think not. It's not the random inputs (no intelligence
or complex system can deal with randomness and turn it into something
meaningful - just like random walk share prices would mean you cannot
consistently beat the stock market :) that make a system complex. It is at
least its structure and internal diversity as well as challenging environment.
Like so much, true complexity (both behaviour or structure) is likely to exist
on/walk a very fine line between 'boring' and 'random' (see Wolfram et al or
anything fractal) which means randomly linking up modules is IMHO not likely to
give rise to AGI (one of my criticisms against traditional connectionism) and
completely random environments are also not likely to give rise to AGI either.
You need a well-configured (through evolution - Ben?, design - my approach - or
experimentation/systematic exploration - Richard's) modular system
working/growing/learning/evolving in a rich i.e. not boring but not random
either environment. Randomness just produces static or noise (i.e. more
randomness). {The parallel with data is too obvious not to point out: neither
completely random data nor highly regular (e.g. infinitely repeating) data
contain much information, the interesting stuff is inbetween those two extremes}
So there is no(t necessarily any) complexity hiding in 'difficult algorithms',
'complex mathematics', 'random data', 'large datasets', etc. Solving a system
of 10000000 linear equations is simple, solving two or three quadratic
differential equations is complex. A map-reduce (assuming a straightforward
transform function) of 20TB of data may require lots of computing power but is
far less complex than running ALife on your PDA.
The reason why I responded to this post is that my AGI architecture is relying
very much on the emerging complexity arising from a "moderately massively
modular" design which seems to be queried by many on this list but one of the
more mainstream hypotheses (tho not necessarily the dominant one) in CogSci (eg
Carruthers). (Aslo note that for CogScientists 'massive modular' is quite
significantly less massive than what it means to CompScientists :)
=Jean-Paul
Research Associate: CITANDA
Post-Graduate Section Head
Department of Information Systems
Phone: (+27)-(0)21-6504256
Fax: (+27)-(0)21-6502280
Office: Leslie Commerce 4.21
>>> "Mike Tintner" <[EMAIL PROTECTED]> 2007/12/06 14:05 >>>
JVPB:You seem to have missed what many A(G)I people (Ben, Richard, etc.)
mean by 'complexity' (as opposed to the common usage of complex meaning
difficult).
Well, I as an ignoramus, was wondering about this - so thankyou. And it
wasn't clear at all to me from Richard's paper what he meant. What I'm
taking out from your account is that it involves random inputs...? Is there
a fuller account of it? Is it the random dimension that he/others hope will
produce emergent/human-like behaviour? (..because if so, I'd disagree - I'd
argue the complications of human behaviour flow from conflict/ conflicting
goals - which happens to be signally missing from his (and cognitive
science's) ideas about emotions).
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?&
-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=73141768-c30744