Ed,
Comments interspersed below:

Ed Porter wrote:

Colin,

Here are my comments re  the following parts of your below post:

=======Colin said======>

I merely point out that there are fundamental limits as to how computer science (CS) can inform/validate basic/physical science - (in an AGI context, brain science). Take the Baars/Franklin "IDA" project.... It predicts nothing neuroscience can poke a stick at...

===ED's reply===>

Different AGI models can have different degrees of correspondence to, and different explanatory relevance to, what is believed to take place in the brain. For example the Thomas Serre's PhD thesis "Learning a Dictionary of Shape-Components in Visual Cortex: Comparison with Neurons, Humans and Machines," at from http://cbcl.mit.edu/projects/cbcl/publications/ps/MIT-CSAIL-TR-2006-028.pdf , is a computer simulation which is rather similar to my concept of how a Novamente-like AGI could perform certain tasks in visual perception, and yet it is designed to model the human visual system to a considerable degree. It shows that a certain model of how Serre and Poggio think a certain aspect of the human brain works, does in fact work surprisingly well when simulated in a computer.

A surprisingly large number of brain science papers are based on computer simulations, many of which are substantially simplified models, but they do given neuroscientists a way to poke a stick at various theories they might have for how the brain operates at various levels of organization. Some of these papers are directly relevant to AGI. And some AGI papers are directly relevant to providing answers to certain brain science questions.

You are quite right! Realistic models can be quite informative and feed back - suggesting new empirical approaches. There can be great cross-fertilisation.

However the point is irrelevant to the discussion at hand.

The phrase "does in fact work surprisingly well when simulated in a computer" illustrates the confusion. 'work'? according to whom? "surprisingly well"? by what criteria? The tacit assumption is that the model's thus implemented on a computer will/can 'behave' indistinguishably from the real thing, when what you are observing is a model of the real thing, not the real thing.

*<HERE> *If you targeting AGI with a benchmark/target of human intellect or problem solving skills, then the claim made on any/all models is that models can attain that goal. A computer implements a model. To make a claim that a model completely captures the reality upon which it was based, you need to have a solid theory of the relationships between models and reality that is not wishful thinking or assumption, but solid science. Here's where you run into the problematic issue that basic physical sciences have with models. There's a boundary to cross - when you claim to have access to human level intellect - then you are demanding a equivalence with a real human, not a model of a human.

=======Colin said======>

I agree with your :

"/At the other end of things, physicists are increasingly viewing physical reality as a computation, and thus the science of computation (and communication which is a part of it), such as information theory, have begun to play an increasingly important role in the most basic of all sciences./"


===ED's reply===>

We are largely on the same page here

=======Colin said======>

I disagree with:

"But the brain is not part of an eternal verity. It is the result of the engineering of evolution. "

Unless I've missed something ... The natural evolutionary 'engineering' that has been going on has /not/ been the creation of a MODEL (aboutness) of things - the 'engineering' has evolved the construction of the /actual/ things. The two are not the same. The brain is indeed 'part of an eternal verity' - it is made of natural components operating in a fashion we attempt to model as 'laws of nature',,,

===ED's reply===>

If you define engineering as a process that involves designing something in the abstract --- i.e., in your "a MODEL (aboutness of things)" --- before physically building it, you could claim evolution is not engineering. But if you define engineering as the designing of things (by a process that has intelligence what ever method) to solve a set of problems or constraints, evolution does perform engineering, and the brain was formed by such engineering.

How can you claim the human brain is an eternal verity, since it is only believed that it has existing in anything close to its current form in the last 30 to 100 thousand years, and there is no guarantee how much longer it will continue to exists. Compared to much of what the natural sciences study, its existence appears quite fleeting.

I think this is just a terminology issue. The 'laws of nature' are the eternal verity, to me. The dynamical output they represent - of course that does whatever it does. The universe is an intrinsically dynamic entity at all levels. Even the persistent expression of total randomness is an 'eternal verity'. No real issue here.

=======Colin said======>

Anyway, for these reasons, folks who use computer models to study human brains/consciousness will encounter some difficulty justifying, to the basic physical sciences, claims made as to the equivalence of the model and reality. That difficulty is fundamental and cannot be 'believed away'.

===ED's reply===>

If you attend brain science lectures and read brain science literature, you will find that computer modeling is playing an ever increasing role in brain science --- so this basic difficulty that you describe largely does not exist.

I think you've missed the actual point at hand for the reasons detailed *<HERE>*.

=======Colin said======>

The intelligence originates in the brain. AGI and brain science must be literally joined at the hip or the AGI enterprise is arguably scientifically impoverished wishful thinking.

===ED's reply===>

I don't know what you mean by "joined at the hip," but I think it is being overly anthropomorphic to think an artificial mind has to slavishly model a human brain to have great power and worth. But I do think it would probably have to accomplish some of the same general functions, such as automatic pattern learning, credit assignment, attention control, etc.

Ed Porter

We are all enthusiastically intent on creating artificial entities with some kind of usefulness (=great power and worth). However, AGI is Artificial *General* Intelligence, seeks to create power and worth through a claim that '/general intelligence/' has been delivered. This is not merely the "same general functions"; it is actual general intelligence. The statement "A /model/ of general intelligence" is oxymoronic. If you can deliver general intelligence then you are not delivering a model of it, you are delivering /actual/ general intelligence. To use models as a basis for it you need to have a scientific basis for a claim that the models that have been used to implement the AGI can (in theory) deliver identical behaviour = general intelligence. Models of a human brain could be involved. Models of outward human behaviour could be involved. ... in any case - Each AGI-er needs to have a cogent, scientifically based claim in respect of the models as deliverers of the claimed outcomes - or the beliefs underlying the AGI-er's approach have a critical weakness in the eyes of science.

I don't think there's any real issue here. Mostly semantics being mixed a bit.

Gotta get back to xmas! Yuletide stuff to you.

Colin




-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=123753653-47f84b
Powered by Listbox: http://www.listbox.com

Reply via email to