> #####ED PORTERS CURRENT RESPONSE ########>
> Forward and backward chaining are not hacks. They has been two of the most
> commonly and often successfully techniques in AI search for at least 30
> years. They are not some sort of wave of the hand. They are much more
> concretely grounded in successful AI experience than many of your much more
> ethereal, and very arguably hand waving, statements about having many of the
> difficult problems in AI are to be cured by some as yet unclearly defined
> emergence from complexity.
Richard Loosemore's response:
Oh dear: yet again I have to turn a blind eye to the ad hominem insults.
----------------------------------------------
There were no ad hominem insults in Ed's response. His comment about Richard's
ethereal hand waiving was clearly and unmistakably within the boundaries that
Richard has set in his own criticisms again and again. And Ed specified the
target of the criticism when he spoke of the "difficult problems in AI
...[which]... are to be cured by some as yet unclearly defined emergence from
complexity." All Richard had to do was to answer the question, and instead he
ran for cover behind this bogus charge of being the victim of an ad hominem
insult.
If upon reflection, Richard sincerely believes that Ed's comment was an ad
hominem insult, then we can take this comment as a basis for detecting the true
motivation behind those comments of Richard which are so similar in form.
For example, Richard said, " Understanding that they only have the status of
hacks is a very important sign of maturity as an AI researcher. There is a
very deep truth buried in that fact."
While I have some partial agreement with Richard's side on this one particular
statement, I can only conclude that by using Richard's own measure of "ad
hominem insults" that Richard must have intended this remark to have that kind
of effect. Similarly, I feel comfortable with the conclusion that every time
Richard uses his "hand waiving" argument, there is a good chance that he is
just using it as an all-purpose ad hominem insult.
It is too bad that Richard cannot discuss his complexity theory without running
from the fact that his solution to the problem is based on his non-explanation
that,
"...in this "emergent" (or, to be precise, "complex system") answer to
the question, there is no guarantee that binding will happen. The
binding problem in effect disappears - it does not need to be explicitly
solved because it simply never arises. There is no specific mechanism
designed to construct bindings (although there are lots of small
mechanisms that enforce constraints), there is only a general style of
computation, which is the relaxation-of-constraints style."
>From reading Richard's postings I think that Richard does not believe there is
>a problem because the nature of complexity itself will solve the problem -
>once someone is lucky enough to find the right combination of initial rules.
For those who believe that problems are solved through study and
experimentation, Richard has no response to the most difficult problems in
contemporary AI research except to cry foul. He does not even consider such
questions to be valid.
Jim Bromer
-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
https://www.listbox.com/member/?member_id=8660244&id_secret=106510220-47b225
Powered by Listbox: http://www.listbox.com