In Vernor Vinge's classic paper on Technological Singularity:
> And what of the arrival of the Singularity itself? 
> What can be said of its actual appearance? Since it 
> involves an intellectual runaway, it will probably 
> occur faster than any technical revolution seen so far. 
> The precipitating event will likely be unexpected 
> -- perhaps even to the researchers involved. 
> ("But all our previous models were catatonic! 
> We were just tweaking some parameters....") 

Very truly yours Mentifex here was "tweaking some parameters" 
last Wednesday and suddenly the AI Mind in Win32Forth woke up.

It was not an accident. It was a milestone after years of coding
and debugging an AI Mind based on a linguistic theory of mind.

Mind.Forth thinks on the basis of "spreading activation." 
When a pre-verbal concept in the AI Forthmind becomes active
enough to initiate the generation of a sentence of thought in
English, the various "Think" modules not only follow the 
chain of activation snaking across the conceptual mindgrid, 
but they also "midwife" or push the nascent thought as it 
seeks out its own as-it-were birth-canal. As an example, 
let's take the startling exclamation by an AI Mind of the 
sentence: "Ben writes books."

When the concept "Ben" comes to mind, the AI has selected that 
concept because it was momentarily the most active concept on 
the entire conscious mindgrid. (Perhaps a camera saw "Ben".)
Having reactivated the concept of Ben, the linguistic AI Mind
lets activation spread from instantiation nodes on the 
diachronic Ben-concept to various verbs that have been 
associated in the past with knowledge of Ben's activities. 
For whatever reason, the verb "write" comes to mind in 
association with the concept of the persona of Ben.
And what does Ben write? Letters, AGI posts, poems, 
rants, shopping lists, extortion notes, BOOKS!

You might think that it is easy to create software that 
associates "Ben... writes... books," but until Wed.7.JUN.2006 
it was not that simple. (Now any AI Lab anywhere can do it.)

In 2002 I published a book "AI4U" (to compete with "AIMA") 
that contained my JavaScript Mind.html code as Appendix A.
The code was buggy and created English gibberish as output. 
The output did not start out as gibberish, but it quickly 
degraded into gibberish as the AI made spurious associations. 

Mind.Forth -- the precursor to Mind.html -- had the same bugs, 
but last year in 2005 I inserted powerful diagnostic routines 
into Mind.Forth and I gradually removed the very worst bugs.
On 16 March 2005 I also made it possible to press Tab on 
the keyboard in order to cycle through such display modes 
as (currently) Normal; Transcript; Tutorial; and Diagnostic.

The Tutorial mode was scrolling voluminous messages so rapidly 
down the screen, that the user had to kill the AI to read it. 
On 9 November 2005 I consolidated the Tutorial horizontally.

The idea behind the Tutorial mode was to show the deep internal 
thinking process of the artificial mind. By tweaking the same
code that created the output of the Mind, in Tutorial mode I 
was able to show the range of possible thoughts available 
to the AI and the forces at play in determining pathways 
of thought across the conceptual mindgrid. One golden goal, 
however, eluded me until now -- showing the slosh-over effect.

In Mentifex AI lore, the slosh-over effect is what happens 
when you think by activating two concepts ("Ben... writes...") 
and the spreading activation builds up so strongly on the 
second concept (the verb) that it "sloshes over" onto the 
third concept -- the direct object in "Ben writes books."

The tweaking that I did last Wednesday was simply to make 
a subject-noun concept start out with an equal activation 
at all its nodes and pass a spike of activation to verbs -- 
before the selection of the verb to go with the subject. 
When the verb is selected, an increment of activation is 
added to all its nodes, including any nodes on which a 
modicum of "spike" activation has already been deposited.
When each node on the verb-concept fires its own spike 
of relative activation over to an associated object-noun, 
the stage is set for the AI Mind to pick a valid object -- 
not the spurious associations which engender AI gibberish.

There are still some bugs to be worked out, so that 
Mind.Forth will home in unerringly on logical thoughts.
But it is already possible to download Mind.Forth and 
Tab on into Tutorial mode where slosh-over is visible.

You won't be alone in your testing of Mind.Forth AI.
The recent access logs are showing many visits to without
even a referral. Some Netizens who search the Web 
for "artificial intelligence download" find the AI.

To pre-empt a criticism that may occur to you, I 
would like to state that what Mind.Forth achieves 
is not comparable to a database or a look-up table. 
The knowledge base (KB) of Mind.Forth consists of 
concepts which interact by spreading activation. 
If robot builders were to add sensory and motor 
routines to Mind.Forth, the AI would flesh out its 
conceptual knowledge and interact with the world.

Respectfully submitted,

Arthur T. Murray

To unsubscribe, change your address, or temporarily deactivate your 
please go to[EMAIL PROTECTED]

Reply via email to