Hey, Ben. Arthur here. Remember the prediction of
the 2012 Singularity? Well, guess what happened.
I just had the following conversation with my
http://www.scn.org/~mentifex/mindforth.txt AI:
Human: boys make cars
Robot: CARS WHAT ARE CARS
Human: boys make guns
Robot: GUNS WHAT ARE GUNS
Human: boys
Robot: THE BOYS MAKE THE CARS
Human: boys
Robot: THE BOYS MAKE THE GUNS
Chief AGI guru Dr. Goertzel! The above is not
a cherry-picked, post-mucho experimentation
routine test result put out for PR purposes.
It just happened during hard-core AI coding.
Now, before everybody jumps in and snipes,
"Aw, there's that crackpot Mentifex again,"
let me explain the significance of the I/O.
I have been coding MindForth AI for twelve
years, since 1998 and maybe even earlier.
Today on Mon.6.SEP.2010 for the first time
I introduced quasi-neuronal inhibition into
the free open-source AI source code. Why?
A year ago, around August or September of
2009, you and I had our set-to (rumble?)
concerning the AGI Roadmap and my posts
there which were deleted ("rolled back")
by Itamar Arel. No biggy. I did not fix
Itamar's wagon last Halloween, so I won't
fix it this Halloween, either. You see, I
was maintaining my own AI Roadmap at
http://code.google.com/p/mindforth/wiki/RoadMap
concurrently with my contributions to
you guys' Roadmap.
The main thing is, I was entering into
the Roadmap Milestone of trying to achieve
"self-referential thought" with my AI.
That particular achievement requires
covering a lot of ground, not just
"you and I" interactions between the
human user and the artificial AI Mind.
The AI needs to acquire a general knowledge
of the surrounding world, so that man and
machine may discuss the AI as a participant
in its world.
So at the end of 2009 I was coding the
ability of the AI to respond to who-queries
and what-queries, so that the AI can deal
with questions like "Who are you?" and
"What are you?"
Recently I have perceived the need to
get the AI to respond with multiple answers
to queries about topics where the AI knows
not a single fact but multiple facts,
such as, "What do robots make?" I want
the AI to be able to say such things as:
"Robots make cars."
"Robots make tools."
"Robots make parts."
"Robots make robots."
It dawned on me a few days ago that the
AI software would have to suppress each
given answer in order to move on to the
next answer available in the knowledge
base (KB). In other words, for the first
time ever, I had to code _inhibition_
into the AI Mind. Tonight I have done
so, and that simple conversation near the
top of this message shows the results.
The same query, of just the word "boys...",
elicits two different answers from the KB
because each response from the AI goes
immediately into inhibition in such a way
as to allow access to the next fact
queued up in the recesses of the AI KB.
This "Singularity Alert" from Mentifex
may generate a collective "Huh?" from
the list readership, but here it is.
Bye for now (and back to the salt mines :-)
Arthur
--
http://AiMind-i.com
http://code.google.com/p/mindforth
http://doi.acm.org/10.1145/307824.307853
http://robots.net/person/AI4U/diary/40.html
-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription:
https://www.listbox.com/member/?member_id=8660244&id_secret=8660244-6e7fb59c
Powered by Listbox: http://www.listbox.com