Re: [agi] Do the inference rules.. P.S.Edward,

Thanks again for a detailed response (I really do appreciate it).

Your interesting examples of systems confirm my casual impressions of what can 
actually be done -  and my reluctance to shell out money on "Fluid Concepts." 
Inferences like MARY DOES OWN A BOOK and those from "IF I CHANGED "ABC TO ABD", 
HOW WOULD YOU (THE COPYCAT) MAKE AN ANALOGOUS CHANGE TO "MRRJJJ". strike me as 
fairly trivial, though by no means useless, and not really AGI . (Yes those are 
what I mean by "purely symbolic" systems, although I take your point that there 
are no absolute boundaries between different kinds of signs and particularly 
sign systems - even networks of symbols are used in complex ways that are not 
just symbolic).

Inferences like those you mention in:

IF YOU ASKED SUCH A SYSTEM WHAT LOVE BETWEEN A MAN AND A WOMAN WAS, IT WOULD BE 
ABLE TO GIVE YOU ALL SORTS OF MEANINGFUL GENERALIZATIONS ABOUT WHAT LOVE WAS, 
BASED ON ALL THE DESCRIPTIONS OF LOVE AND HOW IT MAKES CHARACTERS ACT IN THE 
BOOKS IT HAS READ.

might be v. productive and into AGI territory, but I note that you are talking 
hypothetically, not about real systems.

Ben, if you followed our exchange, has claimed a v. definite form of true AGI 
analogy - his system inferring from being able to "fetch", how to play 
hide-and-seek. I would like more explanation and evidence, though. But that's 
the sort of inference/analogy I think we should all be talking about.

Vis-a-vis neuroscience & what it tells us about what information is laid down 
in the brain, & in what form, I would be vastly more cautious than you. For 
instance, we see images as properly shaped, right? But the images on the retina 
have a severely distorted form - see Hawkin's photo in On Intelligence. So 
where in the brain or in the world is the properly shaped image? (The main 
point of that question is simply: hey, there's still masses we don't know - 
although if you have an answer, I'd be v. interested).

P.S. I hope you receive my privately emailed post with the definitions you 
requested.
  ----- Original Message ----- 
  From: Edward W. Porter 
  To: [email protected] 
  Sent: Friday, October 12, 2007 5:49 PM
  Subject: Re: [agi] Do the inference rules.. P.S.


  IN RESPONSE TO MIKE TINTNER'S Thu 10/11/2007 11:47 PM POST.  AGAIN MY 
RESPONSE IS IN BLUE ALL CAPS. 
  ================================================= 

  Edward, 
    
  Thanks for interesting info - but if I may press you once more. You talk of 
different systems, but you don't give one specific example of the kind of 
useful (& significant for AGI) inferences any of them can produce -as I do with 
my cat example. I'd especially like to hear of one or more from Novamente, or 
Copycat. 

  TO THE BEST OF MY KNOWLEDGE THERE IS NO AGI THAT CURRENTLY DOES ANYTHING 
CLOSE TO HUMAN LEVEL INFERENCING, IF THAT IS WHAT YOU MEAN.  NOVAMENTE IS THE 
CLOSEST THING I KNOW OF.  BUT, UNFORTUNATELY, AS OF THIS WRITING, I DON'T KNOW 
ENOUGH ABOUT IT TO KNOW EXACTLY HOW POWERFUL ITS CURRENT CAPABILITIES ARE.  

  BAYESIAN NETS ARE CURRENTLY USED TO DO A TON USEFUL INFERENCING OF A GENERAL 
TYPE THAT COULD BE VALUABLE TO AGI.  FOR EXAMPLE, A LOT OF COMPUTER-BASED 
DIAGNOSTIC SYSTEMS US IT.  BAYESIAN NETS HAVE SOME LIMITS, BUT THE LIMITS ARE 
BEING LOOSENED BY BRIGHT PEOPLE LIKE DAPHNE KOLLER (REALLY BRIGHT!).  I 
ATTENDED A LECTURE SHE GAVE AT MIT ABOUT A YEAR AND A HALF AGO IN WHICH SHE 
TALKED ABOUT HER GROUPS' WORK ON GETTING BAYSIAN NETS TO HANDLE RELATIONAL 
REASONING, SOMETHING THAT WOULD SUBSTAINTIALLY INCREASE THEIR POWER.  SHE HAS 
ALSO DONE WORK ON INTRODUCING BAYESIAN INFERENCING INTO FORMAL LOGICS.  

   
  SHRUITI (DESCRIBED IN "ADVANCES IN SHRUT -- A NEURALLY MOTIVATED MODEL ...', 
BY LOKENDRA SHASTRI  (A REALLY GREAT PIECE OF WORK)) PROVIDES A SYSTEM IN WHICH 
INFORMATION IS REPRESENTED IN A PREDICATE LOGIC FORM, IN GENERALIZATION 
HIERARCHIES, AND WITH A FORM OF PROBABILISTIC IMPLICATION.  THIS SYSTEM ANSWERS 
A QUESTION, SUCH AS "DOES MARY OWN A BOOK" BY APPROPRIATELY REMEMBERING 
(THROUGH PROBABILISTIC SPREADING ACTIVATION) THAT "JOHN GAVE MARY BOOK 17", 
THAT IF A DONOR GIVES A RECIPIENT SOMETHING, THE RECIPIENT WILL NORMALLY THEN 
OWN THAT SOMETHING, AND, THUS THAT MARY OWNS BOOK17, AND THAT BOOK 17 IS A 
BOOK, AND, THUS, FINALLY, THAT MARY DOES OWN A BOOK.  SHRUITI HAS LIMITS, BUT 
IT IS CAPABLE OF PERFORMING INTERESTING INFERENCES.

  Can you think of a single analogy or metaphor, in addition, that is purely 
symbolic? 

  LIKE BEN, I DON'T KNOW HOW YOU ARE USING YOUR TERMS.  FOR EXAMPLE, WHAT DO 
YOU MEAN BY "PURELY SYMBOLIC."  (A QUESTION I HAVE ASKED BEFORE.)

  COPYCAT MADE ANALOGIES BETWEEN STRINGS OF CHARACTERS, WHICH ARE SYMBOLIC.  AN 
EXAMPLE OF THE TYPE OF PROBLEM IT HANDLED IS:  IF I CHANGED "ABC TO ABD", HOW 
WOULD YOU (THE COPYCAT) MAKE AN ANALOGOUS CHANGE TO "MRRJJJ". THE COMPUTER 
PROGRAM CAME UP WITH MULTIPLE DIFFERENT ANSWERS.  ONE OF THEM WAS "MRRJJJ TO 
MRRJJJJ".  THE DESCRIPTION OF THE SYSTEM I KNOW IS IN CHAPTER 5 OF DOUGLAS 
HOFSTADTER'S FLUID CONCEPTS AND CREATIVE ANALOGIES, PUBLISHED IN 1995 BY BASIC 
BOOKS.

  THIS SYSTEM IS ARGUABLY "PURELY SYMBOLIC, BUT ITS ALGORITHM DOES EMBODY SOME 
LIMITED FORM SEMANTICS, SUCH THE RELATIONS OF INDIVIDUAL LETTERS TO 
ALPHABETICAL ORDER, NOTIONS OF INDIVIDUAL SIMPLE INTEGERS AND THEIR ORDERING, 
THE NOTION OF ORDERING IN GENERAL, THE NOTION OF BEFORE AND AFTER IN AN 
ORDERING, THE NOTION OF NEXT TO IN AN ORDERING INDEPENDENT OF DIRECTION, ETC.  
THIS KNOWLEDGE IS REPRESENTING IN A SLIPNET WHICH VARIABLY DEFINES MEASURE OF 
SIMILARITY BETWEEN ITS CONCEPTS.  

  DOES THAT MATCH YOUR DEFINITION OF PURELY SYMBOLIC? 

  I ASSUME NOVAMENTE WOULD BE QUITE CAPABLE OF MAKING ANALOGIES.  I ALSO ASSUME 
THERE HAS BEEN MUCH MORE WORK ON ANALOGY, BUT OFF THE TOP OF MY HEAD I DON'T 
KNOW OF IT, OTHER THAN NARS AND NOVAMENTE.

   
  I don't, and didn't, deny that logical thought is important. But it's only a 
small part, I'm arguing of most productive, AGI-type reasoning.

   
  IT DEPENDS ON WHAT YOU MEAN BY LOGICAL THOUGHT.  IF ALL REASONED INFERENCE IS 
LOGIC, THEN LOGIC WOULD BE QUITE IMPORTANT TO AGI-TYPE REASONING. EVEN REASONED 
INFERENCES FROM IMAGES COULD BE VIEWED AS LOGICAL THOUGHT. 

  A POWERFUL AGI WOULD PRESUMABLY INVOLVE MASSIVELY PARALLEL LOGICAL 
INFERENCING (AS INDICATED BY THE TERM "PROBABILISTIC LOGIC NETWORKS" FOR 
NOVAMENTE'S MAJOR INFERENCING MECHANISM) AND WOULD USE IT NOT ONLY FOR THINGS 
TRADITIONALLY CONSIDERED LOGICAL REASONING, BUT ALSO FOR MUCH MORE SUBTLE 
THINGS LIKE USING CONTEXT TO CHANGE THE PROBABLE INTERPRETATION OF A SENTENCE, 
SUBCONSCIOUS THOUGHT, AND INTUITIVE FEELINGS OR INSIGHTS.  THUS, LOGICAL 
INFERENCE IN AN AGI CAN BE USED FOR MUCH MORE THAN WHAT HAS TRADITIONALLY BEEN 
CONSIDERED "LOGIC" OR EVEN "REASONING", AND WOULD LIKELY BE A VITAL PART OF THE 
THINKING OF ANY POWERFUL AGI.

  Nor BTW are am I arguing  at all against symbols, (you might care to look at 
the "Picture Tree" thread I started a few months ago to better understand my 
thinking here) - the brain (and any true AGI, I believe) uses symbols, outline 
graphics [or Johnson's image schemata] and images in parallel, interdependently 
and continuously, to reason about the world. (Note: "continuously." You seem to 
think that some occasional sensory grounding of an AGI system here and there 
will do. No, I'm arguing, it has to be, and is, continuous and applied to all 
information and subjects).

   
  I THINK IT IS OBVIOUS THAT CERTAIN TYPES OF LEARNING REQUIRE REASONABLY HIGH 
TEMPORAL RESOLUTION IN THE REPRESENTATIONS THEY LEARN FROM.  A PRIME EXAMPLE 
WOULD BE , HUMAN MOTOR CONTROL.  

  What I am arguing against, rather than symbols,  is what you might call the 
"bookroom illusion" - which you saw graphically illustrated in John Rose's post 
- the illusion that you can "learn about the world just from books" - or, to be 
precise, that you can learn, and think about and build models of the world with 
symbols/ text alone. It's an understandable illusion given that we often  spend 
hours apparently doing nothing but read text. But it is an illusion. The brain 
does, and has to, continuously make sense (in images) of everything we read.And 
if it can't then that text won't make sense - & it's a case of "I can't see 
what you are talking about."

  (1) WITH REGARD TO WHETHER AN AGI HAS TO MAKE SENSE USING IMAGES -- I DON'T 
KNOW HOW YOU ARE DEFINING IMAGES? 

  IT IS NOT CLEAR TO ME THAT ALL POWERFUL, CREATIVE, ADAPTIVE, THINKING, 
INTUITIVE AGI'S REQUIRE 2- OR 3-D SPATIAL MODELS REPRESENTING SENSATIONS 
OBTAINED BY VISION, TOUCH, OR STEREOPHONIC SPATIAL SENSING.  

  AGAIN, FOR EXAMPLE, I REFER TO THE TYPE OF AGI REFERRED TO IN MY Thu 
10/11/2007 8:14 AM POST, AN AGI WHOSE WORLD IS LIMITED TO A PROGRAMMING 
LANGUAGE, PROGRAMS IT HAS CREATED IN THAT LANGUAGE, THE RESULTING OUTPUT IN ITS 
WORKSPACE FROM THAT PROGRAM, ITS OBSERVATIONS OF THE CHANGES TO THE WORKSPACE 
MADE BY ITS VARIOUS PROGRAMS, AND HOW WELL THE CHANGES SATISFY ITS GOALS AND 
VALUES.

  I SPEND MUCH OF MY THOUGHT ABOUT AGI THINKING ABOUT REPRESENTATION IN WHAT I 
CALL SEMANTIC HYPERSPACE.  THIS IS THE SPACE OF PATTERNS AND RELATIONSHIPS 
BETWEEN PATTERNS, INCLUDING TEMPORAL RELATIONS.  IT IS A HYPERSPACE BECAUSE IT 
CAN CONTAIN MILLIONS OR BILLIONS OF PATTERNS, EACH OF WHICH IS A POTENTIAL 
DIMENSION, AND EACH OF ITS SUCCESSION OF ACTIVATION STATES HAS A POTENTIAL 
COMBINATORIAL EXPLOSION OF POSSIBLE STATES, AND THE SPACE DEFINED BY A SEQUENCE 
OF JUST TWO SUCH STATES IS A COMBINATORIAL EXPLOSION TIMES A COMBINATORIAL 
EXPLOSION.  SO WE ARE NOT TALKING ABOUT ANYTHING EVEN CLOSE TO A 2- OR 
3-DIMENSIONAL SPACE, NOR ANYTHING EVEN CLOSE TO 2-D IMAGE.

  EVEN SHRUITI (CITED ABOVE), WHEN SOLVING THE "DOES MARY OWN A BOOK" PROBLEM 
CREATES A TEMPORARY NETWORKED ACTIVATION STATE IN A SIMPLE SEMANTIC SPACE.  I 
AM ATTACHING A PDF OF A MARK-UP I MADE FROM THE ABOVE QUOTED ARTICLE ABOUT 
SHRUITI.  THE MARK-UP BETTER ILLUSTRATES THE ACTIVATION NET IT CREATES IN 
SOLVING THIS PROBLEM.  

  WOULD THE NETWORKED REPRESENTATIONS AN AGI MAKES IN SUCH A SEMANTIC 
HYPERSPACE BE "IMAGES" OF THE TYPE YOU ARE ADVOCATING?  IF SO, IMAGES ARE 
IMPORTANT TO MY VISION OF AGI.  BUT SUCH IMAGES CAN BE CREATED AND REPRESENTED 
IN "LOGICAL" SYSTEMS, SUCH AS SHRUITI.

  (2)  WITH REGARD TO BOOKWORLD -- IF ALL THE WORLD'S BOOKS WERE IN ELECTRONIC 
FORM AND YOU HAD A MASSIVE AMOUNT OF AGI HARDWARD TO READ THEM ALL I THINK YOU 
WOULD BE ABLE TO GAIN A TREMENDOUS AMOUNT OF WORLD KNOWLEDGE FROM THEM, AND 
THAT SUCH WORLD KNOWLEDGE WOULD PROVIDE A SURPRISING AMOUNT OF GROUNDING AND BE 
QUITE USEFUL.  

  BUT, AS I SAID IN MY 10/11/2007 7:33 PM  POSTING, I THINK THERE WOULD 
PROBABLY BE SOME PRETTY BIG GAPS IN ITS UNDERSTANDING. SUCH A SYSTEM WOULD 
PROBABLY BE BRILLIANT IN SOME WAYS AND REALLY DUMB IN OTHERS. 

  WHAT YOU SEEM TO FAIL TO UNDERSTAND IS THAT A REALLY POWERFUL BRAIN HAS AN 
ABILITY TO CREATE A SENSE OF REALITY OUT OF BITS AND BYTES, IF THOSE BITS AND 
BYTES ARE PROPERLY GROUNDED IN A SET OF COHERENT RELATIONSHIPS.  WE ARE NOT 
CONSCIOUS OF THE WORLD DIRECTLY, INSTEAD WE ARE CONSCIOUS OF MENTAL CONSTRUCTS 
THAT HAVE BEEN CREATED BY GENERALIZATIONS OUT OF SENSORY DATA. .

  FOR EXAMPLE, BECAUSE OF  FOVIATION AND CONSTANT SACCADES, OUR EYES AND V1 SEE 
WITH SUCH DISCONTINUOUS FISHEYED VISION THAT IF WE WERE TO FORCED TO WATCH A TV 
IMAGE OF WHAT WAS PROJECTED V1 NOT ONLY WOULD WE PROBABLY NOT BE ABLE TO 
RECOGNIZED MOST OF IT, BUT IF IT WERE ON A LARGE SCREEN TV WE MIGHT PUKE.  YET 
BECAUSE OF THE CORRELATIONS BETWEEN ALL OF THE DIFFERENTLY SHAPED PATTERNS 
CREATED BY GIVEN SHAPE WHEN SEEN IN DIFFERENT PARTS OF THE FOVIATED FIELD OF 
VIEW, OUR BRAIN HAS LEARNED A RELATIVELY INVARIANT REPRESENTATION OF THAT SHAPE 
AND THAT INVARIANT SHAPE IS WHAT WE THINK WE "SEE" EVEN THOUGH THAT IS NOT WHAT 
IS BEING PROJECT ON V1.  SIMILARLY OUR BRAIN STITCHES TOGETHER THE MULTIPLE 
VIEWS CREATED BY OUR RAPIDLY SACCADING EYES INTO A SENSE OF A VISUALLY 
CONTINUOUS SPACE (A TRICK MADE DOUBLY HARD BY FOVIATION.)

  THUS, OUR SENSE OF REALITY INCLUDES ALL SORTS OF MENTAL FABRICATIONS, CREATED 
AS REASONABLE REPRESENTATIONS OF RELATIONSHIPS CONTAINED IN THE SENSORY DATA WE 
RECEIVE.  WE HAVE A NOTION THAT THE TOP OF A TABLE IS CONTINUOUS AND SOLID, YET 
FROM CHEMISTRY AND PHYSICS WE KNOW IT IS NOT.

  IN FACT, CURRENT BRAIN SCIENCE INDICATES WE DON'T STORE PICTURES IN ANYTHING 
LIKE THE FORM OF A PHOTOGRAPH OR A LINE DRAWING.  INSTEAD WE NORMALLY STORE A 
NETWORK OF ONE OR MORE NODES FROM A GEN/COMP HIEARARCHY, EACH OF WHICH MAPS TO 
MULTIPLE POSSIBLE LOWER LEVEL REPRESENTATIONS UNTIL YOU GET DOWN TO THE 
EQUIVALENT OF THE PIXEL LEVEL.  IT IS GENERALLY BELIEVED THERE IS NO ONE NODE 
THAT STORES A PARTICULAR IMAGE.  

  SO EVEN OUR MEMORIES OF THE PICTURES YOU CONSIDER SO IMPORTANT ARE SYMBOLIC, 
IN THAT THEY ARE MADE UP OF NODES THAT SYMBOLIZE PATTERNS OF OTHER NODES.

  SO GETTING BACK TO BOOKWORLD, WHAT I AM TRYING TO SAY IS THAT JUST AS OUR 
MINDS FABRICATE CONCEPTS OF "PHYSICAL REALITY" BASED ON CORRELATIONS AND 
RELATIONS WITHIN A HUGE AMOUNT OF DATA, AN EXTREMELY POWERFUL AGI THAT HAD A 
REASONABLE DEEP STRUCTURE REPRESENTATION OF ALL CURRENTLY EXISTING BOOKS WOULD 
SIMILARLY HAVE FABRICATED CONCEPTS OF A "BOOK-WORLD REALITY", AND THAT SUCH 
CONCEPTS WOULD BE WELL GROUNDED IN THE SENSE THAT THEY WOULD BE CONNECTED BY 
MANY RELATIONSHIPS, ORDERINGS, GENERALIZATIONS, AND BEHAVIORS.

  I DON'T REALLY KNOW EXACTLY HOW MUCH KNOWLEDGE COULD BE EXTRACTED FROM 
BOOKWORLD.  I KNOW THAT LINGUISTS PLAYING WITH ROUGHLY 1G WORD TEXT CORPORA 
BITCH ABOUT HOW SPARSE THE DATA IS.  BUT MY HUNCH IS THAT IF YOU READ SAY 30 
MILLION BOOKS AND THE WEB WITH A GOOD AGI YOU WOULD BE ABLE TO LEARN A LOT. 

  IF YOU ASKED THE BOOKWORLD AGI WHAT HAPPENS WHEN A PERSON DROPS SOMETHING, IT 
WOULD PROBABLY BE ABLE TO GUESS IT OFTEN FALLS TO THE GROUND, AND THAT IF IT IS 
MAKE OF GLASS IT MIGHT BREAK.

  IF YOU ASKED SUCH A SYSTEM WHAT LOVE BETWEEN A MAN AND A WOMAN WAS, IT WOULD 
BE ABLE TO GIVE YOU ALL SORTS OF MEANINGFUL GENERALIZATIONS ABOUT WHAT LOVE 
WAS, BASED ON ALL THE DESCRIPTIONS OF LOVE AND HOW IT MAKES CHARACTERS ACT IN 
THE BOOKS IT HAS READ.  I WOULD NOT BE SURPRISED IF SUCH A SYSTEM UPON READING 
A ROMANTIC NOVEL WOULD PROBABLY HAVE ABOUT AS GOOD A CHANCE AS THE AVERAGE 
HUMAN READER OF PREDICTING WHETHER TO THE TWO LOVERS WILL OR WILL NOT BE 
TOGETHER AT THE END OF THE NOVEL.  

  IF YOU ASKED IT ABOUT HOW PEOPLE MENTALLY ADJUST TO GROWING OLD, IT WOULD 
PROBABLY BE ABLE TO GENERATE A MORE THOUGHTFUL ANSWER THAN MOST YOUNG HUMAN 
BEINGS.

  IN SHORT, IT IS MY HUNCH THAT A POWERFUL BOOKWORLD AGI COULD BE EXTREMELY 
VALUABLE.  AND AS I SAID IN MY Thu 10/11/2007 7:33 PM POST, THERE IS NO REASON 
WHY KNOWLEDGE LEARNED FROM BOOKWORLD COULD NOT BE COMBINED KNOWLEDGE LEARNED BY 
OTHER MEANS, INCLUDING THE IMAGE SEQUENCES YOU ARE SO FOND OF. 





  Edward W. Porter 
  Porter & Associates 
  24 String Bridge S12 
  Exeter, NH 03833 
  (617) 494-1722 
  Fax (617) 494-1822 
  [EMAIL PROTECTED] 


------------------------------------------------------------------------------
  This list is sponsored by AGIRI: http://www.agiri.org/email
  To unsubscribe or change your options, please go to:
  http://v2.listbox.com/member/?&; 


------------------------------------------------------------------------------


  No virus found in this incoming message.
  Checked by AVG Free Edition. 
  Version: 7.5.488 / Virus Database: 269.14.8/1064 - Release Date: 11/10/2007 
15:09

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=53192154-94404e

Reply via email to