Jim,

You are right to call me on that. I need to provide an argument that,
if no logic satisfying B exists, human-level AGI is impossible.

B1: A foundational logic for a human-level intelligence should be
capable of expressing any concept that a human can meaningfully
express.

If a broad enough interpretation of the word "logic" is taken, this
statement is obvious; it could amount to simply "A human level
intelligence should be capable of expressing anything it can
meaningfully express". (ie, logic = way of operating.) So, with this
interpretation, it doesn't even make sense for B to be false. But,
this is not quite what I mean.

The key idea for me is that logic is not the way we *do* think, it is
the way we *should* think, in the ideal situation of infinite
computational resources. So, a more refined B would state:

B2: The theoretical ideal of how a human-level intelligence should
think, should capture everything worth capturing about the way humans
actually do think.

"Everything worth capturing" means everything that could lead to good results.

So, I argue, if no logic exists satisfying B2, then human-level
artificial intelligence is not possible. In fact, I think the negation
of B2 is nonsensical:

not-B2: There is no concept of how a human-level intelligence should
think that captures everything worth capturing about how humans do
think.

This seems to imply that humans do not exist, since the way humans
actually *do* think captures everything worth capturing (as well as
some things not worth capturing) about how we think.

-Abram

On Thu, Aug 14, 2008 at 2:04 PM, Jim Bromer <[EMAIL PROTECTED]> wrote:
> On Thu, Aug 14, 2008 at 12:59 PM, Abram Demski <[EMAIL PROTECTED]> wrote:
>> A more worrisome problem is that B may be contradictory in and of
>> itself. If (1) I can as a human meaningfully explain logical system X,
>> and (2) logical system X can meaningfully explain anything that humans
>> can, then (3) system X can meaningfully explain itself. Tarski's
>> Indefineability Theorem shows that any such system (under some
>> seemingly reasonable assumptions) can express the concept "This
>> concept is false", and is therefore (again under some seemingly
>> reasonable assumptions) contradictory. So, if we accept those
>> "seemingly reasonable assumptions", no logic satisfying B exists.
>>
>> But, this implies that AI is impossible.
>
> At risk of being really annoying I have to say: it does not imply
> anything of the sort!  How could it imply some idea if it can't even
> represent it?  It implies impossibility to you because you are capable
> of dealing with fictions of boundaries as if they were real until you
> either conflate two or more bounded concepts, or simply transcend them
> by virtue of the extent of your everyday experience.
>
> In order to deal with logic you have to be taught how to consider such
> a thing to be bounded from the rest of reality.  That is not difficult
> because that is a requirement of all thought and it is a fundamental
> necessity of dealing with the real universe.  To study anything you
> have to limit your attention to the subject.
>
> I believe we can use logic in AI to detect possible errors and
> boundary issues.  But then we have to build up a system of knowledge
> from experience which seems to transcend the usual boundaries when
> that kind of transcendent insight becomes useful. After a while
> transcendent insight is itself recognized to be bounded and so it
> becomes part of the mundane.
>
> Think of this way, boundaries are primarily dependent on limitations.
> Jim Bromer
>
>
>
>
>
>
> On Thu, Aug 14, 2008 at 12:59 PM, Abram Demski <[EMAIL PROTECTED]> wrote:
>> This looks like it could be an interesting thread.
>>
>> However, I disagree with your distinction between ad hoc and post hoc.
>> The programmer may see things from the high-level "maze" view, but the
>> program itself typically deals with the "mess". So, I don't think
>> there is a real distinction to be made between post-hoc AI systems and
>> ad-hoc ones.
>>
>> When we decide on the knowledge representation, we predefine the space
>> of solutions that the AI can find. This cannot be avoided. The space
>> can be made wider by restricting the knowledge representation less
>> (for example, allowing the AI to create arbitrary assembly-language
>> programs is less of a restriction than requiring it to learn
>> production-rule programs that get executed by some implementation of
>> the rete algorithm). But obviously we run into hardware restrictions.
>> The broadest space to search is the space of all possible
>> configurations of 1s and 0s inside the computer we're using. An AI
>> method called "Godel Machines" is supposed to do that. William Pearson
>> is also interested in this.
>>
>> Since we're doing philosophy here, I'll take a philosophical stance.
>> Here are my assumptions.
>>
>> 0. I assume that there is some proper logic.
>>
>> 1. I assume probability theory and utility theory, acting upon
>> statements in this logic, are good descriptions of the ideal
>> decision-making process (if we do not need to worry about
>> computational resources).
>>
>> 2. I assume that there is some reasonable bayesian prior over the
>> logic, and therefore (given #1) that bayesian updating is the ideal
>> learning method (again given infinite computation).
>>
>> This philosophy is not exactly the one you outlined as the AI/AGI
>> standard: there is no searching. #2 should ideally be carried out by
>> computing the probability of *all* models. With finite computational
>> resources, this is typically approximated by searching for
>> high-probability models, which works well because the low-probability
>> models contribute little to the decision-making process in most cases.
>>
>> Now, to do some philosophy on my assumptions :).
>>
>> Consideration of #0:
>>
>> This is my chief concern. We must find the proper logic. This is very
>> close to your concern, because the search space is determined by the
>> logic we choose (given the above-mentioned approximation to #2). If
>> you think the search space is too restricted, then essentially you are
>> saying we need a broader logic. My requirements for the logic are:
>> A. The logic should be grounded
>> B. The logic should be able to say any meaningful thing a human can say
>> The two requirements are not jointly satisfied by any existing logic
>> (using my personal definition of grounded, at least). Set theory is
>> the broadest logic typically considered, so it comes closest to B, but
>> it (and most other logics considered strong enough to serve as a
>> foundation of mathematics) do not pass the test of A because the
>> manipulation rules do not match up to their semantics. I explain this
>> at some length here:
>>
>> http://groups.google.com/group/opencog/browse_thread/thread/28755f668e2d4267/10245c1d4b3984ca?lnk=gst&q=abramdemski#10245c1d4b3984ca
>>
>> A more worrisome problem is that B may be contradictory in and of
>> itself. If (1) I can as a human meaningfully explain logical system X,
>> and (2) logical system X can meaningfully explain anything that humans
>> can, then (3) system X can meaningfully explain itself. Tarski's
>> Indefineability Theorem shows that any such system (under some
>> seemingly reasonable assumptions) can express the concept "This
>> concept is false", and is therefore (again under some seemingly
>> reasonable assumptions) contradictory. So, if we accept those
>> "seemingly reasonable assumptions", no logic satisfying B exists.
>>
>> But, this implies that AI is impossible. So, some of the seemingly
>> reasonable assumptions need to be dismissed. (But I don't know which
>> ones.)
>>
>> Consideration of #2:
>>
>> Assumption 3 is that there exists some reasonable prior probability
>> distribution that we can use for learning. A now-common way of
>> choosing this prior is the minimum description length principle, which
>> tells us that shorter theories are more probable.
>>
>> The following argument was sent to me by private email by Wei Dai, and
>> I think it is very revealing:
>>
>> "I did suggest a prior based on set theory, but then I realized that
>> it doesn't really solve the entire problem. The real problem seems to
>> be that if we formalize induction as Bayesian sequence prediction with
>> a well-defined prior, we can immediately produce a sequence that an
>> ideal predictor should be able to predict, but this one doesn't, no
>> matter what the prior is. Specifically, the sequence is the "least
>> expected" sequence of the predictor. We generate each symbol in this
>> sequence by feeding the previous symbols to the predictor and then
>> pick the next symbol as the one that it predicts with the smallest
>> probability. (Pick the lexicographically first symbol if more than one
>> has the smallest probability.)
>>
>> This least expected sequence has a simple description, and therefore
>> should not be the least expected, right? Why should it not be more
>> expected than a completely random sequence, for example?
>>
>> Do you think your approach can avoid this problem?"
>>
>> Again, if the argument is right, then AI is impossible (or, as Wei Dai
>> put it, induction cannot be formalized). So, I again conclude that
>> some assumption needs to be dismissed. In this case the most obvious
>> assumption is that the prior should be based on minimum description
>> length. I thought at the time that that was the wrong one, but I am
>> not sure at the moment.
>>
>> So, any ideas how to resolve these two problems?
>>
>> --Abram
>>
>>
>>
>> PS-- I don't want to leave the quote from Wei Dai completely out of
>> context, so here is Wei Dai's full argument:
>>
>> http://groups.google.com/group/everything-list/browse_frm/thread/c7442c13ff1396ec/804e134c70d4a203
>>
>>
>> On Wed, Aug 13, 2008 at 10:15 AM, Mike Tintner <[EMAIL PROTECTED]> wrote:
>>> THE POINT OF PHILOSOPHY:  There seemed to be some confusion re this - the
>>> main point of philosophy is that it makes us aware of the frameworks that
>>> are brought to bear on any subject, from sci to tech to business to arts -
>>> and therefore the limitations of those frameworks. Crudely, it says: hey
>>> you're looking in 2D, you could be loooking in 3D or nD.
>>>
>>> Classic example: Kuhn. Hey, he said, we've thought science discovers bodies
>>> feature-by-feature, with a steady-accumulation-of-facts. Actually those
>>> studies are largely governed by paradigms [or frameworks] of bodies, which
>>> heavily determine  what features we even look for in the first place. A
>>> beatiful piece of philosophical analysis.
>>>
>>> AGI: PROBLEM-SOLVING VS LEARNING.
>>>
>>> I have difficulties with AGI-ers, because my philosophical approach to AGI
>>> is -  start with the end-problems that an AGI must solve, and how they
>>> differ from AI. No one though is interested in discussing them - to a great
>>> extent, perhaps, because the general discussion of such problem distinctions
>>> throughout AI's history (and through psychology's and philosophy's history)
>>> has been pretty poor.
>>>
>>> AGI-ers, it seems to me, focus on learning - on how AGI's must *learn* to
>>> solve problems. The attitude is : if we can just develop a good way for
>>> AGI's to learn here, then they can learn to solve any problem, and gradually
>>> their intelligence will just take off, (hence superAGI). And there is a
>>> great deal of learning theory in AI, and detailed analysis of different
>>> modes of learning, that is logic- and maths-based. So AGI-ers are more
>>> comfortable with this approach.
>>>
>>> PHILOSOPHY OF LEARNING
>>>
>>> However there is relatively little broad-based philosophy of learning. Let's
>>> do some.
>>>
>>> V. broadly, the basic framework, it seems to me, that AGI imposes on
>>> learning to solve problems is:
>>>
>>> 1) define a *set of options* for solving a problem,  and attach if you can,
>>> certain probabilities to them
>>>
>>> 2) test those options,  and carry the best, if any, forward
>>>
>>> 3) find a further set of options from the problem environment, and test
>>> those, updating your probabilities and also perhaps your basic rules for
>>> applying them, as you go
>>>
>>> And, basically, just keep going like that, grinding your way to a solution,
>>> and adapting your program.
>>>
>>> What separates AI from AGI is that in the former:
>>>
>>> * the set of options [or problem space]  is well-defined, [as say, for how a
>>> program can play chess] and the environnment is highly accessible.AGI-ers
>>> recognize their world is much more complicated and not so clearly defined,
>>> and full of *uncertainty*.
>>>
>>> But the common philosophy of both AI and AGI and programming, period, it
>>> seems to me, is : test a set of options.
>>>
>>> THE $1M QUESTION with both approaches is: *how do you define your set of
>>> options*? That's the question I'd like you to try and answer. Let's make it
>>> more concrete.
>>>
>>> a) Defining A Set of Actions?   Take AGI agents, like Ben's, in virtual
>>> worlds. Such agents must learn to perform physical actions and move about
>>> their world. Ben's had to learn how to move to a ball and pick it up.
>>>
>>> So how do you define the set of options here - the set of
>>> actions/trajectories-from-A-to-B that an agent must test? For,say, moving
>>> to, or picking up/hitting a ball. Ben's tried a load - how were they
>>> defined? And by whom? The AGI programmer or the agent?
>>>
>>> b)Defining A Set of Associations ?Essentially, a great deal of formal
>>> problem-solving comes down to working out that A is associated with B,  (if
>>> C,D,E, and however many conditions apply) -   whether A "means," "causes,"
>>> or "contains" B etc etc .
>>>
>>> So basically you go out and test a set of associations, involving A and B
>>> etc, to solve the problem. If you're translating or defining language, you
>>> go and test a whole set of statements involving the relevant words, say "He
>>> jumped over the limit" to know what it means.
>>>
>>> So, again, how do you define the set of options here - the set of
>>> associations to be tested, e.g. the set of texts to be used on Google, say,
>>> for reference for your translation?
>>>
>>> c)What's The Total Possible Set of Options [Actions/Associations] -  how can
>>> you work out the *total* possible set of options to be tested (as opposed to
>>> the set you initially choose) ? Is there one with any AGI problem?
>>>
>>> Can the set of options be definitively defined at all? Is it infinite say
>>> for that set of trajectories, or somehow limited?   (Is there a definitive
>>> or guaranteed way to learn language?)
>>>
>>> d) How Can You Insure the Set of Options is not arbitrary?  That you won't
>>> entirely miss out the crucial options no matter how many more you add? Is
>>> defining a set of options an art not a science - the art of programming,
>>> pace Matt?
>>>
>>> POST HOC VS AD HOC APPROACHES TO LEARNING:  It seems to me there should be a
>>> further condition to how you define your set of options.
>>>
>>> Basically, IMO, AGI learns to solve problems, and AI solves them, *post
>>> hoc.* AFTER the problem has already been solved/learned.
>>>
>>> The perspective of  both on developing a program for
>>> problem-solving/learning is this:
>>>
>>> http://www.danradcliffe.com/12days2005/12days2005_maze_solution.jpg
>>>
>>> you work from the end, with the luxury of a grand overview, after sets of
>>> options and solutions have already been arrived at,  and develop your
>>> program from there.
>>>
>>> But in real life, general intelligences such as humans and animals have to
>>> solve most problems and acquire most skills AD HOC, starting from an
>>> extremely limited view of them :
>>>
>>> http://graphics.stanford.edu/~merrie/Europe/photos/inside%20the%20maze%202.jpg
>>>
>>> where you DON'T know what you're getting into, and you can't be sure what
>>> kind of maze this is, or whether it's a proper maze at all.  That's how YOU
>>> learn to do most things in your life. How can you develop a set of options
>>> from such a position?
>>>
>>> MAZES VS MESSES. Another way of phrasing the question of :"how do you define
>>> the set of options?" is :
>>>
>>> is the set of options along with the problem a maze [clearly definable, even
>>> if only in stages] or a mess:
>>>
>>> http://www.leninimports.com/jackson_pollock_gallery_12.jpg
>>>
>>> [where not a lot is definable]?
>>>
>>> Testing a set of options, it seems to me, is the essence of AI/AGI so far.
>>> It's worth taking time to think about. Philosophically.
>>>
>>>
>>>
>>>
>>> -------------------------------------------
>>> agi
>>> Archives: https://www.listbox.com/member/archive/303/=now
>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>>> Modify Your Subscription:
>>> https://www.listbox.com/member/?&;
>>> Powered by Listbox: http://www.listbox.com
>>>
>>
>>
>> -------------------------------------------
>> agi
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>>
>
>
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>


-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51
Powered by Listbox: http://www.listbox.com

Reply via email to