Jim, You are right to call me on that. I need to provide an argument that, if no logic satisfying B exists, human-level AGI is impossible.
B1: A foundational logic for a human-level intelligence should be capable of expressing any concept that a human can meaningfully express. If a broad enough interpretation of the word "logic" is taken, this statement is obvious; it could amount to simply "A human level intelligence should be capable of expressing anything it can meaningfully express". (ie, logic = way of operating.) So, with this interpretation, it doesn't even make sense for B to be false. But, this is not quite what I mean. The key idea for me is that logic is not the way we *do* think, it is the way we *should* think, in the ideal situation of infinite computational resources. So, a more refined B would state: B2: The theoretical ideal of how a human-level intelligence should think, should capture everything worth capturing about the way humans actually do think. "Everything worth capturing" means everything that could lead to good results. So, I argue, if no logic exists satisfying B2, then human-level artificial intelligence is not possible. In fact, I think the negation of B2 is nonsensical: not-B2: There is no concept of how a human-level intelligence should think that captures everything worth capturing about how humans do think. This seems to imply that humans do not exist, since the way humans actually *do* think captures everything worth capturing (as well as some things not worth capturing) about how we think. -Abram On Thu, Aug 14, 2008 at 2:04 PM, Jim Bromer <[EMAIL PROTECTED]> wrote: > On Thu, Aug 14, 2008 at 12:59 PM, Abram Demski <[EMAIL PROTECTED]> wrote: >> A more worrisome problem is that B may be contradictory in and of >> itself. If (1) I can as a human meaningfully explain logical system X, >> and (2) logical system X can meaningfully explain anything that humans >> can, then (3) system X can meaningfully explain itself. Tarski's >> Indefineability Theorem shows that any such system (under some >> seemingly reasonable assumptions) can express the concept "This >> concept is false", and is therefore (again under some seemingly >> reasonable assumptions) contradictory. So, if we accept those >> "seemingly reasonable assumptions", no logic satisfying B exists. >> >> But, this implies that AI is impossible. > > At risk of being really annoying I have to say: it does not imply > anything of the sort! How could it imply some idea if it can't even > represent it? It implies impossibility to you because you are capable > of dealing with fictions of boundaries as if they were real until you > either conflate two or more bounded concepts, or simply transcend them > by virtue of the extent of your everyday experience. > > In order to deal with logic you have to be taught how to consider such > a thing to be bounded from the rest of reality. That is not difficult > because that is a requirement of all thought and it is a fundamental > necessity of dealing with the real universe. To study anything you > have to limit your attention to the subject. > > I believe we can use logic in AI to detect possible errors and > boundary issues. But then we have to build up a system of knowledge > from experience which seems to transcend the usual boundaries when > that kind of transcendent insight becomes useful. After a while > transcendent insight is itself recognized to be bounded and so it > becomes part of the mundane. > > Think of this way, boundaries are primarily dependent on limitations. > Jim Bromer > > > > > > > On Thu, Aug 14, 2008 at 12:59 PM, Abram Demski <[EMAIL PROTECTED]> wrote: >> This looks like it could be an interesting thread. >> >> However, I disagree with your distinction between ad hoc and post hoc. >> The programmer may see things from the high-level "maze" view, but the >> program itself typically deals with the "mess". So, I don't think >> there is a real distinction to be made between post-hoc AI systems and >> ad-hoc ones. >> >> When we decide on the knowledge representation, we predefine the space >> of solutions that the AI can find. This cannot be avoided. The space >> can be made wider by restricting the knowledge representation less >> (for example, allowing the AI to create arbitrary assembly-language >> programs is less of a restriction than requiring it to learn >> production-rule programs that get executed by some implementation of >> the rete algorithm). But obviously we run into hardware restrictions. >> The broadest space to search is the space of all possible >> configurations of 1s and 0s inside the computer we're using. An AI >> method called "Godel Machines" is supposed to do that. William Pearson >> is also interested in this. >> >> Since we're doing philosophy here, I'll take a philosophical stance. >> Here are my assumptions. >> >> 0. I assume that there is some proper logic. >> >> 1. I assume probability theory and utility theory, acting upon >> statements in this logic, are good descriptions of the ideal >> decision-making process (if we do not need to worry about >> computational resources). >> >> 2. I assume that there is some reasonable bayesian prior over the >> logic, and therefore (given #1) that bayesian updating is the ideal >> learning method (again given infinite computation). >> >> This philosophy is not exactly the one you outlined as the AI/AGI >> standard: there is no searching. #2 should ideally be carried out by >> computing the probability of *all* models. With finite computational >> resources, this is typically approximated by searching for >> high-probability models, which works well because the low-probability >> models contribute little to the decision-making process in most cases. >> >> Now, to do some philosophy on my assumptions :). >> >> Consideration of #0: >> >> This is my chief concern. We must find the proper logic. This is very >> close to your concern, because the search space is determined by the >> logic we choose (given the above-mentioned approximation to #2). If >> you think the search space is too restricted, then essentially you are >> saying we need a broader logic. My requirements for the logic are: >> A. The logic should be grounded >> B. The logic should be able to say any meaningful thing a human can say >> The two requirements are not jointly satisfied by any existing logic >> (using my personal definition of grounded, at least). Set theory is >> the broadest logic typically considered, so it comes closest to B, but >> it (and most other logics considered strong enough to serve as a >> foundation of mathematics) do not pass the test of A because the >> manipulation rules do not match up to their semantics. I explain this >> at some length here: >> >> http://groups.google.com/group/opencog/browse_thread/thread/28755f668e2d4267/10245c1d4b3984ca?lnk=gst&q=abramdemski#10245c1d4b3984ca >> >> A more worrisome problem is that B may be contradictory in and of >> itself. If (1) I can as a human meaningfully explain logical system X, >> and (2) logical system X can meaningfully explain anything that humans >> can, then (3) system X can meaningfully explain itself. Tarski's >> Indefineability Theorem shows that any such system (under some >> seemingly reasonable assumptions) can express the concept "This >> concept is false", and is therefore (again under some seemingly >> reasonable assumptions) contradictory. So, if we accept those >> "seemingly reasonable assumptions", no logic satisfying B exists. >> >> But, this implies that AI is impossible. So, some of the seemingly >> reasonable assumptions need to be dismissed. (But I don't know which >> ones.) >> >> Consideration of #2: >> >> Assumption 3 is that there exists some reasonable prior probability >> distribution that we can use for learning. A now-common way of >> choosing this prior is the minimum description length principle, which >> tells us that shorter theories are more probable. >> >> The following argument was sent to me by private email by Wei Dai, and >> I think it is very revealing: >> >> "I did suggest a prior based on set theory, but then I realized that >> it doesn't really solve the entire problem. The real problem seems to >> be that if we formalize induction as Bayesian sequence prediction with >> a well-defined prior, we can immediately produce a sequence that an >> ideal predictor should be able to predict, but this one doesn't, no >> matter what the prior is. Specifically, the sequence is the "least >> expected" sequence of the predictor. We generate each symbol in this >> sequence by feeding the previous symbols to the predictor and then >> pick the next symbol as the one that it predicts with the smallest >> probability. (Pick the lexicographically first symbol if more than one >> has the smallest probability.) >> >> This least expected sequence has a simple description, and therefore >> should not be the least expected, right? Why should it not be more >> expected than a completely random sequence, for example? >> >> Do you think your approach can avoid this problem?" >> >> Again, if the argument is right, then AI is impossible (or, as Wei Dai >> put it, induction cannot be formalized). So, I again conclude that >> some assumption needs to be dismissed. In this case the most obvious >> assumption is that the prior should be based on minimum description >> length. I thought at the time that that was the wrong one, but I am >> not sure at the moment. >> >> So, any ideas how to resolve these two problems? >> >> --Abram >> >> >> >> PS-- I don't want to leave the quote from Wei Dai completely out of >> context, so here is Wei Dai's full argument: >> >> http://groups.google.com/group/everything-list/browse_frm/thread/c7442c13ff1396ec/804e134c70d4a203 >> >> >> On Wed, Aug 13, 2008 at 10:15 AM, Mike Tintner <[EMAIL PROTECTED]> wrote: >>> THE POINT OF PHILOSOPHY: There seemed to be some confusion re this - the >>> main point of philosophy is that it makes us aware of the frameworks that >>> are brought to bear on any subject, from sci to tech to business to arts - >>> and therefore the limitations of those frameworks. Crudely, it says: hey >>> you're looking in 2D, you could be loooking in 3D or nD. >>> >>> Classic example: Kuhn. Hey, he said, we've thought science discovers bodies >>> feature-by-feature, with a steady-accumulation-of-facts. Actually those >>> studies are largely governed by paradigms [or frameworks] of bodies, which >>> heavily determine what features we even look for in the first place. A >>> beatiful piece of philosophical analysis. >>> >>> AGI: PROBLEM-SOLVING VS LEARNING. >>> >>> I have difficulties with AGI-ers, because my philosophical approach to AGI >>> is - start with the end-problems that an AGI must solve, and how they >>> differ from AI. No one though is interested in discussing them - to a great >>> extent, perhaps, because the general discussion of such problem distinctions >>> throughout AI's history (and through psychology's and philosophy's history) >>> has been pretty poor. >>> >>> AGI-ers, it seems to me, focus on learning - on how AGI's must *learn* to >>> solve problems. The attitude is : if we can just develop a good way for >>> AGI's to learn here, then they can learn to solve any problem, and gradually >>> their intelligence will just take off, (hence superAGI). And there is a >>> great deal of learning theory in AI, and detailed analysis of different >>> modes of learning, that is logic- and maths-based. So AGI-ers are more >>> comfortable with this approach. >>> >>> PHILOSOPHY OF LEARNING >>> >>> However there is relatively little broad-based philosophy of learning. Let's >>> do some. >>> >>> V. broadly, the basic framework, it seems to me, that AGI imposes on >>> learning to solve problems is: >>> >>> 1) define a *set of options* for solving a problem, and attach if you can, >>> certain probabilities to them >>> >>> 2) test those options, and carry the best, if any, forward >>> >>> 3) find a further set of options from the problem environment, and test >>> those, updating your probabilities and also perhaps your basic rules for >>> applying them, as you go >>> >>> And, basically, just keep going like that, grinding your way to a solution, >>> and adapting your program. >>> >>> What separates AI from AGI is that in the former: >>> >>> * the set of options [or problem space] is well-defined, [as say, for how a >>> program can play chess] and the environnment is highly accessible.AGI-ers >>> recognize their world is much more complicated and not so clearly defined, >>> and full of *uncertainty*. >>> >>> But the common philosophy of both AI and AGI and programming, period, it >>> seems to me, is : test a set of options. >>> >>> THE $1M QUESTION with both approaches is: *how do you define your set of >>> options*? That's the question I'd like you to try and answer. Let's make it >>> more concrete. >>> >>> a) Defining A Set of Actions? Take AGI agents, like Ben's, in virtual >>> worlds. Such agents must learn to perform physical actions and move about >>> their world. Ben's had to learn how to move to a ball and pick it up. >>> >>> So how do you define the set of options here - the set of >>> actions/trajectories-from-A-to-B that an agent must test? For,say, moving >>> to, or picking up/hitting a ball. Ben's tried a load - how were they >>> defined? And by whom? The AGI programmer or the agent? >>> >>> b)Defining A Set of Associations ?Essentially, a great deal of formal >>> problem-solving comes down to working out that A is associated with B, (if >>> C,D,E, and however many conditions apply) - whether A "means," "causes," >>> or "contains" B etc etc . >>> >>> So basically you go out and test a set of associations, involving A and B >>> etc, to solve the problem. If you're translating or defining language, you >>> go and test a whole set of statements involving the relevant words, say "He >>> jumped over the limit" to know what it means. >>> >>> So, again, how do you define the set of options here - the set of >>> associations to be tested, e.g. the set of texts to be used on Google, say, >>> for reference for your translation? >>> >>> c)What's The Total Possible Set of Options [Actions/Associations] - how can >>> you work out the *total* possible set of options to be tested (as opposed to >>> the set you initially choose) ? Is there one with any AGI problem? >>> >>> Can the set of options be definitively defined at all? Is it infinite say >>> for that set of trajectories, or somehow limited? (Is there a definitive >>> or guaranteed way to learn language?) >>> >>> d) How Can You Insure the Set of Options is not arbitrary? That you won't >>> entirely miss out the crucial options no matter how many more you add? Is >>> defining a set of options an art not a science - the art of programming, >>> pace Matt? >>> >>> POST HOC VS AD HOC APPROACHES TO LEARNING: It seems to me there should be a >>> further condition to how you define your set of options. >>> >>> Basically, IMO, AGI learns to solve problems, and AI solves them, *post >>> hoc.* AFTER the problem has already been solved/learned. >>> >>> The perspective of both on developing a program for >>> problem-solving/learning is this: >>> >>> http://www.danradcliffe.com/12days2005/12days2005_maze_solution.jpg >>> >>> you work from the end, with the luxury of a grand overview, after sets of >>> options and solutions have already been arrived at, and develop your >>> program from there. >>> >>> But in real life, general intelligences such as humans and animals have to >>> solve most problems and acquire most skills AD HOC, starting from an >>> extremely limited view of them : >>> >>> http://graphics.stanford.edu/~merrie/Europe/photos/inside%20the%20maze%202.jpg >>> >>> where you DON'T know what you're getting into, and you can't be sure what >>> kind of maze this is, or whether it's a proper maze at all. That's how YOU >>> learn to do most things in your life. How can you develop a set of options >>> from such a position? >>> >>> MAZES VS MESSES. Another way of phrasing the question of :"how do you define >>> the set of options?" is : >>> >>> is the set of options along with the problem a maze [clearly definable, even >>> if only in stages] or a mess: >>> >>> http://www.leninimports.com/jackson_pollock_gallery_12.jpg >>> >>> [where not a lot is definable]? >>> >>> Testing a set of options, it seems to me, is the essence of AI/AGI so far. >>> It's worth taking time to think about. Philosophically. >>> >>> >>> >>> >>> ------------------------------------------- >>> agi >>> Archives: https://www.listbox.com/member/archive/303/=now >>> RSS Feed: https://www.listbox.com/member/archive/rss/303/ >>> Modify Your Subscription: >>> https://www.listbox.com/member/?& >>> Powered by Listbox: http://www.listbox.com >>> >> >> >> ------------------------------------------- >> agi >> Archives: https://www.listbox.com/member/archive/303/=now >> RSS Feed: https://www.listbox.com/member/archive/rss/303/ >> Modify Your Subscription: https://www.listbox.com/member/?& >> Powered by Listbox: http://www.listbox.com >> > > > ------------------------------------------- > agi > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/ > Modify Your Subscription: https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com > ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51 Powered by Listbox: http://www.listbox.com
