Mark Waser wrote:
>How is translating patterns into language different from manipulating patterns? > It seems to me that they are *exactly* the same thing. How do you believe that they differ? Manipulating of patterns needs reading and writing operations. Data structures will be changed. Translation needs just reading operations to the patterns of the internal model. >Do you really believe that if A is easier than B then that makes A easy? > How about if A is leaping a tall building in a single bound and B is jumping to the moon? The word *easy* is not exactly definable. > Do you believe that language is fully specified? That we can program English into an AGI by hand? No. Thats the reason why I would not use human language for the first AGI. >Yes, I imagine that an AGI must have some process for learning language because language is necessary for >learning knowledge and knowledge is necessary for intelligence. >What part of that do you disagree with? Please be specific. I disagree that AGI must have some process for learning language. If we concentrate just on the domain of mathematics we could give AGI all the rules for a sufficient language to express its results and to understand our questions. >>> >And this is where we are not communicating. Since language is not fully specified, then the participants in >many conversations are *constantly* creating and learning language as a part of the process of >communication. This is where Gödel's incompleteness comes in. To be a General Intelligence, you must be able to extend beyond what is currently known and specified into new domains. Any time that we are teaching or learning (i.e. modifying our model of the world), we are also necessarily extending our models of each other and language. The computer database analogy you are basing your entire argument upon does not have the necessary features/complexity to be an accurate or useful analogy. <<< Language must only grow if you make new definitions and want to communicate the definition to another agent. But new definitions are not necessary for general intelligence. If you define Methane := CH4 Then it is your choice whether you say the new word methane or you use the known expression CH4. New definitions makes communication more comfortable but they are not necessary. ***Even if you change your model and your language at the same time then there is still a strict distinction between them. Language would still only be used for communication and not for the data structure of the patterns for the world model or the algorithms which manipulate these patterns.*** >Again, I disagree. You added internal details but the end result after the details are hidden is that e-mail > programs are just point-to-point repeaters. That is why I used the examples (the telephone game and round- >trip (mis)translations) that I did which you did not address. I dont know the telephone game. The details are essential. It is not essential where the data comes from and where it ends. Just the process of translating internal data into a certain language and vice versa is important. >> You *believe* that language cannot be separated from intelligence. I dont and I have described a model which has a strict separation. We both have no proof. >>> Three points. 1. My statement was that intelligence can't be built without language/communication. That is entirely different from the fact that they can't be separated. I also gave reasoning why this was the case that you haven't addressed. <<< The main point in this discussion is whether language /communication can be separated from intelligence. It is clear that an AGI needs an interface for human beings. But the question in this discussion is whether the language interface is a key point in AGI or not. In my opinion it is no key point. It is just a communication protocol. The real intelligence has nothing to do with language understanding. Therefore we should use a simple formal hard coded language for first AGI. >>> 2. Your model has serious flaws that you have not answered. You are relying upon an analogy that has points that you have not shown that you are able to defend. Until you do so, this invalidates your model. <<< I dont see any problems with my model and I do not see any flaws which I dont have answered. >>> 3. You have not provided a disproof or counter-example to what I am saying. I have clearly specified where your analogy comes up short and other inaccuracies in your statements while you have not done so for any of mine (other than of the "tis too, tis not" variety). <<< I havent seen any point where my analogy comes short. >>> I have had the courtesy to directly address your points with clear counter-examples. Please return the favor and do not simply drop my examples without replying to them and revert back to global statements. Global statements are great for an initial exposition but eventually you have to get down to the details and work out the nitty-gritty. Thanks. <<< I havent drop your examples. ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com
