>>Still, you can NOT approach the capabilities of an average person, because we have a real world in which to test the many possible interpretations of what we read, whereas a machine can only accept, reject, or assign a probability with >>NO other information.
I am not interested in an AI program that can only accept reject or assign a probability with no other information, so this reasoning does not apply to me. Jim Bromer On Sun, Nov 8, 2015 at 5:16 PM, Steve Richfield <[email protected]> wrote: > Jim, > > You seem to be suffering from a malady that many others on this forum > appear to be suffering from. The primary symptom is the belief that simple > observation of HIGHLY noise-ridden text can lead machines to useful > epiphanies. > > The problem is that pretty much everything people write is, in a word, > wrong. Overbroad statements (and natural languages can NOT precisely > bracket statements), statements made based on unstated presumptions (and no > one can list all their presumptions), statements made based on faulty > models (and ALL models are faulty, as every physicist knows), confused > meanings or words (most words have multiple meanings), lies made for > economic or other gain, etc., etc., etc. It is difficult to find ANY clear > statements that are unquestionably accurate in all of their potential > meanings. > > OK, so suppose you accept the above and simply want to do the best you > can. Still, you can NOT approach the capabilities of an average person, > because we have a real world in which to test the many possible > interpretations of what we read, whereas a machine can only accept, reject, > or assign a probability with NO other information. > > I used to believe as you do - until 2001 when I became very sick. It took > me 3 months of every conscious minute to figure out what was wrong and what > might be done about it. Sure I got most of my information from books or the > Internet, but there remained several very different models in which this > all fit, which I resolved with phone calls to key researchers, and a little > of my own primary experimentation to sort the flyshit from the pepper. > Still, I remained unsure until what should work did work to cure my > condition. > > When I went back to figure out how the Internet might be reorganized to > reduce my search from 3 months to a few minutes, the structure of DrEliza > emerged, and later my patent. > > The problem is that to usefully solve problems you need MODELS, yet > natural language (and most human thinking) predates this concept and only > provides INFORMATION. Sometimes you can construct a model from information, > but this is RARE. Making models requires highly qualified geniuses who can > get their arms around entire fields and synthesize models that fit ALL > known observations. I have done this in several narrow areas, but it will > take a LOT more of this to transform the Internet to a model-based system > from which an AI/AGI can usefully address problems of all types. > > Further, natural language is highly granular - there are far more things > that you can NOT say than things you CAN say, so people without even > thinking about it round to the nearest syntactically expressible meaning in > EVERYTHING they say or write. This "rounding" completely destroys any > ability to construct accurate models. > > Some languages have weak workarounds to this, e.g. German with its > concatenated words, or Arabic where they alter spellings for emphasis, but > these measures only slightly reduce their granularity. > > All in all, each person here must either find a way to jump from > information to models, or abandon this quest. Sure, information can help > people solve a problems whose solution has already been stated, but there > are already plenty of experts around who do this quite well. It is the > UNsolved problems that are interesting, and it is these UNsolved problems > that can NOT EVER be solved by automated means based on people's writings. > > How can I say not EVER when writings continue to accumulate? Because > society's problems also continue to accumulate, so as people find solutions > to past unsolved problems, even more new problems emerge to replace them. > > I hereby proclaim your apparent quest to be theoretically unachievable for > the many reasons outlined above. Sure it would be of astronomical value, > like the methodology to change lead into gold that so many people put so > much effort into, but why waste your time unless/until you can find SOME > way around the above-listed barriers. > > Steve > ===================== > > On Sun, Nov 8, 2015 at 9:00 AM, Jim Bromer <[email protected]> wrote: > >> Steve, >> It will take me some time to reply carefully so let me respond to >> something I feel strongly about. >> >> >>And because it is not an all encompassing >> language of communication it could be used to test the 'emergence' of >> insight that could arise if enough preparatory work had been done, >> even if I haven't figured out how that could be done without the >> artificial referent language. >> >> >> >> >There is a VAST chasm between being able to define language >> constructions and meanings, and "insight". >> > >> >> I believe there is a vast chasm between 'simple associations' or 'simple >> correlations' or associations derived from 'neural networks' and conceptual >> integration. Sophisticated artificial conceptual integration would make >> 'insight' feasible and simple examples across a wide range of subject >> matter should arise fairly quickly. But since AI programs are only capable >> of the simplest examples of 'insight' then declarations about the chasm >> between AI and 'insight' are expected. So I totally disagree with you about >> this. I feel that your feelings about this are historically accurate but >> have little to do with the potential near-future. As I say, I do not recall >> hearing about an AI program that is capable of learning via conversation >> except for extremely simple domains. I feel that I have a solution for this >> problem but the trial and error process of getting from where I am now and >> where I think I can get is so overwhelming a challenge that my decision to >> use the artificial referent para-language makes a sense. >> >> Jim Bromer >> >> On Sun, Nov 8, 2015 at 11:19 AM, Steve Richfield < >> [email protected]> wrote: >> >>> Jim, >>> >>> FINALLY - SOMEONE who wants to discuss PRACTICAL implementations of >>> TAI/TAGI. >>> >>> Continuing... >>> On Sun, Nov 8, 2015 at 7:14 AM, Jim Bromer <[email protected]> wrote: >>> >>>> After I wrote that message I realized that I had tried to start >>>> discussions about an artificial language that could be used to shape a >>>> general AI program before. Many of these discussions were side tracked >>>> when people started talking about Esperanto or about lambda calculus >>>> based artificial languages and stuff like that. That is not what I am >>>> thinking of. >>>> >>> >>> You mean, having syntax like: >>> >>> *When that I write "xxxx" I mean "yyy".* >>> >>> to define idioms, for more subtle things like: >>> >>> *Consider that when I write "," I may mean ";".* >>> >>> which expresses potential alternative interpretations of future >>> writings? >>> >>>> >>>> The artificial language could be used with video or audio or other >>>> kinds of IO environments, but I would use it along side of an attempt >>>> to get the AI program to learn to use a natural language. >>> >>> >>> I did a VERY similar thing in a FORTRAN/ALGOL/BASIC compiler I once >>> wrote for Remote Time-Sharing Corp. It started out as a very simplistic >>> metacompiler, to which I fed it a description of a more capable >>> metacompiler, in which language I fed it a description of an optimizing >>> metacompiler. >>> >>> This could easily be done in a language like English, where a >>> rule-driven system like I have been discussing here has rules whose >>> function is to introduce new rules. >>> >>> >>>> One of the >>>> dreams of old AI was that if you started instructing the program to >>>> learn using the artificialities of some kind of language it would >>>> eventually have enough information for genuine learning to emerge. >>>> >>> >>> The think that seems to be the killer here is erroneous learning of >>> various sorts. Superstitious learning is theoretical unavoidable. Once you >>> get something erroneous into such a system, it becomes difficult/impossible >>> to get it out. A VERY simple demonstration comes in trying to use Dragon >>> NaturallySpeaking's speech input to correct its errors in your dictation. >>> As you would expect it makes errors in trying to correct the errors, and >>> this often compounds to overwhelm any hope of setting things right. >>> >>> Add to that not knowing exactly what a computer got wrong, or even being >>> able to recognize that the computer got something wrong, and you can see >>> how difficult/impossible it is to correct wrongly "learned" rules. >>> >>> >>>> This never really worked. Why not? Partly because computers were not >>>> powerful enough in the old days >>> >>> >>> And still aren't - unless you use my patented LFU methodology. >>> >>> >>>> and, in my opinion, AI researchers had >>>> not appreciated the necessity of sophisticated data integration >>>> methods for some reason. (Old computer systems might one day be shown >>>> to have been potentially powerful enough to run some future program >>>> but they were not powerful enough to entertain the trial and error >>>> process that would have been required using experimental programs of >>>> the day. >>> >>> >>> The advantage in LFU is about the same as the advantage of a modern PC >>> over an old vacuum tube clunker, so yes, they could have done a LOT more >>> way back then. >>> >>> The "cycle time" of an IBM-709 computer was 12 microseconds, and most >>> instructions took two cycles - one to access and interpret the instruction, >>> and one to access and operate on the operand. >>> >>> >>>> For example, with better conceptual integration methods a >>>> future efficient AI program might be used on an old computer system >>>> just to show that it could be run on it.) >>>> >>> >>> No, except for a few in the Computer Museum's display in Cupertino they >>> have all been melted down for their scrap metal, and the Museum won't turn >>> them back on. >>> >>>> >>>> So the artificial referent language would not be a complete language >>>> (of communication) like Esperanto wants to be. And it would not be a >>>> logically sound language like lambda calculus wants to be. It could be >>>> used to establish referents from the IO data environment. It would >>>> need to be capable of denoting a distinction between how those data >>>> objects can be used. For example in natural language there is an >>>> important distinction between syntax and semantics. So if I used this >>>> referent language with a natural language IO then one of the >>>> artificialities would be to distinguish syntactic relations from >>>> semantic relations. On the other hand, this distinction is not always >>>> necessary, desired or clear cut. To explain this, many (or maybe most) >>>> (what I think are) desirable syntactic relations are based on some >>>> semantic conditions. But then again there is no reason not to design >>>> the artificial language to be able to represent relations that are >>>> mixes of semantics and syntax. >>>> >>> >>> Leaving a stupid computer to untangle such messes is probably a mistake. >>> However, it would be fairly easy to provide a mechanism for people to >>> specify such things. >>> >>>> >>>> As I see it, the main problem with language based AI has been the lack >>>> of a really good conceptual integration solution. >>>> >>> >>> This broad statement could be said about ANYTHING people haven't yet >>> seen a way to make work - like AGI. >>> >>>> >>>> One of the reasons I write to groups like this is that I want to get >>>> some ideas about how an idea might work. >>> >>> >>> Same here. >>> >>> >>>> But when I wrote about an >>>> artificial para-language before I wasn't really sure it I even wanted >>>> to use it. I finally have come to the conclusion that it makes a lot >>>> of sense. I can use it to speed up tests about my AI/AGI theories but >>>> then I could also test those theories with more relaxed instructions. >>>> So the artificial para-referent language would not a all encompassing >>>> language of communication, it would not be a logically sound language >>>> other than to denote semantic and syntactic references and relations >>>> based on mixes of semantic and syntactic references. It could also >>>> denote relations that I think would be important to a text-based >>>> AI/AGI program. Because the logic of the method would not be tight and >>>> a contradicting case would not (always) lead to an artificially >>>> reported error, the AI methods would have to do some learning for >>>> itself. So the para-referent language would not sidetrack the whole >>>> effort because if the AI methods have to have the potential to exhibit >>>> some genuine learning. And because it is not an all encompassing >>>> language of communication it could be used to test the 'emergence' of >>>> insight that could arise if enough preparatory work had been done, >>>> even if I haven't figured out how that could be done without the >>>> artificial referent language. The benefit is that I could use it to >>>> test and develop my AI theories. I am really excited by this idea this >>>> time. >>>> >>> >>> There is a VAST chasm between being able to define language >>> constructions and meanings, and "insight". >>> >>> *Steve* >>> ====================== >>> >>>> Jim Bromer >>>> >>>> >>>> On Sat, Nov 7, 2015 at 10:22 PM, Jim Bromer <[email protected]> >>>> wrote: >>>> > I was just working on my latest p=np? idea and I hit up against method >>>> > that is either in np or is otherwise extremely inefficient. So I have >>>> > to come to the conclusion that the human mind is not capable of SAT in >>>> > p. >>>> > >>>> > So then how do we figure how to deal with so many complicated >>>> > situations? Of course I still don't know because so many situations >>>> > seem similar to a SAT problem. The mind must be able to detect many >>>> > different things that are going on at once or which might be useful to >>>> > recall from memory to deal with a situation. But still, there is >>>> > nothing in my own introspective analysis of my thinking which looks >>>> > anything like a p=np process. >>>> > >>>> > So what is wrong with AI? One thing that AI has been consistently >>>> > lacking is the ability to learn through conversation. My feeling is >>>> > that this is not just a problem with communication but a learning >>>> > problem as well. In other words AI is not able to truly learn except >>>> > in a few special cases. Most of those special cases are examples of >>>> > narrow AI but there are others where the learning that takes place >>>> > isn't necessarily like other narrow AI but where the domain of >>>> > learning is so restricted that it is narrow in the sense that the >>>> > applicability of the method is limited. >>>> > >>>> > Then I started thinking of an artificial language which can refer to >>>> > situations or objects in the IO data environment and which can be used >>>> > to instruct a program as it is running. I think this is an unusual >>>> > idea. >>>> > >>>> > One of the characteristics about programming methods that seem to >>>> > catch on with programmers is that they can be used in a very simple >>>> > manner and in more complicated programming. I think an artificial >>>> > language which could be used to instruct a computer to notice objects >>>> > in the IO data environment and which could also be used to refine >>>> > those instructions using this artificial language with the references >>>> > that it had previously established has a lot of potential. And it can >>>> > help us become more clear about what is needed to make better AGI >>>> > programs. >>>> > Jim Bromer >>>> >>>> >>>> ------------------------------------------- >>>> AGI >>>> Archives: https://www.listbox.com/member/archive/303/=now >>>> RSS Feed: >>>> https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac >>>> Modify Your Subscription: https://www.listbox.com/member/?& >>>> Powered by Listbox: http://www.listbox.com >>>> >>> >>> >>> >>> -- >>> Full employment can be had with the stoke of a pen. Simply institute a >>> six hour workday. That will easily create enough new jobs to bring back >>> full employment. >>> >>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >>> <https://www.listbox.com/member/archive/rss/303/24379807-653794b5> | >>> Modify <https://www.listbox.com/member/?&> Your Subscription >>> <http://www.listbox.com> >>> >> >> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | >> Modify <https://www.listbox.com/member/?&> Your Subscription >> <http://www.listbox.com> >> > > > > -- > Full employment can be had with the stoke of a pen. Simply institute a six > hour workday. That will easily create enough new jobs to bring back full > employment. > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/24379807-653794b5> | > Modify > <https://www.listbox.com/member/?&> > Your Subscription <http://www.listbox.com> > ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
