Jim,

You seem to be suffering from a malady that many others on this forum
appear to be suffering from. The primary symptom is the belief that simple
observation of HIGHLY noise-ridden text can lead machines to useful
epiphanies.

The problem is that pretty much everything people write is, in a word,
wrong. Overbroad statements (and natural languages can NOT precisely
bracket statements), statements made based on unstated presumptions (and no
one can list all their presumptions), statements made based on faulty
models (and ALL models are faulty, as every physicist knows), confused
meanings or words (most words have multiple meanings), lies made for
economic or other gain, etc., etc., etc. It is difficult to find ANY clear
statements that are unquestionably accurate in all of their potential
meanings.

OK, so suppose you accept the above and simply want to do the best you can.
Still, you can NOT approach the capabilities of an average person, because
we have a real world in which to test the many possible interpretations of
what we read, whereas a machine can only accept, reject, or assign a
probability with NO other information.

I used to believe as you do - until 2001 when I became very sick. It took
me 3 months of every conscious minute to figure out what was wrong and what
might be done about it. Sure I got most of my information from books or the
Internet, but there remained several very different models in which this
all fit, which I resolved with phone calls to key researchers, and a little
of my own primary experimentation to sort the flyshit from the pepper.
Still, I remained unsure until what should work did work to cure my
condition.

When I went back to figure out how the Internet might be reorganized to
reduce my search from 3 months to a few minutes, the structure of DrEliza
emerged, and later my patent.

The problem is that to usefully solve problems you need MODELS, yet natural
language (and most human thinking) predates this concept and only provides
INFORMATION. Sometimes you can construct a model from information, but this
is RARE. Making models requires highly qualified geniuses who can get their
arms around entire fields and synthesize models that fit ALL known
observations. I have done this in several narrow areas, but it will take a
LOT more of this to transform the Internet to a model-based system from
which an AI/AGI can usefully address problems of all types.

Further, natural language is highly granular - there are far more things
that you can NOT say than things you CAN say, so people without even
thinking about it round to the nearest syntactically expressible meaning in
EVERYTHING they say or write. This "rounding" completely destroys any
ability to construct accurate models.

Some languages have weak workarounds to this, e.g. German with its
concatenated words, or Arabic where they alter spellings for emphasis, but
these measures only slightly reduce their granularity.

All in all, each person here must either find a way to jump from
information to models, or abandon this quest. Sure, information can help
people solve a problems whose solution has already been stated, but there
are already plenty of experts around who do this quite well. It is the
UNsolved problems that are interesting, and it is these UNsolved problems
that can NOT EVER be solved by automated means based on people's writings.

How can I say not EVER when writings continue to accumulate? Because
society's problems also continue to accumulate, so as people find solutions
to past unsolved problems, even more new problems emerge to replace them.

I hereby proclaim your apparent quest to be theoretically unachievable for
the many reasons outlined above. Sure it would be of astronomical value,
like the methodology to change lead into gold that so many people put so
much effort into, but why waste your time unless/until you can find SOME
way around the above-listed barriers.

Steve
=====================

On Sun, Nov 8, 2015 at 9:00 AM, Jim Bromer <[email protected]> wrote:

> Steve,
> It will take me some time to reply carefully so let me respond to
> something I feel strongly about.
>
> >>And because it is not an all encompassing
> language of communication it could be used to test the 'emergence' of
> insight that could arise if enough preparatory work had been done,
> even if I haven't figured out how that could be done without the
> artificial referent language.
> >>
>
> >There is a VAST chasm between being able to define language constructions
> and meanings, and "insight".
> >
>
> I believe there is a vast chasm between 'simple associations' or 'simple
> correlations' or associations derived from 'neural networks' and conceptual
> integration. Sophisticated artificial conceptual integration would make
> 'insight' feasible and simple examples across a wide range of subject
> matter should arise fairly quickly. But since AI programs are only capable
> of the simplest examples of 'insight' then declarations about the chasm
> between AI and 'insight' are expected. So I totally disagree with you about
> this. I feel that your feelings about this are historically accurate but
> have little to do with the potential near-future. As I say, I do not recall
> hearing about an AI program that is capable of learning via conversation
> except for extremely simple domains. I feel that I have a solution for this
> problem but the trial and error process of getting from where I am now and
> where I think I can get is so overwhelming a challenge that my decision to
> use the artificial referent para-language makes a sense.
>
> Jim Bromer
>
> On Sun, Nov 8, 2015 at 11:19 AM, Steve Richfield <
> [email protected]> wrote:
>
>> Jim,
>>
>> FINALLY - SOMEONE who wants to discuss PRACTICAL implementations of
>> TAI/TAGI.
>>
>> Continuing...
>> On Sun, Nov 8, 2015 at 7:14 AM, Jim Bromer <[email protected]> wrote:
>>
>>> After I wrote that message I realized that I had tried to start
>>> discussions about an artificial language that could be used to shape a
>>> general AI program before. Many of these discussions were side tracked
>>> when people started talking about Esperanto or about lambda calculus
>>> based artificial languages and stuff like that. That is not what I am
>>> thinking of.
>>>
>>
>> You mean, having syntax like:
>>
>> *When that I write "xxxx" I mean "yyy".*
>>
>> to define idioms, for more subtle things like:
>>
>> *Consider that when I write "," I may mean ";".*
>>
>> which expresses potential alternative interpretations of future writings?
>>
>>>
>>> The artificial language could be used with video or audio or other
>>> kinds of IO environments, but I would use it along side of an attempt
>>> to get the AI program to learn to use a natural language.
>>
>>
>> I did a VERY similar thing in a FORTRAN/ALGOL/BASIC compiler I once wrote
>> for Remote Time-Sharing Corp. It started out as a very simplistic
>> metacompiler, to which I fed it a description of a more capable
>> metacompiler, in which language I fed it a description of an optimizing
>> metacompiler.
>>
>> This could easily be done in a language like English, where a rule-driven
>> system like I have been discussing here has rules whose function is to
>> introduce new rules.
>>
>>
>>> One of the
>>> dreams of old AI was that if you started instructing the program to
>>> learn using the artificialities of some kind of language it would
>>> eventually have enough information for genuine learning to emerge.
>>>
>>
>> The think that seems to be the killer here is erroneous learning of
>> various sorts. Superstitious learning is theoretical unavoidable. Once you
>> get something erroneous into such a system, it becomes difficult/impossible
>> to get it out. A VERY simple demonstration comes in trying to use Dragon
>> NaturallySpeaking's speech input to correct its errors in your dictation.
>> As you would expect it makes errors in trying to correct the errors, and
>> this often compounds to overwhelm any hope of setting things right.
>>
>> Add to that not knowing exactly what a computer got wrong, or even being
>> able to recognize that the computer got something wrong, and you can see
>> how difficult/impossible it is to correct wrongly "learned" rules.
>>
>>
>>> This never really worked. Why not? Partly because computers were not
>>> powerful enough in the old days
>>
>>
>> And still aren't - unless you use my patented LFU methodology.
>>
>>
>>> and, in my opinion, AI researchers had
>>> not appreciated the necessity of sophisticated data integration
>>> methods for some reason. (Old computer systems might one day be shown
>>> to have been potentially powerful enough to run some future program
>>> but they were not powerful enough to entertain the trial and error
>>> process that would have been required using experimental programs of
>>> the day.
>>
>>
>> The advantage in LFU is about the same as the advantage of a modern PC
>> over an old vacuum tube clunker, so yes, they could have done a LOT more
>> way back then.
>>
>> The "cycle time" of an IBM-709 computer was 12 microseconds, and most
>> instructions took two cycles - one to access and interpret the instruction,
>> and one to access and operate on the operand.
>>
>>
>>> For example, with better conceptual integration methods a
>>> future efficient AI program might be used on an old computer system
>>> just to show that it could be run on it.)
>>>
>>
>> No, except for a few in the Computer Museum's display in Cupertino they
>> have all been melted down for their scrap metal, and the Museum won't turn
>> them back on.
>>
>>>
>>> So the artificial referent language would not be a complete language
>>> (of communication) like Esperanto wants to be. And it would not be a
>>> logically sound language like lambda calculus wants to be. It could be
>>> used to establish referents from the IO data environment. It would
>>> need to be capable of denoting a distinction between how those data
>>> objects can be used. For example in natural language there is an
>>> important distinction between syntax and semantics. So if I used this
>>> referent language with a natural language IO then one of the
>>> artificialities would be to distinguish syntactic relations from
>>> semantic relations. On the other hand, this distinction is not always
>>> necessary, desired or clear cut. To explain this, many (or maybe most)
>>> (what I think are) desirable syntactic relations are based on some
>>> semantic conditions. But then again there is no reason not to design
>>> the artificial language to be able to represent relations that are
>>> mixes of semantics and syntax.
>>>
>>
>> Leaving a stupid computer to untangle such messes is probably a mistake.
>> However, it would be fairly easy to provide a mechanism for people to
>> specify such things.
>>
>>>
>>> As I see it, the main problem with language based AI has been the lack
>>> of a really good conceptual integration solution.
>>>
>>
>> This broad statement could be said about ANYTHING people haven't yet seen
>> a way to make work - like AGI.
>>
>>>
>>> One of the reasons I write to groups like this is that I want to get
>>> some ideas about how an idea might work.
>>
>>
>> Same here.
>>
>>
>>> But when I wrote about an
>>> artificial para-language before I wasn't really sure it I even wanted
>>> to use it. I finally have come to the conclusion that it makes a lot
>>> of sense. I can use it to speed up tests about my AI/AGI theories but
>>> then I could also test those theories with more relaxed instructions.
>>> So the artificial para-referent language would not a all encompassing
>>> language of communication, it would not be a logically sound language
>>> other than to denote semantic and syntactic references and relations
>>> based on mixes of semantic and syntactic references. It could also
>>> denote relations that I think would be important to a text-based
>>> AI/AGI program. Because the logic of the method would not be tight and
>>> a contradicting case would not (always) lead to an artificially
>>> reported error, the AI methods would have to do some learning for
>>> itself. So the para-referent language would not sidetrack the whole
>>> effort because if the AI methods have to have the potential to exhibit
>>> some genuine learning. And because it is not an all encompassing
>>> language of communication it could be used to test the 'emergence' of
>>> insight that could arise if enough preparatory work had been done,
>>> even if I haven't figured out how that could be done without the
>>> artificial referent language. The benefit is that I could use it to
>>> test and develop my AI theories. I am really excited by this idea this
>>> time.
>>>
>>
>> There is a VAST chasm between being able to define language constructions
>> and meanings, and "insight".
>>
>> *Steve*
>> ======================
>>
>>> Jim Bromer
>>>
>>>
>>> On Sat, Nov 7, 2015 at 10:22 PM, Jim Bromer <[email protected]> wrote:
>>> > I was just working on my latest p=np? idea and I hit up against method
>>> > that is either in np or is otherwise extremely inefficient. So I have
>>> > to come to the conclusion that the human mind is not capable of SAT in
>>> > p.
>>> >
>>> > So then how do we figure how to deal with so many complicated
>>> > situations? Of course I still don't know because so many situations
>>> > seem similar to a SAT problem. The mind must be able to detect many
>>> > different things that are going on at once or which might be useful to
>>> > recall from memory to deal with a situation. But still, there is
>>> > nothing in my own introspective analysis of my thinking which looks
>>> > anything like a p=np process.
>>> >
>>> > So what is wrong with AI? One thing that AI has been consistently
>>> > lacking is the ability to learn through conversation. My feeling is
>>> > that this is not just a problem with communication but a learning
>>> > problem as well. In other words AI is not able to truly learn except
>>> > in a few special cases. Most of those special cases are examples of
>>> > narrow AI but there are others where the learning that takes place
>>> > isn't necessarily like other narrow AI but where the domain of
>>> > learning is so restricted that it is narrow in the sense that the
>>> > applicability of the method is limited.
>>> >
>>> > Then I started thinking of an artificial language which can refer to
>>> > situations or objects in the IO data environment and which can be used
>>> > to instruct a program as it is running. I think this is an unusual
>>> > idea.
>>> >
>>> > One of the characteristics about programming methods that seem to
>>> > catch on with programmers is that they can be used in a very simple
>>> > manner and in more complicated programming. I think an artificial
>>> > language which could be used to instruct a computer to notice objects
>>> > in the IO data environment and which could also be used to refine
>>> > those instructions using this artificial language with the references
>>> > that it had previously established has a lot of potential. And it can
>>> > help us become more clear about what is needed to make better AGI
>>> > programs.
>>> > Jim Bromer
>>>
>>>
>>> -------------------------------------------
>>> AGI
>>> Archives: https://www.listbox.com/member/archive/303/=now
>>> RSS Feed:
>>> https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac
>>> Modify Your Subscription: https://www.listbox.com/member/?&;
>>> Powered by Listbox: http://www.listbox.com
>>>
>>
>>
>>
>> --
>> Full employment can be had with the stoke of a pen. Simply institute a
>> six hour workday. That will easily create enough new jobs to bring back
>> full employment.
>>
>> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/24379807-653794b5> |
>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>> <http://www.listbox.com>
>>
>
> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-- 
Full employment can be had with the stoke of a pen. Simply institute a six
hour workday. That will easily create enough new jobs to bring back full
employment.



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to