Steve,
It will take me some time to reply carefully so let me respond to something
I feel strongly about.

>>And because it is not an all encompassing
language of communication it could be used to test the 'emergence' of
insight that could arise if enough preparatory work had been done,
even if I haven't figured out how that could be done without the
artificial referent language.
>>

>There is a VAST chasm between being able to define language constructions
and meanings, and "insight".
>

I believe there is a vast chasm between 'simple associations' or 'simple
correlations' or associations derived from 'neural networks' and conceptual
integration. Sophisticated artificial conceptual integration would make
'insight' feasible and simple examples across a wide range of subject
matter should arise fairly quickly. But since AI programs are only capable
of the simplest examples of 'insight' then declarations about the chasm
between AI and 'insight' are expected. So I totally disagree with you about
this. I feel that your feelings about this are historically accurate but
have little to do with the potential near-future. As I say, I do not recall
hearing about an AI program that is capable of learning via conversation
except for extremely simple domains. I feel that I have a solution for this
problem but the trial and error process of getting from where I am now and
where I think I can get is so overwhelming a challenge that my decision to
use the artificial referent para-language makes a sense.

Jim Bromer

On Sun, Nov 8, 2015 at 11:19 AM, Steve Richfield <[email protected]>
wrote:

> Jim,
>
> FINALLY - SOMEONE who wants to discuss PRACTICAL implementations of
> TAI/TAGI.
>
> Continuing...
> On Sun, Nov 8, 2015 at 7:14 AM, Jim Bromer <[email protected]> wrote:
>
>> After I wrote that message I realized that I had tried to start
>> discussions about an artificial language that could be used to shape a
>> general AI program before. Many of these discussions were side tracked
>> when people started talking about Esperanto or about lambda calculus
>> based artificial languages and stuff like that. That is not what I am
>> thinking of.
>>
>
> You mean, having syntax like:
>
> *When that I write "xxxx" I mean "yyy".*
>
> to define idioms, for more subtle things like:
>
> *Consider that when I write "," I may mean ";".*
>
> which expresses potential alternative interpretations of future writings?
>
>>
>> The artificial language could be used with video or audio or other
>> kinds of IO environments, but I would use it along side of an attempt
>> to get the AI program to learn to use a natural language.
>
>
> I did a VERY similar thing in a FORTRAN/ALGOL/BASIC compiler I once wrote
> for Remote Time-Sharing Corp. It started out as a very simplistic
> metacompiler, to which I fed it a description of a more capable
> metacompiler, in which language I fed it a description of an optimizing
> metacompiler.
>
> This could easily be done in a language like English, where a rule-driven
> system like I have been discussing here has rules whose function is to
> introduce new rules.
>
>
>> One of the
>> dreams of old AI was that if you started instructing the program to
>> learn using the artificialities of some kind of language it would
>> eventually have enough information for genuine learning to emerge.
>>
>
> The think that seems to be the killer here is erroneous learning of
> various sorts. Superstitious learning is theoretical unavoidable. Once you
> get something erroneous into such a system, it becomes difficult/impossible
> to get it out. A VERY simple demonstration comes in trying to use Dragon
> NaturallySpeaking's speech input to correct its errors in your dictation.
> As you would expect it makes errors in trying to correct the errors, and
> this often compounds to overwhelm any hope of setting things right.
>
> Add to that not knowing exactly what a computer got wrong, or even being
> able to recognize that the computer got something wrong, and you can see
> how difficult/impossible it is to correct wrongly "learned" rules.
>
>
>> This never really worked. Why not? Partly because computers were not
>> powerful enough in the old days
>
>
> And still aren't - unless you use my patented LFU methodology.
>
>
>> and, in my opinion, AI researchers had
>> not appreciated the necessity of sophisticated data integration
>> methods for some reason. (Old computer systems might one day be shown
>> to have been potentially powerful enough to run some future program
>> but they were not powerful enough to entertain the trial and error
>> process that would have been required using experimental programs of
>> the day.
>
>
> The advantage in LFU is about the same as the advantage of a modern PC
> over an old vacuum tube clunker, so yes, they could have done a LOT more
> way back then.
>
> The "cycle time" of an IBM-709 computer was 12 microseconds, and most
> instructions took two cycles - one to access and interpret the instruction,
> and one to access and operate on the operand.
>
>
>> For example, with better conceptual integration methods a
>> future efficient AI program might be used on an old computer system
>> just to show that it could be run on it.)
>>
>
> No, except for a few in the Computer Museum's display in Cupertino they
> have all been melted down for their scrap metal, and the Museum won't turn
> them back on.
>
>>
>> So the artificial referent language would not be a complete language
>> (of communication) like Esperanto wants to be. And it would not be a
>> logically sound language like lambda calculus wants to be. It could be
>> used to establish referents from the IO data environment. It would
>> need to be capable of denoting a distinction between how those data
>> objects can be used. For example in natural language there is an
>> important distinction between syntax and semantics. So if I used this
>> referent language with a natural language IO then one of the
>> artificialities would be to distinguish syntactic relations from
>> semantic relations. On the other hand, this distinction is not always
>> necessary, desired or clear cut. To explain this, many (or maybe most)
>> (what I think are) desirable syntactic relations are based on some
>> semantic conditions. But then again there is no reason not to design
>> the artificial language to be able to represent relations that are
>> mixes of semantics and syntax.
>>
>
> Leaving a stupid computer to untangle such messes is probably a mistake.
> However, it would be fairly easy to provide a mechanism for people to
> specify such things.
>
>>
>> As I see it, the main problem with language based AI has been the lack
>> of a really good conceptual integration solution.
>>
>
> This broad statement could be said about ANYTHING people haven't yet seen
> a way to make work - like AGI.
>
>>
>> One of the reasons I write to groups like this is that I want to get
>> some ideas about how an idea might work.
>
>
> Same here.
>
>
>> But when I wrote about an
>> artificial para-language before I wasn't really sure it I even wanted
>> to use it. I finally have come to the conclusion that it makes a lot
>> of sense. I can use it to speed up tests about my AI/AGI theories but
>> then I could also test those theories with more relaxed instructions.
>> So the artificial para-referent language would not a all encompassing
>> language of communication, it would not be a logically sound language
>> other than to denote semantic and syntactic references and relations
>> based on mixes of semantic and syntactic references. It could also
>> denote relations that I think would be important to a text-based
>> AI/AGI program. Because the logic of the method would not be tight and
>> a contradicting case would not (always) lead to an artificially
>> reported error, the AI methods would have to do some learning for
>> itself. So the para-referent language would not sidetrack the whole
>> effort because if the AI methods have to have the potential to exhibit
>> some genuine learning. And because it is not an all encompassing
>> language of communication it could be used to test the 'emergence' of
>> insight that could arise if enough preparatory work had been done,
>> even if I haven't figured out how that could be done without the
>> artificial referent language. The benefit is that I could use it to
>> test and develop my AI theories. I am really excited by this idea this
>> time.
>>
>
> There is a VAST chasm between being able to define language constructions
> and meanings, and "insight".
>
> *Steve*
> ======================
>
>> Jim Bromer
>>
>>
>> On Sat, Nov 7, 2015 at 10:22 PM, Jim Bromer <[email protected]> wrote:
>> > I was just working on my latest p=np? idea and I hit up against method
>> > that is either in np or is otherwise extremely inefficient. So I have
>> > to come to the conclusion that the human mind is not capable of SAT in
>> > p.
>> >
>> > So then how do we figure how to deal with so many complicated
>> > situations? Of course I still don't know because so many situations
>> > seem similar to a SAT problem. The mind must be able to detect many
>> > different things that are going on at once or which might be useful to
>> > recall from memory to deal with a situation. But still, there is
>> > nothing in my own introspective analysis of my thinking which looks
>> > anything like a p=np process.
>> >
>> > So what is wrong with AI? One thing that AI has been consistently
>> > lacking is the ability to learn through conversation. My feeling is
>> > that this is not just a problem with communication but a learning
>> > problem as well. In other words AI is not able to truly learn except
>> > in a few special cases. Most of those special cases are examples of
>> > narrow AI but there are others where the learning that takes place
>> > isn't necessarily like other narrow AI but where the domain of
>> > learning is so restricted that it is narrow in the sense that the
>> > applicability of the method is limited.
>> >
>> > Then I started thinking of an artificial language which can refer to
>> > situations or objects in the IO data environment and which can be used
>> > to instruct a program as it is running. I think this is an unusual
>> > idea.
>> >
>> > One of the characteristics about programming methods that seem to
>> > catch on with programmers is that they can be used in a very simple
>> > manner and in more complicated programming. I think an artificial
>> > language which could be used to instruct a computer to notice objects
>> > in the IO data environment and which could also be used to refine
>> > those instructions using this artificial language with the references
>> > that it had previously established has a lot of potential. And it can
>> > help us become more clear about what is needed to make better AGI
>> > programs.
>> > Jim Bromer
>>
>>
>> -------------------------------------------
>> AGI
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed:
>> https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>>
>
>
>
> --
> Full employment can be had with the stoke of a pen. Simply institute a six
> hour workday. That will easily create enough new jobs to bring back full
> employment.
>
> *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
> <https://www.listbox.com/member/archive/rss/303/24379807-653794b5> |
> Modify
> <https://www.listbox.com/member/?&;>
> Your Subscription <http://www.listbox.com>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to