Hi Mike,

It's not so much the *kind* of programming that I or anyone else could 
recommend, it's just the general skill of programming - getting used to 
thinking in terms of, "how exactly do I solve this problem - what model or 
procedure do I create?" How do you specify something so completely and 
precisely that a mindless machine can execute it?

It's not just that, it's also understanding how the written specification (the 
program) translates into actions at the processor level. That's important too.

Obviously having these skills and knowledge is not the answer to creating AGI - 
if it was, it'd have been solved decades ago. But without understanding how 
computers work, and how we make them work for us, it is too easy to fall into 
the trap of mistaking a computer's operation in terms of some kind of 
homunculus, or that it has a will of its own, or some other kind of anthropic 
confusion. If you don't understand how to program a computer, you will be 
tempted to say that a chess program that can beat Gary Kasparov is intelligent.

Your repeated appeals to creating programs that can "decide for themselves" 
without specifying what they do underscores your technical weakness, because 
programs are nothing but exact specifications. 

You make good points about what General Intelligence entails, but if you had a 
solid grasp of the technical aspects of computing, you could develop your 
philosophy so much further. Matt Mahoney's suggestion of trying to create an 
Artificial Artist is a great example of a direction that is closed to you until 
you learn the things I'm talking about. 

Terren

in response to your PS: I'm not suggesting everyone be proficient at 
everything, although such folks are extremely valuable... why not become one?  
Anyway, sharing expertise is all well and good but in order to do so, you have 
to give ground to the experts - something I haven't seen you do. You seem (to 
me) to be quite attached to your viewpoint, even regarding topics that you 
admit ignorance to. Am I wrong?


--- On Sun, 9/7/08, Mike Tintner <[EMAIL PROTECTED]> wrote:
> Can you tell me which kind of programming is necessary for
> which 
> end-problem[s] that general intelligence must solve? Which
> kind of 
> programming, IOW, can you *guarantee* me  will definitely
> not be a waste of 
> my time (other than by way of general education) ?  Which
> kind are you 
> *sure* will help solve which unsolved problem of AGI?
> 
> P.S. OTOH the idea that in the kind of general community
> I'm espousing, (and 
> is beginning to crop up in other areas), everyone must be
> proficient in 
> everyone else's speciality is actually a non-starter,
> Terren. It defeats the 
> object of the division of labour central to all parts of
> the economy. If you 
> had to spend as much time thinking about those end-problems
> as I have, I 
> suggest you'd have to drop everything. Let's just
> share expertise instead?
> 
> 
> Terren: Good summary. I think your point of view is
> valuable in the sense of 
> helping engineers in AGI to see what they may be missing.
> And your call for 
> technical AI folks to take up the mantle of more artistic
> modes of 
> intelligence is also important.
> >
> > But it's empty, for you've demonstrated no
> willingness to cross over to 
> > engage in technical arguments beyond a certain, quite
> limited, depth. 
> > Admitting your ignorance is one thing, and it's
> laudable, but it only goes 
> > so far. I think if you're serious about getting
> folks (like Pei Wang) to 
> > take you seriously, then you need to also demonstrate
> your willingness to 
> > get your hands dirty and do some programming, or in
> some other way abolish 
> > your ignorance about technical subjects - exactly what
> you're asking 
> > others to do.
> >
> > Otherwise, you have to admit the folly of trying to
> compel any such folks 
> > to move from their hard-earned perspectives, if
> you're not willing to do 
> > that yourself.
> >
> > Terren
> >
> >
> > --- On Sun, 9/7/08, Mike Tintner
> <[EMAIL PROTECTED]> wrote:
> >
> >> From: Mike Tintner
> <[EMAIL PROTECTED]>
> >> Subject: [agi] Philosophy of General Intelligence
> >> To: agi@v2.listbox.com
> >> Date: Sunday, September 7, 2008, 6:26 PM
> >> Jiri: Mike,
> >>
> >> If you think your AGI know-how is superior to the
> know-how
> >> of those
> >> who already built testable thinking machines then
> why
> >> don't you try to
> >> build one yourself?
> >>
> >> Jiri,
> >>
> >> I don't think I know much at all about
> machines or
> >> software & never claim
> >> to. I think I know certain, only certain, things
> about the
> >> psychological and
> >> philosophical aspects of general intelligence -
> esp. BTW
> >> about the things
> >> you guys almost never discuss, the kinds of
> problems that a
> >> general
> >> intelligence must solve.
> >>
> >> You may think that your objections to me are
> entirely
> >> personal  about my
> >> manner. I suggest that there is also a v. deep
> difference
> >> of philosophy
> >> involved here.
> >>
> >> I believe that GI really is about *general*
> intelligence -
> >> a GI, and the
> >> only serious example we have is human, is,
> crucially, and
> >> must be, able to
> >> cross domains - ANY domain. That means the whole
> of our
> >> culture and society.
> >> It means every kind of representation, not just
> >> mathematical and logical and
> >> linguistic, but everything - visual, aural, solid,
> models,
> >> embodied etc etc.
> >> There is a vast range. That means also every
> subject domain
> >>  - artistic,
> >> historical, scientific, philosophical,
> technological,
> >> politics, business
> >> etc. Yes, you have to start somewhere, but there
> should be
> >> no limit to how
> >> you progress.
> >>
> >> And the subject of general intelligence is
> tberefore, in no
> >> way, just the
> >> property of a small community of programmers, or
> >> roboticists - it's the
> >> property of all the sciences, incl. neuroscience,
> >> psychology, semiology,
> >> developmental psychology, AND the arts and
> philosophy etc.
> >> etc. And it can
> >> only be a collaborative effort. Some robotics
> disciplines,
> >> I believe, do
> >> think somewhat along those lines and align
> themselves with
> >> certain sciences.
> >> Some AI-ers also align themselves broadly with
> scientists
> >> and philosophers.
> >>
> >> By definition, too, general intelligence should
> embrace
> >> every kind of
> >> problem that humans have to deal with - again
> artistic,
> >> practical,
> >> technological, political, marketing etc. etc.
> >>
> >> The idea that general intelligence really could be
> anything
> >> else but truly
> >> general is, I suggest, if you really think about
> it,
> >> absurd. It's like
> >> preaching universal brotherhood, and a global
> society, and
> >> then practising
> >> severe racism.
> >>
> >> But that's exactly what's happening in
> current AGI.
> >> You're actually
> >> practising a highly specialised approach to AGI -
> only
> >> certain kinds of
> >> representation, only certain kinds of problems are
> >> considered - basically
> >> the ones you were taught and are comfortable with
> - a very,
> >> very narrow
> >> range - (to a great extent in line with the v.
> narrow
> >> definition of
> >> intelligence involved in the IQ test).
> >>
> >> When I raised other kinds of problems, Pei
> considered it
> >> not "constructive."
> >> When I recently suggested an in fact brilliant
> game for
> >> producing creative
> >> metaphors, DZ considered it "childish," 
> because
> >> it was visual and
> >> imaginative, and you guys don't do those
> things, or
> >> barely. (Far from being
> >> childish, that game produced a rich series of
> visual/verbal
> >> metaphors, where
> >> AGI has produced nothing).
> >>
> >> If you aren't prepared to use your imagination
> and
> >> recognize the other half
> >> of the brain, you are, frankly, completely
> buggered as far
> >> as AGI is
> >> concerned. In over 2000 years, logic and
> mathematics
> >> haven't produced a
> >> single metaphor or analogy or crossed any domains.
> >> They're not meant to,
> >> that's expressly forbidden. But the arts
> produce
> >> metaphors and analogies on
> >> a daily basis by the thousands. The grand irony
> here is
> >> that creativity
> >> really is - from a strictly technical pov - 
> largely what
> >> our culture has
> >> always said it is - imaginative/artistic and not
> rational..
> >> (Many rational
> >> thinkers are creative - but by using their
> imagination).
> >> AGI will in fact
> >> only work if sciences and arts align.
> >>
> >> Here, then is basically why I think you're
> getting
> >> upset over and over by
> >> me. I'm saying in many different ways, general
> >> intelligence really should be
> >> general, and embrace the whole of culture and
> intelligence,
> >> not just the
> >> very narrow sections you guys espouse. And yes, I
> think you
> >> should be
> >> delighted to defer to, and learn from
> >> "outsiders", (if they deserve it),
> >> just as I'm delighted to learn from you. But
> you're
> >> not - you resent
> >> outsiders like me telling you about
> "your"
> >> subject.
> >>
> >> I think you should also be prepared to admit your
> ignorance
> >> - and most of
> >> you, frankly, don't have much of a clue about
> >> imaginative/visual/artistic
> >> intelligence and vast swathes of problemsolving, (
> just as
> >> I have don't have
> >> much of a clue re your technology and many kinds
> of
> >> problemsolving...etc).
> >> But there is v. little willingness to admit
> ignorance, or
> >> to acknowledge the
> >> value of other disciplines.
> >>
> >> IN the final analysis, I suggest, that's just
> sheer
> >> cultural prejudice. It
> >> doesn't belong in the new millennium when the
> defining
> >> paradigm is global
> >> (and general) as opposed to the local (and
> specialist)
> >> mentality of the old
> >> one - recognizing the value and interdependence of
> ALL
> >> parts of society and
> >> culture. And it doesn't belong in a true field
> of
> >> *General* INtelligence. I
> >> think you need to change your central philosophy
> in a major
> >> way and be
> >> culturally open-minded. (and then just possibly
> you
> >> won't find me quite so
> >> upsetting.)
> >>
> >>
> >>
> >>
> >> -------------------------------------------
> >> agi
> >> Archives:
> https://www.listbox.com/member/archive/303/=now
> >> RSS Feed:
> https://www.listbox.com/member/archive/rss/303/
> >> Modify Your Subscription:
> >> https://www.listbox.com/member/?&;
> >> Powered by Listbox: http://www.listbox.com
> >
> >
> >
> >
> >
> > -------------------------------------------
> > agi
> > Archives:
> https://www.listbox.com/member/archive/303/=now
> > RSS Feed:
> https://www.listbox.com/member/archive/rss/303/
> > Modify Your Subscription: 
> > https://www.listbox.com/member/?&;
> > Powered by Listbox: http://www.listbox.com
> > 
> 
> 
> 
> 
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com


      


-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51
Powered by Listbox: http://www.listbox.com

Reply via email to