On Sat, Mar 18, 2023 at 5:28 AM Telmo Menezes <[email protected]>
wrote:

*> Huge progresses is being made, but we are not at the human level of
> generality of intelligence and autonomy. Not even close.*
>

Not even close? Don't be silly.

*> I fear that you are falling for the very human bias (I fall for it so
> many times myself) of seeing what you want to see.*
>

And I fear you are whistling past the graveyard.

*> A machine learning system can only be objectively evaluated by applying
> it to data that was not used to train it.*
>

I don't know what you mean by that, you're not falling for that old cliché
that computers can only do what they're told to do are you? GPT-4 was not
trained on the exact questions asked; I suppose you could make a case that
some of the training data GPT-4 was educated on was somewhat similar to
some of the questions it was asked, but the exact same thing is true
for human beings. When you ask questions to a human being some of those
questions are somewhat similar to data he was educated on. In fact if some
of the data 2 intelligences were educated on were not similar they would
not be able to ask each other questions because they wouldn't even be able
to communicate.


*>Again, it is important to understand what exactly GPT-4 is doing. It is
> certainly impressive, but it is not the same thing as a human being taking
> an IQ test,*
>

So you must think the following fundamental axiom is true:

*"If a human does something that is smart then the human is smart, but if a
computer does the exact same thing then the computer is NOT smart." *

And from that axiom it's easy to derive the following Corollary:
*"Computers, buy definition, can never be smart."*

I think you need to be more careful in picking your fundamental axioms.

*> I do think that passing the Turing test is impressive,*
>

Probably the greatest understatement of all time.

* > although it is true that most AI researchers never took it very
> seriously,*
>

What?!  I'm sure that in their daily lives AI researchers, like every other
human being on planet earth, have met people in their life that they
considered to be very intelligent, and people they considered to be very
stupid, but if they didn't use the Turing Test to make that determination
then what on earth did they use? All the Turing test is saying is that you
need to play fair, whatever criteria you used to judge the intelligence of
your fellow human beings you should also use on a computer to judge its
intelligence.

It's always the same, I'm old enough to remember when respectable people
were saying a computer would never be able to do better than play a
mediocre game of chess and certainly never be able to beat a grandmaster at
the game. But when a computer did beat a grandmaster at Chess they switched
gears and said such an accomplishment means nothing and insisted a computer
could never beat a human champion at a game like GO because that really
requires true intelligence. Of course when a computer did beat the human
champion at GO they switched gears again and said that accomplishment means
nothing because a computer would never be able to pass the Turing Test
because that really* really* requires true intelligence.  And now that a
computer has passed the Turing Test the human response to that
accomplishment is utterly predictable.  As I said before, they're whistling
past the graveyard.

... and so, just seconds before he was vaporized the last surviving human
being turned to Mr. Jupiter Brain and said "*I still think I'm more
intelligent than you*".


*> GPT-4 and image generators are a type of intelligence that we had never
> seen before. Maybe the first time such a thing arises in this galaxy or
> even universe,*
>

I agree, and I can't think of anything more important that happened in my
lifetime.



>  > *They are probably also similar to stuff that happens in our brain.
> But what they are not is something you can be compare to a human mind with
> an IQ test in any meaningful way.*
>

Not just *an* IQ test but 4 quite different types of IQ tests. And it was a
lobotomized version of GPT-4 that was tested that could not input graphs
and charts or diagrams so any question that contained them was
automatically marked as getting wrong, and yet it STILL got an IQ of 114.
And the computer completed those tests in seconds while it took humans
hours to do the same thing. Imagine what IQ score it will get in two years,
or even two months.  And you say "not even close"?


> *> That is just junk science.*
>

Huh? Creating "a type of intelligence that we had never seen before, maybe
the first time such a thing arises in this galaxy or even the universe", is
junk science?

John K Clark    See what's on my new list at  Extropolis
<https://groups.google.com/g/extropolis>
e4v

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/CAJPayv09UxnUi2NYkBrn7Ti7h9o%3D%2BjDTiuFioY9dyGJ-tnypBQ%40mail.gmail.com.

Reply via email to