The only thing I hope AI will achieve is to be less condescending... if it
achieves true understanding,  I hope it will be humble... and as far as
John Clark dislikes religions and God,  the singularity will be God...

Quentin

Le sam. 5 févr. 2022, 20:51, Terren Suydam <[email protected]> a
écrit :

>
>
> On Fri, Feb 4, 2022 at 6:18 PM John Clark <[email protected]> wrote:
>
>> On Fri, Feb 4, 2022 at 5:34 PM Terren Suydam <[email protected]>
>> wrote:
>>
>> >> Look at this code for a subprogram and make something that does the
>>>> same thing but is smaller or runs faster or both. And that's not a toy
>>>> problem, that's a real problem.
>>>>
>>>
>>> > "does the same thing" is problematic for a couple reasons. The first
>>> is that AlphaCode doesn't know how to read code,
>>>
>>
>> Huh? We already know AlphaCode can write code, how can something know
>> how to write but not read? It's easier to read a novel than write a novel.
>>
>
> This is one case where your intuitions fail. I dug a little deeper into
> how AlphaCode works. It generates millions of candidate solutions using a
> model trained on github code. It then filters out 99% of those candidate
> solutions by running them against test cases provided in the problem
> description and removing the ones that fail. It then uses a different
> technique to whittle down the candidate solutions from several thousand to
> just ten. Nobody, neither the AI nor the humans running AlphaCode, know if
> the 10 solutions picked are correct.
>
> AlphaCode is not capable of reading code. It's a clever version of monkeys
> typing on typewriters until they bang out a Shakespeare play. Still counts
> as AI, but cannot be said to understand code.
>
>
>>
>>> *> The other problem is that with that problem description, it won't
>>> evolve except in the very narrow sense of improving its efficiency.*
>>>
>>
>> It seems to me the ability to write code that was smaller and faster than
>> anybody else is not "very narrow", a human could make a very good living
>> indeed from that talent.  And if I was the guy that signed his enormous
>> paycheck and somebody offered me a program that would do the same thing he
>> did I'd jump at it.
>>
>
> This actually already exists in the form of optimizing compilers - which
> are the programs that translate human-readable code like Java into assembly
> language that microprocessors use to manipulate data. Optimizing compilers
> can make human code more efficient. But these gains are only available in
> very well-understood and limited ways. To do what you're suggesting
> requires machine intelligence capable of understanding things in a much
> broader context.
>
>
>>
>>
>>> *> The kind of problem description that might actually lead to a
>>> singularity is something like "Look at this code and make something that
>>> can solve ever more complex problem descriptions". But my hunch there is
>>> that that problem description is too complex for it to recursively
>>> self-improve towards.*
>>>
>>
>> Just adding more input variables would be less complex than figuring out
>> how to make a program smaller and faster.
>>
>
> Think about it this way. There's diminishing returns on the strategy to
> make the program smaller and faster, but potentially unlimited returns on
> being able to respond to ever greater complexity in the problem
> description.
>
>
>>
>> >> I think if Steven Spielberg's movie had been called AGI instead of AI
>>>> some people today would no longer like the acronym AGI because too many
>>>> people would know exactly what it means and thus would lack that certain
>>>> aura of erudition and mystery that they crave . Everybody knows what AI
>>>> means, but only a small select cognoscenti know the meaning of AGI. A
>>>> Classic case of jargon creep.
>>>>
>>>
>>> >Do you really expect a discipline as technical as AI to not use
>>> jargon?
>>>
>>
>> When totally new concepts come up, as they do occasionally in science,
>> jargon is necessary because there is no previously existing word or short
>> phrase that describes it, but that is not the primary generator of
>> jargon and is not in this case  because a very short word that describes
>> the idea already exists and everybody already knows what AI means, but
>> very few know that AGI means the same thing. And some see that as AGI's
>> great virtue, it's mysterious and sounds brainy.
>>
>>
>>> *> You use physics jargon all the time.*
>>>
>>
>> I do try to keep that to a minimum, perhaps I should try harder.
>>
>
> I don't hold it against you, and I certainly don't think you're trying to
> cultivate an aura of erudition and mystery when you do. I'm not sure why
> you seem to have an axe to grind about the use of AGI, but it is a useful
> distinction to make. It's clear we have AI today. And it's equally clear we
> do not have AGI.
>
> Terren
>
>
>>
>> John K Clark    See what's on my new list at  Extropolis
>> <https://groups.google.com/g/extropolis>
>> pjx
>>
>>
>>
>> --
>> You received this message because you are subscribed to the Google Groups
>> "Everything List" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to [email protected].
>> To view this discussion on the web visit
>> https://groups.google.com/d/msgid/everything-list/CAJPayv0344k0h7t3EjYtgsW5-652P_qieSqyXCtOiAr9zAnmOQ%40mail.gmail.com
>> <https://groups.google.com/d/msgid/everything-list/CAJPayv0344k0h7t3EjYtgsW5-652P_qieSqyXCtOiAr9zAnmOQ%40mail.gmail.com?utm_medium=email&utm_source=footer>
>> .
>>
> --
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/everything-list/CAMy3ZA9AUnjQCeOSP1NVN6djp57oOJLQi6_Se03wuR3yjfEW9A%40mail.gmail.com
> <https://groups.google.com/d/msgid/everything-list/CAMy3ZA9AUnjQCeOSP1NVN6djp57oOJLQi6_Se03wuR3yjfEW9A%40mail.gmail.com?utm_medium=email&utm_source=footer>
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/CAMW2kAo9KTLj49i7UV-rr2U77T2y63H05SBXryoTa-YSZ0a7Sw%40mail.gmail.com.

Reply via email to