Hi,

Is it not how evolution is working ? By iteration and random modification,
new better organisms come to existence ?

Why AI could not use iterating evolution to make better and better AI ?

Also if *we build* a real AGI, isn't it the same thing ? Wouldn't we have
built a better, smarter version of us ? The AI surely would be able to
build another one and by iterating, a better one.

What's wrong with this ?

Quentin

Le ven. 12 juil. 2019 à 06:28, Terren Suydam <[email protected]> a
écrit :

> Sure, but that's not the "FOOM" scenario, in which an AI modifies its own
> source code, gets smarter, and with the increase in intelligence, is able
> to make yet more modifications to its own source code, and so on, until its
> intelligence far outstrips its previous capabilities before the recursive
> self-improvement began. It's hypothesized that such a process could take an
> astonishingly short amount of time, thus "FOOM". See
> https://wiki.lesswrong.com/wiki/AI_takeoff#Hard_takeoff for more.
>
> My point was that the inherent limitation of a mind to understand itself
> completely, makes the FOOM scenario less likely. An AI would be forced to
> model its own cognitive apparatus in a necessarily incomplete way. It might
> still be possible to improve itself using these incomplete models, but
> there would always be some uncertainty.
>
> Another more minor objection is that the FOOM scenario also selects for
> AIs that become massively competent at self-improvement, but it's not clear
> whether this selected-for intelligence is merely a narrow competence, or
> translates generally to other domains of interest.
>
>
> On Thu, Jul 11, 2019 at 2:56 PM 'Brent Meeker' via Everything List <
> [email protected]> wrote:
>
>> Advances in intelligence can just be gaining more factual knowledge,
>> knowing more mathematics, using faster algorithms, etc.  None of that is
>> barred by not being able to model oneself.
>>
>> Brent
>>
>> On 7/11/2019 11:41 AM, Terren Suydam wrote:
>> > Similarly, one can never completely understand one's own mind, for it
>> > would take a bigger mind than one has to do so. This, I believe, is
>> > the best argument against the runaway-intelligence scenarios in which
>> > sufficiently advanced AIs recursively improve their own code to
>> > achieve ever increasing advances in intelligence.
>> >
>> > Terren
>>
>>
>> --
>> You received this message because you are subscribed to the Google Groups
>> "Everything List" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to [email protected].
>> To view this discussion on the web visit
>> https://groups.google.com/d/msgid/everything-list/304332c1-13a6-7006-651b-494e468eefc4%40verizon.net
>> .
>>
> --
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/everything-list/CAMy3ZA9xK%3DibZqo%3DxQcqSVZXjTu3pnAiTvRLF_8-LHVRth8F_w%40mail.gmail.com
> <https://groups.google.com/d/msgid/everything-list/CAMy3ZA9xK%3DibZqo%3DxQcqSVZXjTu3pnAiTvRLF_8-LHVRth8F_w%40mail.gmail.com?utm_medium=email&utm_source=footer>
> .
>


-- 
All those moments will be lost in time, like tears in rain. (Roy
Batty/Rutger Hauer)

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/everything-list/CAMW2kAoZrj4nXJ_EFCCSupSOWP_ows52ECR3w3zLBrNg8UDsyg%40mail.gmail.com.

Reply via email to