Oh, and let's not forget the FHI itself!  When I approached one of its
geniuses during the covid pandemic about setting up something like a Hutter
Prize except using epidemiological data, he insisted on empirical testing
of the efficacy of the Algorithmic Information Criterion.  That sounds
great if you are utterly incapable of rational thought.

On Tue, Apr 23, 2024 at 8:54 AM James Bowery <[email protected]> wrote:

> A book title I've considered:
>
> "The Unfriendly AGI:  How and Why The Global Economy Castrates Our Sons"
>
> Yudowsky is basically a tool of The Unfriendly AGI.   LessWrong
> spearheaded the sophistic attacks on The Hutter Prize.  Why?  So that there
> is no recognition of the Algorithmic Information Criterion in the social
> sciences.  If anything remotely like a Hutter Prize were to take root in
> the social sciences, the TFR disaster being visited on the planet would be
> over in very short order.
>
> On Mon, Apr 22, 2024 at 10:13 PM Matt Mahoney <[email protected]>
> wrote:
>
>> Here is an early (2002) experiment described on SL4 (precursor to
>> Overcoming Bias and Lesswrong) on whether an unfriendly self improving AI
>> could convince humans to let it escape from a box onto the internet.
>> http://sl4.org/archive/0207/4935.html
>>
>> This is how actual science is done on AI safety. The results showed that
>> attempts to contain it would be hopeless. Almost everyone let the (role
>> played) AI escape.
>>
>> Of course the idea that a goal directed, self improving AI could even be
>> developed in isolation from the internet seems hopelessly naïve in
>> hindsight. Eliezer Yudkowsky, who I still regard as brilliant, was young
>> and firmly believe that the unfriendly AI (now called alignment) problem
>> could be and must be solved before it kills everyone, like it was a really
>> hard math problem. Now, after decades of effort it seems he has given up
>> hope. He organized communities of rationalists (Singularity Institute,
>> later MIRI), attempted to formally define human goals (coherent
>> extrapolated volition), timeless decision theory and information hazards
>> (Roko's Basilisk), but to no avail.
>>
>> Vernor Vinge described the Singularity as an event horizon on the future.
>> It cannot be predicted. The best we can do is extrapolate long term trends
>> like Moore's law, increasing quality of life, life expectancy, and economic
>> growth. But who forecast the Internet, social media, social isolation, and
>> population collapse? What are we missing now?
>> *Artificial General Intelligence List <https://agi.topicbox.com/latest>*
>> / AGI / see discussions <https://agi.topicbox.com/groups/agi> +
>> participants <https://agi.topicbox.com/groups/agi/members> +
>> delivery options <https://agi.topicbox.com/groups/agi/subscription>
>> Permalink
>> <https://agi.topicbox.com/groups/agi/Te0da187fd19737a7-M74abe1f60f6dc75c28386a99>
>>

------------------------------------------
Artificial General Intelligence List: AGI
Permalink: 
https://agi.topicbox.com/groups/agi/Te0da187fd19737a7-M38aebe72088cb23a813b1e6e
Delivery options: https://agi.topicbox.com/groups/agi/subscription

Reply via email to