When Nick Bostrom hired Anders Sandberg at FHI he should have made
Anders the director and become a researcher, as Eliezer Yudkowsky
is a researcher at MIRI with someone else as director. I understand
this as a person who should not be an institutional director - my
inevitable personality conficts would damage any institute I ran.
With Anders as director I think FHI would not have been closed.
On Wed, 24 Apr 2024, Ben Goertzel wrote:
FHI shared an office with the Institute for Effective Altruism, which
is not a terribly popular movement since SBF
The EA association combined with off-and-on attention to some old
racist-appearing comments Nick made online years ago, could perhaps
have contributed to FHI's alienation from the Oxford bureaucracy
CSER at Cambridge has essentially the same mission/theme but is run
more-so by UK academic establishment so runs little risk of
cancellation...
-- Ben
On Tue, Apr 23, 2024 at 12:35 PM James Bowery <[email protected]> wrote:
Last year it was reported Bostrom said "nigger" in 1998 or thereabouts.
https://urldefense.com/v3/__https://youtu.be/Lu_i042oaNg__;!!Mak6IKo!NeVEFmCPYayuoNiIaTPuT1I_F5Whk4nK4-UjEla0lvrwoQFO8DTj2M-R_dPdcMLUNapUQYvRzdJobOCnV-rl$
On Tue, Apr 23, 2024 at 9:00 AM James Bowery <[email protected]> wrote:
Oh, and let's not forget the FHI itself! When I approached one of its geniuses
during the covid pandemic about setting up something like a Hutter Prize except
using epidemiological data, he insisted on empirical testing of the efficacy of
the Algorithmic Information Criterion. That sounds great if you are utterly
incapable of rational thought.
On Tue, Apr 23, 2024 at 8:54 AM James Bowery <[email protected]> wrote:
A book title I've considered:
"The Unfriendly AGI: How and Why The Global Economy Castrates Our Sons"
Yudowsky is basically a tool of The Unfriendly AGI. LessWrong spearheaded the
sophistic attacks on The Hutter Prize. Why? So that there is no recognition
of the Algorithmic Information Criterion in the social sciences. If anything
remotely like a Hutter Prize were to take root in the social sciences, the TFR
disaster being visited on the planet would be over in very short order.
On Mon, Apr 22, 2024 at 10:13 PM Matt Mahoney <[email protected]> wrote:
Here is an early (2002) experiment described on SL4 (precursor to Overcoming
Bias and Lesswrong) on whether an unfriendly self improving AI could convince
humans to let it escape from a box onto the internet.
https://urldefense.com/v3/__http://sl4.org/archive/0207/4935.html__;!!Mak6IKo!NeVEFmCPYayuoNiIaTPuT1I_F5Whk4nK4-UjEla0lvrwoQFO8DTj2M-R_dPdcMLUNapUQYvRzdJobMIvRrAn$
This is how actual science is done on AI safety. The results showed that
attempts to contain it would be hopeless. Almost everyone let the (role played)
AI escape.
Of course the idea that a goal directed, self improving AI could even be
developed in isolation from the internet seems hopelessly naïve in hindsight.
Eliezer Yudkowsky, who I still regard as brilliant, was young and firmly
believe that the unfriendly AI (now called alignment) problem could be and must
be solved before it kills everyone, like it was a really hard math problem.
Now, after decades of effort it seems he has given up hope. He organized
communities of rationalists (Singularity Institute, later MIRI), attempted to
formally define human goals (coherent extrapolated volition), timeless decision
theory and information hazards (Roko's Basilisk), but to no avail.
Vernor Vinge described the Singularity as an event horizon on the future. It
cannot be predicted. The best we can do is extrapolate long term trends like
Moore's law, increasing quality of life, life expectancy, and economic growth.
But who forecast the Internet, social media, social isolation, and population
collapse? What are we missing now?
Artificial General Intelligence List / AGI / see discussions + participants +
delivery options Permalink
--
Ben Goertzel, PhD
[email protected]
"One must have chaos in one's heart to give birth to a dancing star"
-- Friedrich Nietzsche
------------------------------------------
Artificial General Intelligence List: AGI
Permalink:
https://urldefense.com/v3/__https://agi.topicbox.com/groups/agi/Te0da187fd19737a7-M65cfae2d16e21ae8f460a758__;!!Mak6IKo!NeVEFmCPYayuoNiIaTPuT1I_F5Whk4nK4-UjEla0lvrwoQFO8DTj2M-R_dPdcMLUNapUQYvRzdJobCx4tUb5$
Delivery options:
https://urldefense.com/v3/__https://agi.topicbox.com/groups/agi/subscription__;!!Mak6IKo!NeVEFmCPYayuoNiIaTPuT1I_F5Whk4nK4-UjEla0lvrwoQFO8DTj2M-R_dPdcMLUNapUQYvRzdJobKWnisFn$
------------------------------------------
Artificial General Intelligence List: AGI
Permalink:
https://agi.topicbox.com/groups/agi/Te0da187fd19737a7-M9ac3f48dade0fd1a8f0fa3b5
Delivery options: https://agi.topicbox.com/groups/agi/subscription