On Tue, Feb 10, 2015 at 11:35 PM, meekerdb <[email protected]> wrote:
> On 2/10/2015 5:49 PM, Jason Resch wrote: > > > > On Tue, Feb 10, 2015 at 6:40 PM, meekerdb <[email protected]> wrote: > >> >> On 2/10/2015 8:47 AM, Jason Resch wrote: >> >> If you define increased intelligence as decreased probability of having a >> false belief on any randomly chosen proposition, then superintelligences >> will be wrong on almost nothing, and their beliefs will converge as their >> intelligence rises. Therefore nearly all superintelligences will operate >> according to the same belief system. We should stop worrying about trying >> to ensure friendly AI, it will either be friendly or it won't according to >> what is right. >> >> >> The problem isn't beliefs, it's values. Humans have certain core values >> selected by evolution; and in addition they have many secondary culturally >> determined values. What values will super-AI have and where will it get >> them and will they evolve? That seems to be the main research topic at the >> Machine Intelligence Research Institute. >> >> > Were all your values set at birth and driven by biology, or are some of > your values based on what you've since learned about the world? > > > Isn't that what I wrote just above? > > If values can be learned, and if morality is a field that has > objective truth, then why wouldn't a super intelligence will approach a > correct value system. > > > What would correct mean? Is vanilla *really* better than chocolate? > > I think there are core values - self-preservation, love of offspring, > desire for companionship, desire for power that are provided by evolution > and adapt people to live in extended families or small tribes. The other > values we learn from our culture are the result of cultural evolution > selecting values and ethics that let us realize our core values while > living in towns and cities and nations. > > Do you think in the long run that human society is evolving toward a more fair, more just, more correct system of values? If so, why can't a machine? Particularly one with the thinking capacity of a billion human minds operating a million times faster? Jason -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. Visit this group at http://groups.google.com/group/everything-list. For more options, visit https://groups.google.com/d/optout.

