On Mon, Apr 29, 2019, at 15:16, [email protected] wrote: > > > On Monday, April 29, 2019 at 7:27:26 AM UTC-5, Bruno Marchal wrote: >> >>> On 26 Apr 2019, at 15:33, [email protected] wrote: >>> >>> ** >>> ** >>> *AIs should have the same ethical protections as animals* >>> ** >>> *John Basl is assistant professor of philosophy at Northeastern University >>> in Boston* >>> >>> https://aeon.co/ideas/ais-should-have-the-same-ethical-protections-as-animals >>> >>> ... >>> >>> A puzzle and difficulty arises here because the scientific study of >>> consciousness has not reached a consensus about what consciousness is, and >>> how we can tell whether or not it is present. On some views – ‘liberal’ >>> views – for consciousness to exist requires nothing but a certain type of >>> well-organised information-processing, such as a flexible informational >>> model of the system in relation to objects in its environment, with guided >>> attentional capacities and long-term action-planning. We might be on the >>> verge of creating such systems already. On other views – ‘conservative’ >>> views – consciousness might require very specific biological features, such >>> as a brain very much like a mammal brain in its low-level structural >>> details: in which case we are nowhere near creating artificial >>> consciousness. >>> >>> It is unclear which type of view is correct or whether some other >>> explanation will in the end prevail. However, if a liberal view is correct, >>> we might soon be creating many subhuman AIs who will deserve ethical >>> protection. There lies the moral risk. >>> >>> Discussions of ‘AI risk’ normally focus on the risks that new AI >>> technologies might pose to us humans, such as taking over the world and >>> destroying us, or at least gumming up our banking system. Much less >>> discussed is the ethical risk we pose to the AIs, through our possible >>> mistreatment of them. >> >> The humans are still the main threat for the human. The idea to give human >> right to AI does not make music sense. It is part of the work of the AI to >> learn to defend themselves. We can be open mind, and listen, but defending >> their right can only threat the human right, I would say. In the theology of >> the machine, it can be proved that hell is paved with the good intentions … >> (amazingly enough, and accepting some definitions, of course). >> >>> >>> >>> My 'conservative' view: information processing (alone) does not achieve >>> experience (consciousness) processing. >> >> Mechanism makes you right on this, although it can depend how information >> processing is defined. Consciousness is not in the processing, but in truth, >> or in the semantic related to that processing,. The processing itself by is >> only a relative concept, where consciousness is an absolute thing. >> >> Bruno >> >>> https://codicalist.wordpress.com/2018/10/14/experience-processing/ >>> >>> >>> - >>> @philipthrift >>> <https://www.google.com/url?q=https%3A%2F%2Ftwitter.com%2Fphilipthrift&sa=D&sntz=1&usg=AFQjCNHxMatNNL0zmgjAMcrtu2m1txb6_A> >>> > > On "Consciousness is not in the processing, but in truth, or in the semantic > related to that processing, ..." I address in the next article: > > https://codicalist.wordpress.com/2018/12/14/material-semantics-for-unconventional-programming/ > > > But my mode of thinking is that of an engineer, not a truth-seeker.
I would say that engineering is a form of truth-seeking. Telmo. > > @philipthrift > <https://www.google.com/url?q=https%3A%2F%2Ftwitter.com%2Fphilipthrift&sa=D&sntz=1&usg=AFQjCNHxMatNNL0zmgjAMcrtu2m1txb6_A> > > -- > You received this message because you are subscribed to the Google Groups > "Everything List" group. > To unsubscribe from this group and stop receiving emails from it, send an > email to [email protected]. > To post to this group, send email to [email protected]. > Visit this group at https://groups.google.com/group/everything-list. > For more options, visit https://groups.google.com/d/optout. -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. Visit this group at https://groups.google.com/group/everything-list. For more options, visit https://groups.google.com/d/optout.

