I plan to be there. This should be fascinating! George Duncan georgeduncanart.com (505) 983-6895 Represented by ViVO Contemporary 725 Canyon Road Santa Fe, NM 87501
My art theme: Dynamic application of matrix order and luminous chaos. "Attempt what is not certain. Certainty may or may not come later. It may then be a valuable delusion." >From "Notes to myself on beginning a painting" by Richard Diebenkorn On Mon, May 18, 2015 at 4:54 PM, Tom Johnson <[email protected]> wrote: > FYI, Santa Fe folks. > -tj > > ============================================ > Tom Johnson > Institute for Analytic Journalism -- Santa Fe, NM USA > 505.577.6482(c) 505.473.9646(h) > Society of Professional Journalists <http://www.spj.org> - Region 9 > <http://www.spj.org/region9.asp> Director > Join more than 1,500 journalists Sept. 18-20 at > Excellence in Journalism 2015 in Orlando. #EIJ15 Orlando > http://www.jtjohnson.com [email protected] > ============================================ > > > Can We Reshape Humanity’s Deep Future?Possibilities & Risks of Artificial > Intelligence (AI), Human Enhancement, and Other Emerging Technologies > ------------------------------ > > WHERE: The James A. Little Theater <https://goo.gl/maps/NfQUu> at the New > Mexico School for the Deaf. > WHEN: Sunday, June 7, 2015, 2:00 pm > TICKETS: Book your seats now > <http://tickets.ticketssantafe.org/single/SelectSeating.aspx?p=2065> | More > info. <http://tickets.ticketssantafe.org/single/EventDetail.aspx?p=2065> > ------------------------------ > > Dr. Nick Bostrom spends much of his time calculating the possible rewards > and dangers of rapid technological advances — how such advances will likely > alter the course of human evolution and life as we know it. One useful > concept in untangling this puzzle is existential risk — the question of > whether an adverse outcome would end human intelligent life or drastically > curtail what we, in the infancy of the twenty-first century, would consider > a viable future. Figuring out how to reduce existential risk even slightly > brings into play an array of thought-provoking issues. In this engaging > lecture, Professor Bostrom will present the factors to be taken into > consideration: > > - Future technology and its capabilities > - Anthropics > - Population ethics > > > - Human enhancement ethics > - Game theory > - Fermi paradox > > ------------------------------ > About Nick Bostrom > > Nick Bostrom <http://www.nickbostrom.com/> is Professor in the Faculty of > Philosophy at Oxford University. He is the founding director of the Future > of Humanity Institute, a multidisciplinary research center that enables a > few exceptional mathematicians, philosophers, and scientists to think > carefully about global priorities and big questions for humanity. > > He is the recipient of a Eugene R. Gannon Award and has been listed on > *Foreign > Policy’s* Top 100 Global Thinkers list. He was included on *Prospect* > magazine’s World Thinkers list, the youngest person in the top fifteen from > all fields and the highest-ranked analytic philosopher. His writings have > been translated into twenty-four languages. > > Bostrom’s background includes physics, computational neuroscience, and > mathematical logic as well as philosophy. He is the author of some 200 > publications, including *Anthropic Bias* (Routledge, 2002), *Global > Catastrophic Risks* (ed., OUP, 2008), *Human Enhancement* (ed., OUP, > 2009), and *Superintelligence: Paths, Dangers, Strategies* (OUP, 2014), a *New > York Times* bestseller. He is best known for his work in five areas: > existential risk; the simulation argument; anthropics; impacts of future > technology; and implications of consequentialism for global strategy. He > has been referred to as one of the most important thinkers of our age. > ------------------------------ > > *SAR thanks these sponsors for underwriting this lecture:* > ------------------------------ > > > > *Slate,* Sept. 2014: > You Should Be Terrified of Superintelligent Machines > > In the recent discussion over the risks of developing superintelligent > machines—that is, machines with general intelligence greater than that of > humans—two narratives have emerged. One side argues that if a machine ever > achieved advanced intelligence, it would automatically know and care about > human values and wouldn’t pose a threat to us. The opposing side argues > that artificial intelligence would “want” to wipe humans out, either out of > revenge or an intrinsic desire for survival. > > As it turns out, both of these views are wrong. > > Read more > > <http://www.slate.com/articles/technology/future_tense/2014/09/will_artificial_intelligence_turn_on_us_robots_are_nothing_like_humans_and.html> > > *Aeon Magazine,* Feb. 2013: > Omens > > To understand why an AI might be dangerous, you have to avoid > anthropomorphising it. When you ask yourself what it might do in a > particular situation, you can’t answer by proxy. You can't picture a > super-smart version of yourself floating above the situation. Human > cognition is only one species of intelligence, one with built-in impulses > like empathy that colour the way we see the world, and limit what we are > willing to do to accomplish our goals. But these biochemical impulses > aren’t essential components of intelligence. They’re incidental software > applications, installed by aeons of evolution and culture. Bostrom told me > that it’s best to think of an AI as a primordial force of nature, like a > star system or a hurricane — something strong, but indifferent. > > Read more > > <http://aeon.co/magazine/philosophy/ross-andersen-human-extinction/> > > *TEDx/Youtube,* Apr. 2015: > TEDx Talks: What happens when our computers get smarter than we are? > > Artificial intelligence is getting smarter by leaps and bounds — within > this century, research suggests, a computer AI could be as “smart” as a > human being. Nick Bostrom asks us to think hard about the world we're > building right now, driven by thinking machines. Will our smart machines > help to preserve humanity and our values — or will they have values of > their own? > Become a Member of SAR! > > A School for Advanced Research membership opens doors to exploring a world > of ideas about past and present peoples around the world and in the > Southwest, as well as Native American life and arts. Become an SAR member > today. Individual memberships start at $50. *Click here to join!* > <http://sarweb.org/?become_a_member> > > > > > Header image, copyright: / 123RF Stock Photo > <http://www.123rf.com/profile_spaxia> > > ============================== > Dorothy H. Bracey -- Santa Fe, NM US > [email protected] > ============================== > > > > ============================================================ > FRIAM Applied Complexity Group listserv > Meets Fridays 9a-11:30 at cafe at St. John's College > to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com >
============================================================ FRIAM Applied Complexity Group listserv Meets Fridays 9a-11:30 at cafe at St. John's College to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com
