Thanks so much, Tom. I've got my ticket. Sounds wonderful. See you there.
On Tue, May 19, 2015 at 10:22 PM, John Dobson <[email protected]> wrote: > cdobson@okstate,edu > > On Mon, May 18, 2015 at 4:54 PM, Tom Johnson <[email protected]> wrote: > >> FYI, Santa Fe folks. >> -tj >> >> ============================================ >> Tom Johnson >> Institute for Analytic Journalism -- Santa Fe, NM USA >> 505.577.6482(c) 505.473.9646(h) >> Society of Professional Journalists <http://www.spj.org> - Region 9 >> <http://www.spj.org/region9.asp> Director >> Join more than 1,500 journalists Sept. 18-20 at >> Excellence in Journalism 2015 in Orlando. #EIJ15 Orlando >> http://www.jtjohnson.com [email protected] >> ============================================ >> >> >> Can We Reshape Humanity’s Deep Future?Possibilities & Risks of >> Artificial Intelligence (AI), Human Enhancement, and Other Emerging >> Technologies >> ------------------------------ >> >> WHERE: The James A. Little Theater <https://goo.gl/maps/NfQUu> at the >> New Mexico School for the Deaf. >> WHEN: Sunday, June 7, 2015, 2:00 pm >> TICKETS: Book your seats now >> <http://tickets.ticketssantafe.org/single/SelectSeating.aspx?p=2065> | More >> info. <http://tickets.ticketssantafe.org/single/EventDetail.aspx?p=2065> >> ------------------------------ >> >> Dr. Nick Bostrom spends much of his time calculating the possible rewards >> and dangers of rapid technological advances — how such advances will likely >> alter the course of human evolution and life as we know it. One useful >> concept in untangling this puzzle is existential risk — the question of >> whether an adverse outcome would end human intelligent life or drastically >> curtail what we, in the infancy of the twenty-first century, would consider >> a viable future. Figuring out how to reduce existential risk even slightly >> brings into play an array of thought-provoking issues. In this engaging >> lecture, Professor Bostrom will present the factors to be taken into >> consideration: >> >> - Future technology and its capabilities >> - Anthropics >> - Population ethics >> >> >> - Human enhancement ethics >> - Game theory >> - Fermi paradox >> >> ------------------------------ >> About Nick Bostrom >> >> Nick Bostrom <http://www.nickbostrom.com/> is Professor in the Faculty >> of Philosophy at Oxford University. He is the founding director of the >> Future of Humanity Institute, a multidisciplinary research center that >> enables a few exceptional mathematicians, philosophers, and scientists to >> think carefully about global priorities and big questions for humanity. >> >> He is the recipient of a Eugene R. Gannon Award and has been listed on >> *Foreign >> Policy’s* Top 100 Global Thinkers list. He was included on *Prospect* >> magazine’s World Thinkers list, the youngest person in the top fifteen from >> all fields and the highest-ranked analytic philosopher. His writings have >> been translated into twenty-four languages. >> >> Bostrom’s background includes physics, computational neuroscience, and >> mathematical logic as well as philosophy. He is the author of some 200 >> publications, including *Anthropic Bias* (Routledge, 2002), *Global >> Catastrophic Risks* (ed., OUP, 2008), *Human Enhancement* (ed., OUP, >> 2009), and *Superintelligence: Paths, Dangers, Strategies* (OUP, 2014), >> a *New York Times* bestseller. He is best known for his work in five >> areas: existential risk; the simulation argument; anthropics; impacts of >> future technology; and implications of consequentialism for global >> strategy. He has been referred to as one of the most important thinkers of >> our age. >> ------------------------------ >> >> *SAR thanks these sponsors for underwriting this lecture:* >> ------------------------------ >> >> >> >> *Slate,* Sept. 2014: >> You Should Be Terrified of Superintelligent Machines >> >> In the recent discussion over the risks of developing superintelligent >> machines—that is, machines with general intelligence greater than that of >> humans—two narratives have emerged. One side argues that if a machine ever >> achieved advanced intelligence, it would automatically know and care about >> human values and wouldn’t pose a threat to us. The opposing side argues >> that artificial intelligence would “want” to wipe humans out, either out of >> revenge or an intrinsic desire for survival. >> >> As it turns out, both of these views are wrong. >> >> Read more > >> <http://www.slate.com/articles/technology/future_tense/2014/09/will_artificial_intelligence_turn_on_us_robots_are_nothing_like_humans_and.html> >> >> *Aeon Magazine,* Feb. 2013: >> Omens >> >> To understand why an AI might be dangerous, you have to avoid >> anthropomorphising it. When you ask yourself what it might do in a >> particular situation, you can’t answer by proxy. You can't picture a >> super-smart version of yourself floating above the situation. Human >> cognition is only one species of intelligence, one with built-in impulses >> like empathy that colour the way we see the world, and limit what we are >> willing to do to accomplish our goals. But these biochemical impulses >> aren’t essential components of intelligence. They’re incidental software >> applications, installed by aeons of evolution and culture. Bostrom told me >> that it’s best to think of an AI as a primordial force of nature, like a >> star system or a hurricane — something strong, but indifferent. >> >> Read more > >> <http://aeon.co/magazine/philosophy/ross-andersen-human-extinction/> >> >> *TEDx/Youtube,* Apr. 2015: >> TEDx Talks: What happens when our computers get smarter than we are? >> >> Artificial intelligence is getting smarter by leaps and bounds — within >> this century, research suggests, a computer AI could be as “smart” as a >> human being. Nick Bostrom asks us to think hard about the world we're >> building right now, driven by thinking machines. Will our smart machines >> help to preserve humanity and our values — or will they have values of >> their own? >> Become a Member of SAR! >> >> A School for Advanced Research membership opens doors to exploring a >> world of ideas about past and present peoples around the world and in the >> Southwest, as well as Native American life and arts. Become an SAR member >> today. Individual memberships start at $50. *Click here to join!* >> <http://sarweb.org/?become_a_member> >> >> >> >> >> Header image, copyright: / 123RF Stock Photo >> <http://www.123rf.com/profile_spaxia> >> >> ============================== >> Dorothy H. Bracey -- Santa Fe, NM US >> [email protected] >> ============================== >> >> >> >> ============================================================ >> FRIAM Applied Complexity Group listserv >> Meets Fridays 9a-11:30 at cafe at St. John's College >> to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com >> > > > ============================================================ > FRIAM Applied Complexity Group listserv > Meets Fridays 9a-11:30 at cafe at St. John's College > to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com > -- Merle Lefkoff, Ph.D. President, Center for Emergent Diplomacy Santa Fe, New Mexico, USA [email protected] mobile: (303) 859-5609 skype: merlelefkoff
============================================================ FRIAM Applied Complexity Group listserv Meets Fridays 9a-11:30 at cafe at St. John's College to unsubscribe http://redfish.com/mailman/listinfo/friam_redfish.com
