Does not look like there is a nice formatting option. They do have an enable 
conversations setting, but I do not thing that provides formatting and 
indentation. If I have some free time -- which I have very little of 
unfortunately -- I will look.




________________________________
 From: Terren Suydam <[email protected]>
To: [email protected] 
Sent: Friday, September 5, 2014 4:02 PM
Subject: Re: Fwd: The Machine Intelligence Research Institute Blog
 


I left Yahoo mail five years ago because they do such a terrible job of 
engineering. I have embraced the Google. Thanks for whatever you can do. 
Usually email clients offer a couple of modes for how to include the original 
email... Is there a different mode you can try?
Terren
On Sep 5, 2014 5:50 PM, "'Chris de Morsella' via Everything List" 
<[email protected]> wrote:

Terren - You should forward your concerns to the folks who code the yahoo 
webmail client... when I am at work I use its webmail client, which does a poor 
job of threading a conversation. Will try to remember that and put in manual 
'>>' marks to show what I am replying to.
>
>
>
>________________________________
> From: Terren Suydam <[email protected]>
>To: [email protected] 
>Sent: Friday, September 5, 2014 12:47 PM
>Subject: Re: Fwd: The Machine Intelligence Research Institute Blog
> 
>
>
>Chris, is there a way you can improve your email client?  Sometimes your 
>responses are very hard to detect because they're at the same indentation and 
>font as the one you are reply to, as below. Someone new to the conversation 
>would have no way of knowing that Brent did not write that entire thing, as 
>you didn't sign your name.
>
>
>Thanks, Terren
>
>
>
>
>
>
>On Fri, Sep 5, 2014 at 2:15 PM, 'Chris de Morsella' via Everything List 
><[email protected]> wrote:
>
>
>>
>>
>>
>>
>>________________________________
>> From: meekerdb <[email protected]>
>>To: EveryThing <[email protected]> 
>>Sent: Friday, September 5, 2014 9:47 AM
>>Subject: Fwd: The Machine Intelligence Research Institute Blog
>> 
>>
>>
>>For you who are worried about the threat of artificial intelligence, MIRI 
>>seems to make it their main concern.  Look up their website and subscribe.  
>>On my list of existential threats it comes well below natural stupidity.
>>
>>
>>On mine as well... judging by how far the google car still has to go before 
>>it does not drive straight into that pothole or require that its every route 
>>be very carefully mapped down to the level of each single driveway. Real 
>>world AI is still mired in the stubbornly, dumb as sand nature of our silicon 
>>based deterministic logic gate architecture.
>>Much higher chance that we will blow ourselves up in some existentially 
>>desperate final energy war, or so poison our earth's biosphere that systemic 
>>collapse is triggered and the deep ocean's flip into an anoxic state favoring 
>>the hydrogen sulfide producing microorganisms that are poisoned by oxygen, 
>>resulting in another great belch of poisonous (to animals and plants) 
>>hydrogen sulfide into the planet's atmosphere -- as occurred during the great 
>>Permian extinction.
>>Speaking of which has anyone read the recent study that concludes the current 
>>anthropocene boundary layer extinction rate is more than one thousand times 
>>the average extinction level that prevailed from the last great extinction 
>>(Jurassic) until now. See: Extinctions during human era one thousand times 
>>more than before
>>
>>Brent
>>
>> 
>>
>>
>>-------- Original Message -------- 
>>Subject: The Machine Intelligence Research Institute Blog 
>>Date: Fri, 05 Sep 2014 12:07:00 +0000 
>>From: Machine Intelligence Research Institute » Blog <[email protected]> 
>>To: [email protected] 
>>
>>
>>The Machine Intelligence Research Institute Blog  
>> 
>>________________________________
>> 
>>John Fox on AI safety 
>>Posted: 04 Sep 2014 12:00 PM PDT
>> John Fox is an interdisciplinary scientist with theoretical interests in AI 
>> and computer science, and an applied focus in medicine and medical software 
>> engineering. After training in experimental psychology at Durham and 
>> Cambridge Universities and post-doctoral fellowships at CMU and Cornell in 
>> the USA and UK (MRC) he joined the Imperial Cancer Research Fund (now Cancer 
>> Research UK) in 1981 as a researcher in medical AI. The group’s research was 
>> explicitly multidisciplinary and it subsequently made significant 
>> contributions in basic computer science, AI and medical informatics, and 
>> developed a number of successful technologies which have been commercialised.
>>In 1996 he and his team were awarded the 20th Anniversary Gold Medal of the 
>>European Federation of Medical Informatics for the development of PROforma, 
>>arguably the first formal computer language for modeling clinical decision 
>>and processes. Fox has published widely in computer science, cognitive 
>>science and biomedical engineering, and was the founding editor of the 
>>Knowledge Engineering Review  (Cambridge University Press). Recent 
>>publications include a research monograph Safe and Sound: Artificial 
>>Intelligence in Hazardous Applications (MIT Press, 2000) which deals with the 
>>use of AI in safety-critical fields such as medicine.
>>Luke Muehlhauser: You’ve spent many years studying AI safety issues, in 
>>particular in medical contexts, e.g. in your 2000 book with Subrata Das, Safe 
>>and Sound: Artificial Intelligence in Hazardous Applications. What kinds of 
>>AI safety challenges have you focused on in the past decade or so?
>>________________________________
>> 
>>John Fox: From my first research job, as a post-doc with AI founders Allen 
>>Newell and Herb Simon at CMU, I have been interested in computational 
>>theories of high level cognition. As a cognitive scientist I have been 
>>interested in theories that subsume a range of cognitive functions, from 
>>perception and reasoning to the uses of knowledge in autonomous 
>>decision-making. After I came back to the UK in 1975 I began to combine my 
>>theoretical interests with the practical goals of designing and deploying AI 
>>systems in medicine.
>>Since our book was published in 2000 I have been committed to testing the 
>>ideas in it by designing and deploying many kind of clinical systems, and 
>>demonstrating that AI techniques can significantly improve quality and safety 
>>of clinical decision-making and process management. Patient safety is 
>>fundamental to clinical practice so, alongside the goals of building systems 
>>that can improve on human performance, safety and ethics have always been 
>>near the top of my research agenda.
>>________________________________
>> 
>>Luke Muehlhauser: Was it straightforward to address issues like safety and 
>>ethics in practice?
>>________________________________
>> 
>>John Fox: While our concepts and technologies have proved to be clinically 
>>successful we have not achieved everything we hoped for. Our attempts to 
>>ensure, for example, that practical and commercial deployments of AI 
>>technologies should explicitly honor ethical principles and carry out active 
>>safety management have not yet achieved the traction that we need to achieve. 
>>I regard this as a serious cause for concern, and unfinished business in both 
>>scientific and engineering terms.
>>The next generation of large-scale knowledge based systems and software 
>>agents that we are now working on will be more intelligent and will have far 
>>more autonomous capabilities than current systems. The challenges for human 
>>safety and ethical use of AI that this implies are beginning to mirror those 
>>raised by the singularity hypothesis. We have much to learn from singularity 
>>researchers, and perhaps our experience in deploying autonomous agents in 
>>human healthcare will offer opportunities to ground some of the singularity 
>>debates as well.
>>________________________________
>> 
>>Luke: You write that your “attempts to ensure… [that] commercial deployments 
>>of AI technologies should… carry out active safety management” have not yet 
>>received as much traction as you would like. Could you go into more detail on 
>>that? What did you try to accomplish on this front that didn’t get adopted by 
>>others, or wasn’t implemented?
>>________________________________
>> 
>>John: Having worked in medical AI from the early ‘seventies I have always 
>>been keenly aware that while AI can help to mitigate the effects of human 
>>error there is a potential downside too. AI systems could be programmed 
>>incorrectly, or their knowledge could prescribe inappropriate practices, or 
>>they could have the effect of deskilling the human professionals who have the 
>>final responsibility for their patients. Despite well-known limitations of 
>>human cognition people remain far and away the most versatile and creative 
>>problem solvers on the planet.
>>In the early ‘nineties I had the opportunity to set up a project whose goal 
>>was to establish a rigorous framework for the design and implementation of AI 
>>systems for safety critical applications. Medicine was our practical focus 
>>but the RED project1 was aimed at the development of a general architecture 
>>for the design of autonomous agents that could be trusted to make decisions 
>>and carry out plans as reliably and safely as possible, certainly to be as 
>>competent and hence as trustworthy as human agents in comparable tasks. This 
>>is obviously a hard problem but we made sufficient progress on theoretical 
>>issues and design principles that I thought there was a good chance the 
>>techniques might be applicable in medical AI and maybe even more widely.
>>I thought AI was like medicine, where we all take it for granted that medical 
>>equipment and drug companies have a duty of care to show that their products 
>>are effective and safe before they can be certificated for commercial use. I 
>>also assumed that AI researchers would similarly recognize that we have a 
>>“duty of care” to all those potentially affected by poor engineering or 
>>misuse in safety critical settings but this was naïve. The commercial tools 
>>that have been based on the technologies derived from AI research have to 
>>date focused on just getting and keeping customers and safety always takes a 
>>back seat.
>>In retrospect I should have predicted that making sure that AI products are 
>>safe is not going to capture the enthusiasm of commercial suppliers. If you 
>>compare AI apps with drugs we all know that pharmaceutical companies have to 
>>be firmly regulated to make sure they fulfill their duty of care to their 
>>customers and patients. However proving drugs are safe is expensive and also 
>>runs the risk of revealing that your new wonder-drug isn’t even as effective 
>>as you claim! It’s the same with AI.
>>I continue to be surprised how optimistic software developers are – they 
>>always seem to have supreme confidence that worst-case scenarios wont happen, 
>>or that if they do happen then their management is someone else’s 
>>responsibility. That kind of technical over-confidence has led to countless 
>>catastrophes in the past, and it amazes me that it persists.
>>There is another piece to this, which concerns the roles and responsibilities 
>>of AI researchers. How many of us take the risks of AI seriously so that it 
>>forms a part of our day-to-day theoretical musings and influences our 
>>projects? MIRI has put one worst case scenario in front of us – the 
>>possibility that our creations might one day decide to obliterate us – but so 
>>far as I can tell the majority of working AI professionals either see safety 
>>issues as irrelevant to the pursuit of interesting scientific questions or, 
>>like the wider public, that the issues are just science fiction.
>>I think experience in medical AI trying to articulate and cope with human 
>>risk and safety may have a couple of important lessons for the wider AI 
>>community. First we have a duty of care that professional scientists cannot 
>>responsibly ignore. Second, the AI business will probably need to be 
>>regulated, in much the same way as the pharmaceutical business is. If these 
>>propositions are correct then the AI research community would be wise to 
>>engage with and lead on discussions around safety issues if it wants to 
>>ensure that the regulatory framework that we get is to our liking!
>>________________________________
>> 
>>Luke: Now you write, “That kind of technical over-confidence has led to 
>>countless catastrophes in the past…” What are some example “catastrophes” 
>>you’re thinking of?
>>________________________________
>> 
>>John:
>>Psychologists have known for years that human decision-making is flawed, even 
>>if amazingly creative sometimes, and overconfidence is an important source of 
>>error in routine settings. A large part of the motivation for applying AI in 
>>medicine comes from the knowledge that, in the words of the Institute of 
>>Medicine, “To err is human” and overconfidence is an established cause of 
>>clinical mistakes.2
>>Over-confidence and its many relatives (complacency, optimism, arrogance and 
>>the like) have a huge influence on our personal successes and failures, and 
>>our collective futures. The outcomes of the US and UK’s recent adventures 
>>around the world can be easily identified as consequences of overconfidence, 
>>and it seems to me that the polarized positions about global warming and 
>>planetary catastrophe are both expressions of overconfidence, just in 
>>opposite directions.
>>________________________________
>> 
>>Luke: Looking much further out… if one day we can engineer AGIs, do you think 
>>we are likely to figure out how to make them safe?
>>________________________________
>> 
>>John: History says that making any technology safe is not an easy business. 
>>It took quite a few boiler explosions before high-pressure steam engines got 
>>their iconic centrifugal governors. Ensuring that new medical treatments are 
>>safe as well as effective is famously difficult and expensive. I think we 
>>should assume that getting to the point where an AGI manufacturer could 
>>guarantee its products are safe will be a hard road, and it is possible that 
>>guarantees are not possible in principle. We are not even clear yet what it 
>>means to be “safe”, at least not in computational terms.
>>It seems pretty obvious that entry level robotic products like the robots 
>>that carry out simple domestic chores or the “nursebots” that are being 
>>trialed for hospital use, have such a simple repertoire of behaviors that it 
>>should not be difficult to design their software controllers to operate 
>>safely in most conceivable circumstances. Standard safety engineering 
>>techniques like HAZOP3 are probably up to the job I think, and where software 
>>failures simply cannot be tolerated software engineering techniques like 
>>formal specification and model-checking are available.
>>There is also quite a lot of optimism around more challenging robotic 
>>applications like autonomous vehicles and medical robotics. Moustris et al.4 
>>say that autonomous surgical robots are emerging that can be used in various 
>>roles, automating important steps in complex operations like open-heart 
>>surgery for example, and they expect them to become standard in – and to 
>>revolutionize the practice of – surgery. However at this point it doesn’t 
>>seem to me that surgical robots with a significant cognitive repertoire are 
>>feasible and a human surgeon will be in the loop for the foreseeable future.
>>________________________________
>> 
>>Luke: So what might artificial intelligence learn from natural intelligence?
>>________________________________
>> 
>>As a cognitive scientist working in medicine my interests are co-extensive 
>>with those of scientists working on AGIs. Medicine is such a vast domain that 
>>practicing it safely requires the ability to deal with countless clinical 
>>scenarios and interactions and even when working in a single specialist 
>>subfield requires substantial knowledge from other subfields. So much so that 
>>it is now well known that even very experienced humans with a large clinical 
>>repertoire are subject to significant levels of error.5 An artificial 
>>intelligence that could be helpful across medicine will require great 
>>versatility, and this will require a general understanding of medical 
>>expertise and a range of cognitive capabilities like reasoning, 
>>decision-making, planning, communication, reflection, learning and so forth.
>>If human experts are not safe is it well possible to ensure that an AGI, 
>>however sophisticated, will be? I think that it is pretty clear that the 
>>range of techniques currently available for assuring system safety will be 
>>useful in making specialist AI systems reliable and minimizing the likelihood 
>>of errors in situations that their human designers can anticipate. However, 
>>AI systems with general intelligence will be expected to address scenarios 
>>and hazards that are beyond us to solve currently and often beyond designers 
>>even to anticipate. I am optimistic but at the moment I don’t see any 
>>convincing reason to believe that we have the techniques that would be 
>>sufficient to guarantee that a clinical super-intelligence is safe, let alone 
>>an AGI that might be deployed in many domains.
>> 
>>________________________________
>> 
>>Luke: Thanks, John!
>>________________________________
>> 
>>      1. Rigorously Engineered Decisions
>>      2. Overconfidence in major disasters: 
>>• D. Lucas. Understanding the Human Factor in Disasters. Interdisciplinary 
>>Science Reviews. Volume 17 Issue 2 (01 June 1992), pp. 185-190.
>>• “Nuclear safety and security.
>>Psychology of overconfidence:
>>• Overconfidence effect.
>>• C. Riordan. Three Ways Overconfidence Can Make a Fool of You Forbes 
>>Leadership Forum.
>>Overconfidence in medicine:
>>• R. Hanson. Overconfidence Erases Doc Advantage. Overcoming Bias, 2007.
>>• E. Berner, M. Graber. Overconfidence as a Cause of Diagnostic Error in 
>>Medicine. The American Journal of Medicine. Volume 121, Issue 5, Supplement, 
>>Pages S2–S23, May 2008.
>>• T. Ackerman. Doctors overconfident, study finds, even in hardest cases. 
>>Houston Chronicle, 2013.
>>General technology example:
>>• J. Vetter, A. Benlian, T. Hess. Overconfidence in IT Investment Decisions: 
>>Why Knowledge can be a Boon and Bane at the same Time. ICIS 2011 Proceedings. 
>>Paper 4. December 6, 2011.
>>      3. Hazard and operability study
>>      4. Int J Med Robotics Comput Assist Surg 2011; 7: 375–39
>>      5. A. Ford. Domestic Robotics – Leave it to Roll-Oh, our Fun loving 
>> Retrobot. Institute for Ethics and Emerging Technologies, 2014.
>>The post John Fox on AI safety appeared first on Machine Intelligence 
>>Research Institute. 
>>Daniel Roy on probabilistic programming and AI 
>>Posted: 04 Sep 2014 08:03 AM PDT
>> Daniel Roy is an Assistant Professor of Statistics at the University of 
>> Toronto. Roy earned an S.B. and M.Eng. in Electrical Engineering and 
>> Computer Science, and a Ph.D. in Computer Science, from MIT.  His 
>> dissertation on probabilistic programming received the department’s George M 
>> Sprowls Thesis Award.  Subsequently, he held a Newton International 
>> Fellowship of the Royal Society, hosted by the Machine Learning Group at the 
>> University of Cambridge, and then held a Research Fellowship at Emmanuel 
>> College. Roy’s research focuses on theoretical questions that mix computer 
>> science, statistics, and probability.
>>Luke Muehlhauser: The abstract of Ackerman, Freer, and Roy (2010) begins:
>>As inductive inference and machine learning methods in computer science see 
>>continued success, researchers are aiming to describe even more complex 
>>probabilistic models and inference algorithms. What are the limits of 
>>mechanizing probabilistic inference? We investigate the computability of 
>>conditional probability… and show that there are computable joint 
>>distributions with noncomputable conditional distributions, ruling out the 
>>prospect of general inference algorithms.
>>In what sense does your result (with Ackerman & Freer) rule out the prospect 
>>of general inference algorithms?
>>________________________________
>> 
>>Daniel Roy: First, it’s important to highlight that when we say 
>>“probabilistic inference” we are referring to the problem of computing 
>>conditional probabilities, while highlighting the role of conditioning in 
>>Bayesian statistical analysis.
>>Bayesian inference centers around so-called posterior distributions. From a 
>>subjectivist standpoint, the posterior represents one’s updated beliefs after 
>>seeing (i.e., conditioning on) the data. Mathematically, a posterior 
>>distribution is simply a conditional distribution (and every conditional 
>>distribution can be interpreted as a posterior distribution in some 
>>statistical model), and so our study of the computability of conditioning 
>>also bears on the problem of computing posterior distributions, which is 
>>arguably one of the core computational problems in Bayesian analyses.
>>Second, it’s important to clarify what we mean by “general inference”. In 
>>machine learning and artificial intelligence (AI), there is a long tradition 
>>of defining formal languages in which one can specify probabilistic models 
>>over a collection of variables. Defining distributions can be difficult, but 
>>these languages can make it much more straightforward.
>>The goal is then to design algorithms that can use these representations to 
>>support important operations, like computing conditional distributions. 
>>Bayesian networks can be thought of as such a language: You specify a 
>>distribution over a collection of variables by specifying a graph over these 
>>variables, which breaks down the entire distribution into “local” conditional 
>>distributions corresponding with each node, which are themselves often 
>>represented as tables of probabilities (at least in the case where all 
>>variables take on only a finite set of values). Together, the graph and the 
>>local conditional distributions determine a unique distribution over all the 
>>variables.
>>An inference algorithms that support the entire class of all finite, 
>>discrete, Bayesian networks might be called general, but as a class of 
>>distributions, those having finite, discrete Bayesian networks is a rather 
>>small one.
>>In this work, we are interested in the prospect of algorithms that work on 
>>very large classes of distributions. Namely, we are considering the class of 
>>samplable distributions, i.e., the class of distributions for which there 
>>exists a probabilistic program that can generate a sample using, e.g., 
>>uniformly distributed random numbers or independent coin flips as a source of 
>>randomness. The class of samplable distributions is a natural one: indeed it 
>>is equivalent to the class of computable distributions, i.e., those for which 
>>we can devise algorithms to compute lower bounds on probabilities from 
>>descriptions of open sets. The class of samplable distributions is also 
>>equivalent to the class of distributions for which we can compute 
>>expectations from descriptions of bounded continuous functions.
>>The class of samplable distributions is, in a sense, the richest class you 
>>might hope to deal with. The question we asked was: is there an algorithm 
>>that, given a samplable distribution on two variables X and Y, represented by 
>>a program that samples values for both variables, can compute the conditional 
>>distribution of, say, Y given X=x, for almost all values for X? When X takes 
>>values in a finite, discrete set, e.g., when X is binary valued, there is a 
>>general algorithm, although it is inefficient. But when X is continuous, 
>>e.g., when it can take on every value in the unit interval [0,1], then 
>>problems can arise. In particular, there exists a distribution on a pair of 
>>numbers in [0,1] from which one can generate perfect samples, but for which 
>>it is impossible to compute conditional probabilities for one of the 
>>variables given the other. As one might expect, the proof reduces the halting 
>>problem to that of conditioning a specially crafted distribution.
>>This pathological distribution rules out the possibility of a general 
>>algorithm for conditioning (equivalently, for probabilistic inference). The 
>>paper ends by giving some further conditions that, when present, allow one to 
>>devise general inference algorithms. Those familiar with computing 
>>conditional distributions for finite-dimensional statistical models will not 
>>be surprised that conditions necessary for Bayes’ theorem are one example.
>>
>>________________________________
>> 
>>Luke: In your dissertation (and perhaps elsewhere) you express a particular 
>>interest in the relevance of probabilistic programming to AI, including the 
>>original aim of AI to build machines which rival the general intelligence of 
>>a human. How would you describe the relevance of probabilistic programming to 
>>the long-term dream of AI?
>>________________________________
>> 
>>Daniel: If you look at early probabilistic programming systems, they were 
>>built by AI researchers: De Raedt, Koller, McAllester, Muggleton, Pfeffer, 
>>Poole, Sato, to name a few. The Church language, which was introduced in 
>>joint work with Bonawitz, Mansinghka, Goodman, and Tenenbaum while I was a 
>>graduate student at MIT, was conceived inside a cognitive science laboratory, 
>>foremost to give us a language rich enough to express the range of models 
>>that people were inventing all around us. So, for me, there’s always been a 
>>deep connection. On the other hand, the machine learning community as a whole 
>>is somewhat allergic to AI and so the pitch to that community has more often 
>>been pragmatic: these systems may someday allow experts to conceive, 
>>prototype, and deploy much larger probabilistic systems, and at the same 
>>time, empower a much larger community of nonexperts to use probabilistic 
>>modeling techniques to understand their data. This is the basis for
 the DARPA PPAML program, which is funding 8 or so teams to engineer scalable 
systems over the next 4 years.
>>From an AI perspective, probabilistic programs are an extremely general 
>>representation of knowledge, and one that identifies uncertainty with 
>>stochastic computation. Freer, Tenenbaum, and I recently wrote a book chapter 
>>for the Turing centennial that uses a classical medical diagnosis example to 
>>showcase the flexibility of probabilistic programs and a general QUERY 
>>operator for performing probabilistic conditioning. Admittedly, the book 
>>chapter ignores the computational complexity of the QUERY operator, and any 
>>serious proposal towards AI cannot do this indefinitely. Understanding when 
>>we can hope to efficiently update our knowledge in light of new observations 
>>is a rich source of research questions, both applied and theoretical, 
>>spanning not only AI and machine learning, but also statistics, probability, 
>>physics, theoretical computer science, etc.
>>________________________________
>> 
>>Luke: Is it fair to think of QUERY as a “toy model” that we can work with in 
>>concrete ways to gain more general insights into certain parts of the 
>>long-term AI research agenda, even though QUERY is unlikely to be directly 
>>implemented in advanced AI systems? (E.g. that’s how I think of AIXI.)
>>________________________________
>> 
>>Daniel: I would hesitate to call QUERY a toy model. Conditional probability 
>>is a difficult concept to master, but, for those adept at reasoning about the 
>>execution of programs, QUERY demystifies the concept considerably. QUERY is 
>>an important conceptual model of probabilistic conditioning.
>>That said, the simple guess-and-check algorithm we present in our Turing 
>>article runs in time inversely proportional to the probability of the 
>>event/data on which one is conditioning. In most statistical settings, the 
>>probability of a data set decays exponentially towards 0 as a function of the 
>>number of data points, and so guess-and-check is only useful for reasoning 
>>with toy data sets in these settings. It should come as no surprise to hear 
>>that state-of-the-art probabilistic programming systems work nothing like 
>>this.
>>On the other hand, QUERY, whether implemented in a rudimentary fashion or 
>>not, can be used to represent and reason probabilistically about arbitrary 
>>computational processes, whether they are models of the arrival time of spam, 
>>the spread of disease through networks, or the light hitting our retinas. 
>>Computer scientists, especially those who might have had a narrow view of the 
>>purview of probability and statistics, will see a much greater overlap 
>>between these fields and their own once they understand QUERY.
>>To those familiar with AIXI, the difference is hopefully clear: QUERY 
>>performs probabilistic reasoning in a model given as input. AIXI, on the 
>>other hand, is itself a “universal” model that, although not computable, 
>>would likely predict (hyper)intelligent behavior, were we (counterfactually) 
>>able to perform the requisite probabilistic inferences (and feed it enough 
>>data). Hutter gives an algorithm implementing an approximation to AIXI, but 
>>its computational complexity still scales exponentially in space. AIXI is 
>>fascinating in many ways: If we ignore computational realities, we get a 
>>complete proposal for AI. On the other hand, AIXI and its approximations take 
>>maximal advantage of this computational leeway and are, therefore, ultimately 
>>unsatisfying. For me, AIXI and related ideas highlight that AI must be as 
>>much a study of the particular as it of the universal. Which potentially 
>>unverifiable, but useful, assumptions will enable us to efficiently
 represent, update, and act upon knowledge under uncertainty?
>>________________________________
>> 
>>Luke: You write that “AI must be as much a study of the particular as it is 
>>of the universal.” Naturally, most AI scientists are working on the 
>>particular, the near term, the applied. In your view, what are some other 
>>examples of work on the universal, in AI? Schmidhuber’s Gödel machine comes 
>>to mind, and also some work that is as likely to be done in a logic or formal 
>>philosophy department as a computer science department — e.g. perhaps work on 
>>logical priors — but I’d love to hear what kinds of work you’re thinking of.
>>________________________________
>> 
>>Daniel: I wouldn’t equate any two of the particular, near-term, or applied. 
>>By the word particular, I am referring to, e.g., the way that our environment 
>>affects, but is also affected by, our minds, especially through society. More 
>>concretely, both the physical spaces in which most of us spend our days and 
>>the mental concepts we regularly use to think about our daily activities are 
>>products of the human mind. But more importantly, these physical and mental 
>>spaces are necessarily ones that are easily navigated by our minds. The 
>>coevolution by which this interaction plays out is not well studied in the 
>>context of AI. And to the extent that this cycle dominates, we would expect a 
>>universal AI to be truly alien. On the other hand, exploiting the constraints 
>>of human constructs may allow us to build more effective AIs.
>>As for the universal, I have an interest in the way that noise can render 
>>idealized operations computable or even efficiently computable. In our work 
>>on the computability of conditioning that came up earlier in the discussion, 
>>we show that adding sufficiently smooth independent noise to a random 
>>variable allows us to perform conditioning in situations where we would not 
>>have been able to otherwise. There are examples of this idea elsewhere. For 
>>example, Braverman, Grigo, and Rojas study noise and intractability in 
>>dynamical systems. Specifically, they show that computing the invariant 
>>measure characterizing the long-term statistical behavior of dynamical 
>>systems is not possible. The road block is the computational power of the 
>>dynamical system itself. The addition of a small amount of noise to the 
>>dynamics, however, decreases the computational power of the dynamical system, 
>>and suffices to make the invariant measure computable. In a world subject to
 noise (or, at least, well modeled as such), it seems that many theoretical 
obstructions melt away.
>>________________________________
>> 
>>Luke: Thanks, Daniel!
>>The post Daniel Roy on probabilistic programming and AI appeared first on 
>>Machine Intelligence Research Institute. 
>>You are subscribed to email updates from Machine Intelligence Research 
>>Institute » Blog 
>>To stop receiving these emails, you may unsubscribe now. Email delivery 
>>powered by Google 
>>Google Inc., 20 West Kinzie, Chicago IL USA 60610 
>>
>>
>>  
>>             
>>Extinctions during human era one thousand times more th...
>>The gravity of the world's current extinction rate becomes clearer upon 
>>knowing what it was before people came along. A new estimate finds that 
>>species d...  
>>View on www.sciencedaily.com Preview by Yahoo  
>>  
>> 
-- 
>>You received this message because you are subscribed to the Google Groups 
>>"Everything List" group.
>>To unsubscribe from this group and stop receiving emails from it, send an 
>>email to [email protected].
>>To post to this group, send email to [email protected].
>>Visit this group at http://groups.google.com/group/everything-list.
>>For more options, visit https://groups.google.com/d/optout.
>>
>>
>> -- 
>>You received this message because you are subscribed to the Google Groups 
>>"Everything List" group.
>>To unsubscribe from this group and stop receiving emails from it, send an 
>>email to [email protected].
>>To post to this group, send email to [email protected].
>>Visit this group at http://groups.google.com/group/everything-list.
>>For more options, visit https://groups.google.com/d/optout.
>>
>
-- 
>You received this message because you are subscribed to the Google Groups 
>"Everything List" group.
>To unsubscribe from this group and stop receiving emails from it, send an 
>email to [email protected].
>To post to this group, send email to [email protected].
>Visit this group at http://groups.google.com/group/everything-list.
>For more options, visit https://groups.google.com/d/optout.
>
>
>
-- 
>You received this message because you are subscribed to the Google Groups 
>"Everything List" group.
>To unsubscribe from this group and stop receiving emails from it, send an 
>email to [email protected].
>To post to this group, send email to [email protected].
>Visit this group at http://groups.google.com/group/everything-list.
>For more options, visit https://groups.google.com/d/optout.
>
-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to