I never saw how the AIXI idealization represented true intelligence.  Of course 
Bayesian learning and a model derived from algorithmic information theory might 
be used as tools of intelligence, but that model has a problem similar to 
logic.  You can't possibly think that a simple deductive system can be used to 
represent intelligence. Deductive logic can be used as a tool of logic but it 
is not in itself equivalent to artificial intelligence.  Similar arguments go 
against ideas like strict behaviorist psychology. So I am explicitly making the 
claim that I think that I could create something that, even though it was a 
simple application, would be a more complete base for AGI then the contemporary 
paradigms. There is no evidence that we make the most rational choice in a 
situation relative to our previous experiences (where the term rational has a 
precise idealized mathematical definition.)  We need to be able to explore 
other pathways because the real world always can present itself with paths that 
do not fall within our expectations.  We need to be able to apply our 
imagination to a problem without being bounded by an approximation to a most 
rational choice.   Like extreme Behaviorism, the AIXI Bayesian ideals do not 
explain independent inner directed learning, they deny it.  Now, the use of 
something like the Monte Carlo Method is an example of an imaginative use of 
the principles (and I like the idea) but it certainly is a constrained use of 
the artificial imagination. I want to create a simplified model which would 
include learning through structured concept based interactions.  Since the 
potential for complexity is so high, my application would have to use a 
simplified base set of these kinds of interactions.  Perhaps, as computer 
science continues to advance any base set which I would select at this time 
will turn out to be inadequate for some future computer.  But, in general, I 
believe that I can go further than contemporary AGI just because the 
foundational principals of AGI are so poorly designed in contemporary models. 
Jim Bromer
 
> Date: Thu, 25 Jul 2013 22:33:59 +0200
> Subject: Re: [agi] A Very Simple AGI Project
> From: [email protected]
> To: [email protected]
> 
> Matt Mahoney wrote
> 
> > 1. AIXI won't solve AGI because AIXI is not computable, and good
> > approximations are intractable beyond toy problems.
> 
> Yes. I offered it as an inspiration, not as a solution. Including
> awareness of the issues you mention. If he wants to go the "simple" way,
> AIXI (with approximations) is something "simple" enough and it comes with
> known problems. Many man-years and CPU cycles were spent on defining AIXI
> and finding the issues, starting from scratch is often wasteful.
> At least he should look into Hutter's first principle approach
> (video/notes below).
> 
> As I said, I might have misunderstood his requirements and he didn't ask
> for that kind of answer.
> 
> Anyway, everyone interested may find below my notes on several AIXI
> videos. Note that they're not beautified, some parts are just my
> additional comments, and they're not consistent in the use of names (e.g.,
> Solomonoff/Solomonov).
> 
> cu Jan
> 
> 
> ======================
> Marcus Hutter: "What is intelligence? AIXI and induction" [18:56]
> http://www.youtube.com/watch?v=F2bQ5TSB-cE
> 
> Real-world intelligence is resource-bounded. But it's hard to define. So
> we take another road:
> 
> phase 1: Define the problem ("intelligence") first, start with the
> unbounded version (non-computable). Once we're sure to have this solved:
> 
> phase 2: try to approximate it and make a computational theory out of it.
> 
> (phase 3: now you can still try to create a theory of resource-bounded
> intelligence if you want.)
> 
> (Like with universal turing machines: unbounded space and time resources).
> 
> AIXI is theoretical computer science with theoretical general
> intelligence. It gives us a model of the capabilities and limitations of
> intelligent agents in correspondence to environments.
> 
> Hutter: "Or the short answer may be I am not smart enough to come up with
> a resource bounded theory of intelligence, therefore I only developed one
> without resource constraints."
> 
> Hutter: "(...) informal definition that intelligence is an agent's ability
> to succeed or achieve goals in a wide range of environments."
> 
> Hutter: "Or Universal AI is the general field theory and AIXI is the
> particular agent which acts optimally in this sense."
> 
> Planning component, learning compontent.
> AIXI agent starts blank (no data/knowledge). Acquire data/knowledge of the
> world and build its own model from those data.
> How to learn a model from data -> Roots: Kolmogorov complexity,
> algorithmic information theory.
> 
> * look for the simplest model that describes your data sufficiently well.
> (learning part)
> * take this knowledge and think about the best possible outcomes of all
> possible actions where "best" is evaluated according to a utility function
> (value function) -> rewards. (prediction part)
> * Maximize the reward of its lifetime. (Planning part)
> 
> AIXI: it's a mathematical theory of intelligence, one can prove properties
> (and one can prove that it's the most intelligent system possible).
> Downside: it's incomputable (needs infinite computational resources).
> There's the need to approximate it. One of those approximations: Pac-Man:
> 
> Pacman via AIXI Approximation [5:42]
> http://www.youtube.com/watch?v=RhQTWidQQ8U
> 
> Playing Pacman using AIXI Approximation [1:52]
> http://www.youtube.com/watch?v=yfsMHtmGDKE
> 
> (it starts blank, then via interacting with its environment gains
> knowledge. A value function is given before to compute positive and
> negative rewards.)
> 
> What's so cool about it is that it's not tailored to any particular
> application (like only playing chess or go): interface it with any
> problem, it could (theoretically) learn to solve this problem optimally.
> There's no built-in pacman knowledge, only the value function. Getting
> feedback it learns everything else by itself.
> 
> in approximations:
> For the learning part: standard compressors / data compressors.
> for the planning part: standard monte carlo random search
> 
> monte carlo algorithms: to search through enormous trees; if one could
> search through those huge trees, one would arrive at an optimal solution
> (but in reality that's computationally infeasable), but MCs are for
> approximations/heuristics (stochastic search).
> here: Upper Confidence Bound for Trees (UCT MC algorithm) -> very balanced
> way of exploration and exploitation: you search where you think things are
> good or where you have very little knowledge and maybe there's a gold
> nugget. Fundamental problem: stay where you believe things are good or
> explore.
> (nice to have: only one parameter to control, where in other algos like
> NNs there are sometimes several thousends).
> 
> Essential part of the AI problem: get induction right -> derive models
> from data.
> use Occam's razor (take the simplest theory consistent with your data),
> which has been formalized and quantified -> Kolmogorov complexity
> (quantification what complexity or simplicity means).
> -> universal theory of induction/prediction: take past data stream, ask
> "what comes next".
> universal predictor that works in any kind of situation. (it's
> incomputable, but beautiful, later you approximate it)
> 
> Bayesian reasoning is built into AIXI.
> 
> sequential decision theory
> 
> =====================
> 
> Tim Tyler: On AIXI
> http://www.youtube.com/watch?v=xDMN4zi7wb4
> 
> 1. problem: Has no representation of self. It's not embedded in its
> environment. But that's not a serious flaw.
> 
> 2. problem: wirehead problem -> hacking its own reward feedback, which
> will endanger its long-term survival.
> 
> 3. problem: world is parallel, AIXI agent is a serial agent modeled by a
> TM. While parallelism can be modeled sequentially, the reward model is
> also serial and thus unsuitable for a parallel world. (not a serious
> problem).
> 
> 4. problem: solomonoff induction is a formalized version of Occam's razor
> using kolmogoroff complexity (not a serious problem) -> it's not
> language-independent and it's not known whether there exists an optimal
> description of Occam's razor.
> 
> Ben Goertzel: AIXI shows that AGI is a problem of resource restrictions,
> if there were no space and time constraints, it'd be a trivial program.
> 
> From the video comments: AIXI has no access to its own reasoning. That's
> true for reinforcement learning that treats the brain as a black box, thus
> it can't explain its own reasoning.
> 
> =====================
> 
> Marcus Hutter - AI, the Scientific Method & Philosophy
> http://www.youtube.com/watch?v=slTuDZIJqkQ
> 
> Science is very much about induction: get data, derive models. ->
> Solomonoff Induction
> It's also about decision making and planning (that's the active part). ->
> AIXI
> 
> You can always ask "why", but to prevent an infinite regress, you have to
> stop somewhere and declare that something are the axioms and ask about
> their consequences. When they are useful, you can stop questioning (you
> could go on but for practical reason you stop somewhere and proceed with
> what you have).
> When you do this process (ask "why why why why why") often enough, you
> arrive at the Occam's Razor principle. It seems to be necessary and
> sufficient for science. That's defining science and OR is about the
> scientific method. There might be better principles than OR, but currently
> it's the best we have. Just use it until someone has found something
> better.
> 
> Issues on free will.
> Closed system: can be predicted from outside.
> Open system (here: give feedback into it): put yourself into the closed
> system and everything's fine again.
> 
> =====================
> 
> A computational approximation to the AIXI model (AGI 2008)
> http://www.youtube.com/watch?v=SpgXXfRqNAk
> 
> AIXI: control theory (expectation maximazation) + universal induction
> (Solomonoff induction) -> optimal behavior.
> 
> Problem: Find a computationally efficient (if not optimal) approximation
> for the optimal but incomputable AIXI theory.
> 
> Universal induction solves problem of choosing a prior to achieve optimal
> inductive inference.
> 
> =====================
> 
> Marcus Hutter: Foundations of Intelligent Agents
> http://www.youtube.com/watch?v=x8btbKaRfoc
> 
> Informal working definition: Intelligence measures an agent's ability to
> perform well in a wide range of environments.
> 
> Design from first principles of Artificial Intelligent Systems:
> 
> * Logic/language based: expert/reasoning/proving/cognitive systems
> * Economics inspired: utility, sequential decisions, game theory
> * Cybernetics: adaptive dynamic control
> * Machine Learning: reinforcement learning
> * Information processing: data compression -> intelligence
> Separately too limited for AGI, but jointly very powerful.
> 
> Foundations of Universal Artificial Intelligence:
> 
> * Philosophy: Ockham, Epicurus, Induction
> * Mathematics: Information, complexity, Bayesian & Algorithmic
> Probability, Solomonoff Induction, Sequential Decision
> * Frameworks: Rational Agents (in known and unknown environments)
> * Computation: Universal search and feature Reinforcement Learning
> 
> Science is about induction (Ockham's Razor): take the simplest hypothesis
> consistent with the data
> Induction: go from one to the next
> 1. construct set of possible nexts
> 2. choose one next
> Is the most important principle in science and ML
> 
> Problem: Quantification of simplicity/complexity (because a machine has to
> apply Ockham's Razor)
> -> Due to the Turing's Thesis, everything computable by a human using a
> fixed procedure can also be computed by a (universal) Turing Machine
> -> Measure of complexity: Kolmogorov Complexity, Algorithmic Information
> Theory => Kolmogorov complexity of a string is the length of the shortest
> program on U describing the string.
> 
> K(s) := min_p {Length(p): U(p) = s}. // U(p) is a program computing s,
> pick the shortest one.
> 
> -> Bayesian Probability Theory: update prior degree of belief in
> hypothesis H, given new observations D, to posterior belief in H.
> 
> Pr(H|D) \propto Pr(D|H)Pr(H).
> 
> Alg. Inf. Theo: how to initialize beliefs
> Bayes: how to update beliefs
> 
> -> Algorithmic Probability
> Epicurus: if more than one theory=hypothesis=model is consistent with the
> observations, keep them all.
> Refinement with Ockham: Give simpler theories higher a-priori weight.
> Quantitative: Pr(H) := 2^{-K(H)}
> i.e., keep them, but weight them
> 
> => Universal Induction (by Solomonoff):
> combined Ockham, Epicurus, Bayes, Turing into one formal theory of
> sequential prediction.
> 
> Universal a-priori probability: M(x) := prob that U fed with noise outputs
> x, i.e., what is the prob that randomness produces x.
> M(x_{t+1} | x_1, ..., x_t) best predicts x_{t+1} from x_1, ..., x_t.
> 
> => Sequential Decision Theory (Optimal Control Theory)
> 
> for t = 1, 2, ..., given sequence x_1, x_2, ..., x_{t-1}:
> 1) make decision y_t
> 2) observe x_t
> 3) suffer Loss(x_t, y_t)
> 4) t -> t+1, goto 1)
> Goal: minimize expected Loss
> Problem: true prob unknown
> Solution: use Solomonoff's M(x)
> 
> => Agent Model (extremely general):
> agent interacts with environments in cycles t, t+1, ... and receives
> pos/neg reinforcement feedback
> 
> AIXI = AI + greek letter Xi
> * universally optimal rational agent
> * ultimate Super Intelligence
> * computationally intractable
> * could serve as a gold standard for AGI
> 
> => Towards practical universal AI (efficient general-purpose intelligent
> agents)
> Additional ingredients:
> * universal search (Schmidhuber)
> * learning: mostly Reinforcement Learning
> * information: Minimal Description Length Principle
> * complexity/similarity
> * optimization, esp. Monte Carlo
> 
> Feature Reinforcement Learning
> reduce real-world problem into a (tractable) Markov Decision Process by
> learning relevant features.
> 
> =====================
> 
> 
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/24379807-f5817f28
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com

                                          


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to