Matt,

Ok, you have me, I admit defeat.

I could only continue my argument if I could pin down what sorts of
facts need to be learned with high probability for RSI, and show
somehow that this set does not include unlearnable facts. Learnable
facts form a larger set than provable facts, since for example we can
probabilistically declare that a program never halts if we run it for
a while and it doesn't. But there are certain facts that are not even
probabilistically learnable, so until I can show that none of these
are absolutely essential to RSI, I concede.

--Abram Demski

On Wed, Aug 27, 2008 at 6:48 PM, Matt Mahoney <[EMAIL PROTECTED]> wrote:
> Abram Demski <[EMAIL PROTECTED]> wrote:
>>First, I do not think it is terribly difficult to define a Goedel
>>machine that does not halt. It interacts with its environment, and
>>there is some utility value attached to this interaction, and it
>>attempts to rewrite its code to maximize this utility.
>
> It's not that the machine halts, but that it makes no further improvements 
> once the best solution is found. This might not be a practical concern if the 
> environment is very complex.
>
> However, I doubt that a Goedel machine could even be built. Legg showed [1] 
> that Goedel incompleteness is ubiquitous. To paraphrase, beyond some low 
> level of complexity, you can't prove anything. Perhaps this is the reason we 
> have not (AFAIK) built a software model, even for very simple sets of axioms.
>
> If we resort to probabilistic evidence of improvement rather than proofs, 
> then it is no longer a Goedel machine, and I think we would need experimental 
> verification of RSI. Random modifications of code are much more likely to be 
> harmful than helpful, so we would need to show that improvements could be 
> detected with a very low false positive rate.
>
> 1. http://www.vetta.org/documents/IDSIA-12-06-1.pdf
>
>
>  -- Matt Mahoney, [EMAIL PROTECTED]
>
>
>
> ----- Original Message ----
> From: Abram Demski <[EMAIL PROTECTED]>
> To: [email protected]
> Sent: Wednesday, August 27, 2008 11:40:24 AM
> Subject: Re: Goedel machines (was Re: Information theoretic approaches to AGI 
> (was Re: [agi] The Necessity of Embodiment))
>
> Matt,
>
> Thanks for the reply. There are 3 reasons that I can think of for
> calling Goedel machines "bounded":
>
> 1. As you assert, once a solution is found, it stops.
> 2. It will be on a finite computer, so it will eventually reach the
> one best version of itself that it can reach.
> 3. It can only make provably correct steps, which is very limiting
> thanks to Godel's incompleteness theorem.
>
> I'll try to argue that each of these limits can be overcome in
> principle, and we'll see if the result satisfies your RSI criteria.
>
> First, I do not think it is terribly difficult to define a Goedel
> machine that does not halt. It interacts with its environment, and
> there is some utility value attached to this interaction, and it
> attempts to rewrite its code to maximize this utility.
>
> The second and third need to be tackled together, because the main
> reason that a Goedel machine can't improve its own hardware is because
> there is uncertainty involved, so it would never be provably better.
> There is always some chance of hardware malfunction. So, I think it is
> necessary to grant the possibility of modifications that are merely
> very probably correct. Once this is done, 2 and 3 fall fairly easily,
> assuming that the machine begins life with a good probabilistic
> learning system. That is a big assumption, but we can grant it for the
> moment I think?
>
> For the sake of concreteness, let's say that the utility value is some
> (probably very complex) attempt to logically describe Eliezer-style
> Friendliness, and that the probabilistic learning system is an
> approximation of AIXI (which the system will improve over time along
> with everything else). (These two choices don't reflect my personal
> tastes, they are just examples.)
>
> By tweaking the allowances the system makes, we might either have a
> slow self-improver that is, say, 99.999% probable to only improve
> itself in the next 100 years, or a faster self-improver that is 50%
> guaranteed.
>
> Does this satisfy your criteria?
>
> On Wed, Aug 27, 2008 at 9:14 AM, Matt Mahoney <[EMAIL PROTECTED]> wrote:
>> Abram Demski <[EMAIL PROTECTED]> wrote:
>>>Matt,
>>>
>>>What is your opinion on Goedel machines?
>>>
>>> http://www.idsia.ch/~juergen/goedelmachine.html
>>
>>
>> Thanks for the link. If I understand correctly, this is a form of bounded 
>> RSI, so it could not lead to a singularity. A Goedel machine is functionally 
>> equivalent to AIXI^tl in that it finds the optimal reinforcement learning 
>> solution given a fixed environment and utility function. The difference is 
>> that AIXI^tl does a brute force search of all machines up to length l for 
>> time t each, so it run in O(t 2^l) time. A Goedel machine achieves the same 
>> result more efficiently through a series of self improvments by proving that 
>> each proposed modification (including modifications to its own proof search 
>> code) is a actual improvement. It does this by using an instruction set such 
>> that it is impossible to construct incorrect proof verification code.
>>
>> What I am looking for is unbounded RSI capable of increasing intelligence. A 
>> Goedel machine doesn't do this because once it finds a solution, it stops. 
>> This is the same problem as a chess playing program that plays randomly 
>> modified copies of itself in death matches. At some point, it completely 
>> solves the chess problem and stops improving.
>>
>> Ideally we should use a scalable test for intelligence such as Legg and 
>> Hutter's universal intelligence, which measures expected accumulated reward 
>> over a Solomonoff distribution of environments (random programs). We can't 
>> compute this exactly because it requires testing an infinite number of 
>> environments, but we can approximate it to arbitrary precision by randomly 
>> sampling environments.
>>
>> RSI would require a series of increasingly complex test environments because 
>> otherwise there is an exact solution such that RSI would stop once found. 
>> For any environment with Kolmogorov complexity l, and agent can guess all 
>> environments up to length l. But this means that RSI cannot be implemented 
>> by a Turing machine because a parent with complexity l cannot test its 
>> children because it cannot create environments with complexity greater than 
>> l.
>>
>> RSI would be possible with a true source of randomness. A parent could 
>> create arbitrarily complex environments by flipping a coin. In practice, we 
>> usually ignore the difference between pseudo-random sources and true random 
>> sources. But in the context of Turing machines that can execute exponential 
>> complexity algorithms efficiently, we can't do this because the child could 
>> easily guess the parent's generator, which has low complexity.
>>
>> One could argue that the real universe does have true random sources, such 
>> as quantum mechanics. I am not convinced. The universe does have a definite 
>> quantum state, but it is not possible to know it because a memory within the 
>> universe cannot have more information than the universe. Therefore, any 
>> theory of physics must appear random.
>>  -- Matt Mahoney, [EMAIL PROTECTED]
>>
>>
>>
>> ----- Original Message ----
>> From: Abram Demski <[EMAIL PROTECTED]>
>> To: [email protected]
>> Sent: Monday, August 25, 2008 3:30:59 PM
>> Subject: Re: Information theoretic approaches to AGI (was Re: [agi] The 
>> Necessity of Embodiment)
>>
>> Matt,
>>
>> What is your opinion on Goedel machines?
>>
>> http://www.idsia.ch/~juergen/goedelmachine.html
>>
>> --Abram
>>
>> On Sun, Aug 24, 2008 at 5:46 PM, Matt Mahoney <[EMAIL PROTECTED]> wrote:
>>> Eric Burton <[EMAIL PROTECTED]> wrote:
>>>
>>>
>>>>>These have profound impacts on AGI design. First, AIXI is (provably) not 
>>>>>computable,
>>>>>which means there is no easy shortcut to AGI. Second, universal 
>>>>>intelligence is not
>>>>>computable because it requires testing in an infinite number of 
>>>>>environments. Since
>>>>>there is no other well accepted test of intelligence above human level, it 
>>>>>casts doubt on
>>>>>the main premise of the singularity: that if humans can create agents with 
>>>>>greater than
>>>>>human intelligence, then so can they.
>>>>
>>>>I don't know for sure that these statements logically follow from one
>>>>another.
>>>
>>> They don't. I cannot prove that there is no non-evolutionary model of 
>>> recursive self improvement (RSI). Nor can I prove that there is. But it is 
>>> a question we need to answer before an evolutionary model becomes 
>>> technically feasible, because an evolutionary model is definitely 
>>> unfriendly.
>>>
>>>>Higher intelligence bootstrapping itself has already been proven on
>>>>Earth. Presumably it can happen in a simulation space as well, right?
>>>
>>> If you mean the evolution of humans, that is not an example of RSI. One 
>>> requirement of friendly AI is that an AI cannot alter its human-designed 
>>> goals. (Another is that we get the goals right, which is unsolved). 
>>> However, in an evolutionary environment, the parents do not get to choose 
>>> the goals of their children. Evolution chooses goals that maximize 
>>> reproductive fitness, regardless of what you want.
>>>
>>> I have challenged this list as well as the singularity and SL4 lists to 
>>> come up with an example of a mathematical, software, biological, or 
>>> physical example of RSI, or at least a plausible argument that one could be 
>>> created, and nobody has. To qualify, an agent has to modify itself or 
>>> create a more intelligent copy of itself according to an intelligence test 
>>> chosen by the original. The following are not examples of RSI:
>>>
>>> 1. Evolution of life, including humans.
>>> 2. Emergence of language, culture, writing, communication technology, and 
>>> computers.
>>> 3. A chess playing (or tic-tac-toe, or factoring, or SAT solving) program 
>>> that makes modified copies of itself by
>>> randomly flipping bits in a compressed representation of its source
>>> code, and playing its copies in death matches.
>>> 4. Selective breeding of children for those that get higher grades in 
>>> school.
>>> 5. Genetic engineering of humans for larger brains.
>>>
>>> 1 fails because evolution is smarter than all of human civilization if you 
>>> measure intelligence in bits of memory. A model of evolution uses 10^37 
>>> bits (10^10 bits of DNA per cell x 10^14 cells in the human body x 10^10 
>>> humans x 10^3 ratio of biomass to human mass). Human civilization has at 
>>> most 10^25 bits (10^15 synapses in the human brain x 10^10 humans).
>>>
>>> 2 fails because individual humans are not getting smarter with each 
>>> generation, at least not nearly as fast as civilization is advancing. 
>>> Rather, there are more humans, and we are getting better organized through 
>>> specialization of tasks. Human brains are not much different than they were 
>>> 10,000 years ago.
>>>
>>> 3 fails because there are no known classes of problems that are provably 
>>> hard to solve but easy to verify. Tic-tac-toe and chess have bounded 
>>> complexity. It has not been proven that factoring is harder than 
>>> multiplication. We don't know that P != NP, and even if we did, many 
>>> NP-complete problems have special cases that are easy to solve, and we 
>>> don't know how to program the parent to avoid these cases through 
>>> successive generations.
>>>
>>> 4 fails because there is no evidence that above a certain level (about IQ 
>>> 200) that childhood intelligence correlates with adult success. The problem 
>>> is that adults of average intelligence can't agree on how success should be 
>>> measured*.
>>>
>>> 5 fails for the same reason.
>>>
>>> *For example, the average person recognizes Einstein as a genius not 
>>> because they are
>>> awed by his theories of general relativity, but because other people
>>> have said so. If you just read his papers (without understanding their 
>>> great insights) and knew that he never learned to drive a car, you might 
>>> conclude differently.
>>>
>>>  -- Matt Mahoney, [EMAIL PROTECTED]
>>>
>>>
>>>
>>> -------------------------------------------
>>> agi
>>> Archives: https://www.listbox.com/member/archive/303/=now
>>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>>> Modify Your Subscription: https://www.listbox.com/member/?&;
>>> Powered by Listbox: http://www.listbox.com
>>>
>>
>>
>> -------------------------------------------
>> agi
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>>
>>
>>
>> -------------------------------------------
>> agi
>> Archives: https://www.listbox.com/member/archive/303/=now
>> RSS Feed: https://www.listbox.com/member/archive/rss/303/
>> Modify Your Subscription: https://www.listbox.com/member/?&;
>> Powered by Listbox: http://www.listbox.com
>>
>
>
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>
>
>
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription: https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>


-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=111637683-c8fa51
Powered by Listbox: http://www.listbox.com

Reply via email to