A nice exposition, Jesse. But it bothers me that it seems to rely on the idea of "output"
and a kind of isolation like invoking a meta-level. What if instead of "Craig Weinberg
will never in his lifetime assert that this statement is true" we considered "Craig
Weinberg will never in his lifetime think that this statement is true"? Then it seems
that one invokes a kind of paraconsistent logic in which one just refuses to draw any
inferences from this sentence that one cannot think either true or false.
Brent
On 12/19/2013 8:08 AM, Jesse Mazer wrote:
The argument only works if you assume from the beginning that an A.I. is unconscious or
doesn't have the same sort of "mind" as a human (and given your views you probably do
presuppose these things--but if the conclusion *requires* such presuppositions, then
it's an exercise in circular reasoning). If you are instead willing to consider that an
A.I. mind works basically like a human mind (including things like being able to make
mistakes, and being able to understand things it doesn't "say out loud"), and are
willing to "put yourself in the place" of an A.I. being faced with its own Godel
statement, then you can see it's like a more formal equivalent of me asking you to
evaluate the statement "Craig Weinberg will never in his lifetime assert that this
statement is true". You can understand that if you *did* assert that it's true, that
would of course make it false, but you can likewise understand that as long as you try
to refrain from uttering any false statements including that one, it *will* end up being
true.
Similarly, an A.I. who is capable of making erroneous statements, and of understanding
things distinct from its "output" to the world outside the program, might well
understand that its own Godel statement is true--provided it never outputs a formal
judgment that the statement is true, which would mean it's false! So if the A.I. in fact
avoided ever giving as output a judgment about that the statement is true, it need not
be because it lacks an understanding of what's going on, but rather just because it's
caught in a bind similar to the one you're caught in with "Craig Weinberg will never in
his lifetime assert that this statement is true".
To flesh this out a bit, imagine a community of human-like A.I. mathematicians (mind
uploads, say), living in a self-contained simulated world with no input from the
outside, who have the ability to reflect on various arithmetical propositions. Once
there is a consensus in this community that a proposition has been proven true or false,
they can go to a special terminal (call it the "output terminal") and enter it on the
list of proven statements, which will constitute the simulation's "output" to those of
us watching it run in the real world. Suppose also that the simulated world is
constantly growing, and that they have an internal simulated supercomputer within their
world to help with their mathematical investigations, and this supercomputer is
constantly growing in memory too. So if we imagine a string encoding the *initial* state
of the simulation along with the rules determining its evolution, although this string
may be very large, after some time has passed the memory of the simulated supercomputer
will be much larger than that, so it's feasible to have this string appear within the
supercomputer's memory (and it's part of the rules of the simulation that the string
automatically appears in the supercomputer's memory after some finite time T within the
simulation, and all the A.I. mathematicians knew that this was scheduled to happen).
Once the A.I. mathematicians have the program's initial conditions and the rules
governing subsequent evolution, they can construct their own Godel statement. Of course
they can never really be sure that the string they are given correctly describes the
true initial conditions of their own simulated universe, but let's say they have a high
degree of trust that it is--for example, they might be mind uploads of the humans who
designed the original simulation, and they remember having designed it to ensure that
the string that would appear in the supercomputer's memory is the correct one. They
could even use the growing supercomputer to run a simulation-within-the-simulation of
their own history, starting from those initial conditions--the sub-simulation would
always lag behind what they were experiencing, but they could continually verify that
the events in the sub-simulation matched their historical records and memories up to
some point in the past.
So, they have a high degree of confidence that the Godel statement they've constructed
actually is the correct one for their own simulated universe. They can therefore
interpret the conceptual meaning of the statement as something like "you guys living in
the simulation will never enter into your output terminal a judgment that this statement
is true". So they could understand perfectly well that if they ever *did* enter such a
judgment into their output terminal, that would mean the statement was a false statement
about arithmetic. But provided that they *don't* ever enter any such judgment into their
output terminal, they can see it's a true statement about arithmetic (and can discuss
this fact among themselves and reach a consensus about this fact, as long as they don't
enter it as output to the terminal). If they are mathematical platonists, they realize
that this feeling of it being their choice whether to output the statement or not, with
the statement's truth or falsity depending on that choice, is a sort of illusion--really
the truth-value of the statement is a timeless fact about arithmetic. But presumably, in
such a situation they would adopt a "compatibilist" view of free will as many real-world
philosophers have done (http://plato.stanford.edu/entries/compatibilism/), a view which
sees no conflict between the feeling of free will and the idea that our actions are
ultimately completely determined by natural laws and initial conditions.
Jesse
--
You received this message because you are subscribed to the Google Groups
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email
to everything-list+unsubscr...@googlegroups.com.
To post to this group, send email to everything-list@googlegroups.com.
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/groups/opt_out.