On 12 December 2014 at 12:22, Jason Resch <[email protected]> wrote:
>
>
> On Thu, Dec 11, 2014 at 3:10 PM, LizR <[email protected]> wrote:
>>
>> On 11 December 2014 at 18:59, Stathis Papaioannou <[email protected]>
>> wrote:
>>>
>>>
>>> On Thursday, December 11, 2014, LizR <[email protected]> wrote:
>>>>
>>>> Maybe it's a delayed choice experiment and retroactively collapses the
>>>> wave function, so your choice actually does determine the contents of the
>>>> boxes.
>>>>
>>>> (Just a thought...maybe the second box has a cat in it...)
>>>>
>>> No such trickery is required. Consider the experiment where the subject
>>> is a computer program and the clairvoyant is you, with the program's source
>>> code and inputs. You will always know exactly what the program will do by
>>> running it, including all its deliberations. If it is the sort of program
>>> that decides to choose both boxes it will lose the million dollars. The
>>> question of whether it *ought to* choose both boxes or one is meaningless if
>>> it is a deterministic program, and the paradox arises from failing to
>>> understand this.
>>
>> Not trickery, how dare you?! An attempt to give a meaningful answer which
>> actually makes something worthwhile from what appears to be a trivial
>> "paradox" without any real teeth.
>>
>> But OK since you are determined to belittle my efforts, let's try your
>> approach.
>>
>> 1 wait 10 seconds
>> 2 print "after careful consideration, I have decided to open both boxes"
>> 3 stop
>>
>> This is what ANY deterministic computer programme (with no added random
>> inputs) would boil down to, although millions of lines of code might take a
>> while to analyse, and the simplest way to find out the answer in practice
>> might be to run it (but each run would give the same result, so once it's
>> been run once we can replace it with my simpler version).
>>
>> I have to admit I can't see where the paradox is, or why there is any
>> interest in discussing it.
>>
>
> It's probably not a true paradox, but why it seems like one is that
> depending on which version of decision theory you use, you can be led to two
> opposite conclusions. About half of people think one-boxing is best, and the
> other half think two-boxing is best, and more often then not, people from
> either side think people on the other side are idiots. However, for whatever
> reason, everyone on this list seems to agree one-boxing is best, so you are
> missing out on the interesting discussions that can arise from seeing people
> justify their alternate decision.
>
> Often two-boxers will say: the predictor's already made his decision, what
> you decide now can't change the past or alter what's already been done. So
> you're just leaving money on the table by not taking both boxes. An
> interesting twist one two-boxer told me was: what would you do if both boxes
> were transparent, and how does that additional information change what the
> best choice is?

If both boxes were transparent, that would screw up the oracle's
ability to make the prediction, since there would be a feedback from
the oracle's attempt at prediction to the subject. The oracle can
predict if I'm going to pick head or tails, but the oracle *can't*
predict if I'm going to pick heads or tails if he tells me his
prediction then waits for me to make a decision.


-- 
Stathis Papaioannou

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to