On Friday, December 12, 2014, meekerdb <[email protected]> wrote: > On 12/11/2014 5:49 PM, Stathis Papaioannou wrote: > >> On 12 December 2014 at 12:22, Jason Resch <[email protected]> wrote: >> >>> >>> On Thu, Dec 11, 2014 at 3:10 PM, LizR <[email protected]> wrote: >>> >>>> On 11 December 2014 at 18:59, Stathis Papaioannou <[email protected]> >>>> wrote: >>>> >>>>> >>>>> On Thursday, December 11, 2014, LizR <[email protected]> wrote: >>>>> >>>>>> Maybe it's a delayed choice experiment and retroactively collapses the >>>>>> wave function, so your choice actually does determine the contents of >>>>>> the >>>>>> boxes. >>>>>> >>>>>> (Just a thought...maybe the second box has a cat in it...) >>>>>> >>>>>> No such trickery is required. Consider the experiment where the >>>>> subject >>>>> is a computer program and the clairvoyant is you, with the program's >>>>> source >>>>> code and inputs. You will always know exactly what the program will do >>>>> by >>>>> running it, including all its deliberations. If it is the sort of >>>>> program >>>>> that decides to choose both boxes it will lose the million dollars. The >>>>> question of whether it *ought to* choose both boxes or one is >>>>> meaningless if >>>>> it is a deterministic program, and the paradox arises from failing to >>>>> understand this. >>>>> >>>> Not trickery, how dare you?! An attempt to give a meaningful answer >>>> which >>>> actually makes something worthwhile from what appears to be a trivial >>>> "paradox" without any real teeth. >>>> >>>> But OK since you are determined to belittle my efforts, let's try your >>>> approach. >>>> >>>> 1 wait 10 seconds >>>> 2 print "after careful consideration, I have decided to open both boxes" >>>> 3 stop >>>> >>>> This is what ANY deterministic computer programme (with no added random >>>> inputs) would boil down to, although millions of lines of code might >>>> take a >>>> while to analyse, and the simplest way to find out the answer in >>>> practice >>>> might be to run it (but each run would give the same result, so once >>>> it's >>>> been run once we can replace it with my simpler version). >>>> >>>> I have to admit I can't see where the paradox is, or why there is any >>>> interest in discussing it. >>>> >>>> It's probably not a true paradox, but why it seems like one is that >>> depending on which version of decision theory you use, you can be led to >>> two >>> opposite conclusions. About half of people think one-boxing is best, and >>> the >>> other half think two-boxing is best, and more often then not, people from >>> either side think people on the other side are idiots. However, for >>> whatever >>> reason, everyone on this list seems to agree one-boxing is best, so you >>> are >>> missing out on the interesting discussions that can arise from seeing >>> people >>> justify their alternate decision. >>> >>> Often two-boxers will say: the predictor's already made his decision, >>> what >>> you decide now can't change the past or alter what's already been done. >>> So >>> you're just leaving money on the table by not taking both boxes. An >>> interesting twist one two-boxer told me was: what would you do if both >>> boxes >>> were transparent, and how does that additional information change what >>> the >>> best choice is? >>> >> If both boxes were transparent, that would screw up the oracle's >> ability to make the prediction, since there would be a feedback from >> the oracle's attempt at prediction to the subject. The oracle can >> predict if I'm going to pick head or tails, but the oracle *can't* >> predict if I'm going to pick heads or tails if he tells me his >> prediction then waits for me to make a decision. >> > Why not? If the oracle has a complete and accurate simulation of you then > he can predict your response to what he tells you - it's just that what he > told you may then no longer be truthful. Suppose he tells you he predicts > you'll pick tails, but he actually predicts that after hearing this you'll > pick heads. It just means he lied. It's somewhat like the unexpected > hanging problem, if you reason from the premises given and you reach a > contradiction then the contradiction was implicit in the premises and no > valid conclusion follows from them. >
If the oracle lies then he controls the inputs and can still make a prediction. The difficulty arises when he tells the truth, which is effectively what happens with transparent boxes. I think then the answer for the oracle is then indeterminate since you can always do the opposite of what he tells you; although a super-oracle outside the system might still be able to predict what will actually happen. -- Stathis Papaioannou -- You received this message because you are subscribed to the Google Groups "Everything List" group. To unsubscribe from this group and stop receiving emails from it, send an email to [email protected]. To post to this group, send email to [email protected]. Visit this group at http://groups.google.com/group/everything-list. For more options, visit https://groups.google.com/d/optout.

