On Thu, Dec 11, 2014 at 7:49 PM, Stathis Papaioannou <[email protected]>
wrote:
>
> On 12 December 2014 at 12:22, Jason Resch <[email protected]> wrote:
> >
> >
> > On Thu, Dec 11, 2014 at 3:10 PM, LizR <[email protected]> wrote:
> >>
> >> On 11 December 2014 at 18:59, Stathis Papaioannou <[email protected]>
> >> wrote:
> >>>
> >>>
> >>> On Thursday, December 11, 2014, LizR <[email protected]> wrote:
> >>>>
> >>>> Maybe it's a delayed choice experiment and retroactively collapses the
> >>>> wave function, so your choice actually does determine the contents of
> the
> >>>> boxes.
> >>>>
> >>>> (Just a thought...maybe the second box has a cat in it...)
> >>>>
> >>> No such trickery is required. Consider the experiment where the subject
> >>> is a computer program and the clairvoyant is you, with the program's
> source
> >>> code and inputs. You will always know exactly what the program will do
> by
> >>> running it, including all its deliberations. If it is the sort of
> program
> >>> that decides to choose both boxes it will lose the million dollars. The
> >>> question of whether it *ought to* choose both boxes or one is
> meaningless if
> >>> it is a deterministic program, and the paradox arises from failing to
> >>> understand this.
> >>
> >> Not trickery, how dare you?! An attempt to give a meaningful answer
> which
> >> actually makes something worthwhile from what appears to be a trivial
> >> "paradox" without any real teeth.
> >>
> >> But OK since you are determined to belittle my efforts, let's try your
> >> approach.
> >>
> >> 1 wait 10 seconds
> >> 2 print "after careful consideration, I have decided to open both boxes"
> >> 3 stop
> >>
> >> This is what ANY deterministic computer programme (with no added random
> >> inputs) would boil down to, although millions of lines of code might
> take a
> >> while to analyse, and the simplest way to find out the answer in
> practice
> >> might be to run it (but each run would give the same result, so once
> it's
> >> been run once we can replace it with my simpler version).
> >>
> >> I have to admit I can't see where the paradox is, or why there is any
> >> interest in discussing it.
> >>
> >
> > It's probably not a true paradox, but why it seems like one is that
> > depending on which version of decision theory you use, you can be led to
> two
> > opposite conclusions. About half of people think one-boxing is best, and
> the
> > other half think two-boxing is best, and more often then not, people from
> > either side think people on the other side are idiots. However, for
> whatever
> > reason, everyone on this list seems to agree one-boxing is best, so you
> are
> > missing out on the interesting discussions that can arise from seeing
> people
> > justify their alternate decision.
> >
> > Often two-boxers will say: the predictor's already made his decision,
> what
> > you decide now can't change the past or alter what's already been done.
> So
> > you're just leaving money on the table by not taking both boxes. An
> > interesting twist one two-boxer told me was: what would you do if both
> boxes
> > were transparent, and how does that additional information change what
> the
> > best choice is?
>
> If both boxes were transparent, that would screw up the oracle's
> ability to make the prediction, since there would be a feedback from
> the oracle's attempt at prediction to the subject. The oracle can
> predict if I'm going to pick head or tails, but the oracle *can't*
> predict if I'm going to pick heads or tails if he tells me his
> prediction then waits for me to make a decision.
>

Right that's what I told him. :-)

Jason

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to