2017-05-08 9:14 GMT+02:00 Bruce Kellett <[email protected]>:

> On 8/05/2017 5:01 pm, Quentin Anciaux wrote:
>
> 2017-05-08 8:58 GMT+02:00 Bruce Kellett <[email protected]>:
>
>> On 8/05/2017 4:41 pm, Quentin Anciaux wrote:
>>
>> 2017-05-08 8:26 GMT+02:00 Bruce Kellett < <[email protected]>
>> [email protected]>:
>>
>>> On 8/05/2017 3:59 pm, David Nyman wrote:
>>>
>>> On 8 May 2017 4:53 a.m., "Bruce Kellett" < <[email protected]>
>>> [email protected]> wrote:
>>>
>>> On 8/05/2017 3:14 am, David Nyman wrote:
>>>
>>>
>>> I've been thinking a bit more about this and I'd like to set out some
>>> further tentative remarks about the above. Your professional expertise in
>>> these matters is orders of magnitude greater than mine and consequently any
>>> comments you might make would be very helpful. By the way, it would also be
>>> helpful if you would read beyond the next paragraph before commenting
>>> because I hope I will come by myself to the fly in the ointment.
>>>
>>> Firstly, and "assuming computationalism" on the basis of CT + YD, we are
>>> led to the view that UD* must include all possible "physical" computational
>>> continuations (actually infinitely reiterated). This of course is also to
>>> assume that all such continuations are finitely computable (i.e. halting).
>>> Now, again on the same assumptions, it might seem reasonable that our
>>> observing such a physics in concrete substantial form is evidence of its
>>> emergence (i.e. epistemologically) as the predominant computational
>>> mechanism underlying those very perceptions. Hence it might seem equally
>>> reasonable to conclude that this is the reason that these latter
>>> correspondingly appear to supervene on concrete physical manifestations in
>>> their effective environment.
>>>
>>> Now wait a minute. We cannot escape the question of measure. Why would
>>> it be reasonable to assume that a physics of this sort should predominate
>>> in the manner outlined above? Well, firstly, it would seem that the
>>> generator of the set of possible physical computations is infinitely
>>> reiterative​ and hence very robust (both in the sense of computational
>>> inclusiveness a la step 7, and that of internal self-consistency). But who
>>> is to say that the generators of "magical" or simply inconsistent
>>> continuations aren't equally or even more prevalent? After all we're
>>> dealing with a Library of Babel here and the Vast majority of any such
>>> library is bound to be gibberish. Well, I'm wondering​ about an analogy
>>> with Feynman's path integral idea (comments particularly appreciated here).
>>> Might a kind of least action principle be applicable here, such that
>>> internally consistent computations self-reinforce, whereas inconsistent
>>> ones in effect self-cancel?
>>>
>>> Also, absence of evidence isn't evidence of absence. I'm thinking here
>>> about the evaluation of what we typically remember having experienced. I
>>> can't help invoking Hoyle here again (sorry). Subjectively speaking,
>>> there's a kind of struggle always in process between remembering and
>>> forgetting. So on the basis suggested above, and from the abstract point of
>>> view of Hoyle's singular agent (or equally Bruno's virgin machine),
>>> inconsistent paths might plausibly tend to result, in effect, in a net
>>> (unintelligible) forgetting and contrariwise, self-consistent paths might
>>> equally plausibly result in a net (intelligible) remembering. I'm speaking
>>> of consistent and hence intelligible "personal histories" here. But perhaps
>>> you would substitute "implausibly" above. Anyway, your comments as ever
>>> particularly appreciated.
>>>
>>>
>>> I think the problem here is the use of the word "consistent". You refer
>>> to "internally consistent computations" and "consistent and hence
>>> intelligible 'personal histories'." But what is the measure of such
>>> consistency? You cannot use the idea of 'consistent according to some
>>> physical laws', because it is those laws that you are supposedly deriving
>>> -- they cannot form part of the derivation. I don't think any notion of
>>> logical consistency can fill the bill here. It is logically consistent that
>>> my present conscious moment, with its rich record of memories of a physical
>>> world, stretching back to childhood, is all an illusion of the momentary
>>> point in a computational history: the continuation of this computation back
>>> into the past, and forward into the future, could be just white noise! That
>>> is not logically inconsistent, or comutationally inconsistent. It is
>>> inconsistent only with the physical laws of conservation and persistence.
>>> But at this point, you do not have such laws!
>>>
>>> In fact, just as Boltzmann realized in the Boltzmann brain problem,
>>> states of complete randomness both before and after our current conscious
>>> moment are overwhelmingly more likley than that our present moment is
>>> immersed in a physics that involves exceptionless conservation laws, so
>>> that the past and future can both be evolved from our present state by the
>>> application of persistent and pervasive physical laws.
>>>
>>> Unless you can give some meaning to the concept of "consistent" that
>>> does not just beg the question, then I think Boltzmann's problem will
>>> destroy your search for some 'measure' that makes our experience of
>>> physical laws (any physical laws, not just those we actually observe)
>>> overwhelmingly likely.
>>>
>>>
>>> Thanks for this. However I'm not sure you've fully addressed my "path
>>> integral" point, for what it's worth. Feynman's idea, if I've got the gist
>>> of it, was that an electron could be considered as taking every possible
>>> path from A to B, but that the direct or short paths could be considered as
>>> mutually reinforcing and the indirect or longer paths as mutually
>>> cancelling.
>>>
>>>
>>> Feynman's ideas relies on a physical theory within which one can
>>> calculate the phase change along each possible path. The upshot is that
>>> paths far away from the path of least action have phases that cancel in the
>>> quantum superposition sense. Note that the crucial input into this picture
>>> is that there is an underlying physical theory, in terms of which one can
>>> calculate the phase changes along each path. Also, it is important to
>>> remember that Feynman's path integral is only one means of calculating
>>> probability amplitudes in QM -- there are many other means of calculating
>>> these, and all give the same results.
>>>
>>> Hence the derivation of the principle of least action. So the analogy,
>>> more or less, that I have in mind is that Boltzmann-type random subjective
>>> states would, computationally speaking, mutually reinforce identical states
>>> supervening on the generator of "consistent" physical continuations (bear
>>> with me for a moment on the applicable sense of "consistent" here). IOW "If
>>> I am a machine I cannot know which machine I am". So as long as the
>>> generator of those consistent states is encapsulated by UD* - which is
>>> equivalent to saying as long as the computable evolution of physical states
>>> is so encapsulated (which it is by assumption) - then we can plausibly
>>> suppose that the net subjective consequences would be indistinguishable.
>>>
>>>
>>> I agree that conscious states, whether Boltzmann brains or parts of a
>>> longer calculation that does not start and end in white noise, are, insofar
>>> as thise states are conscious, they are indistinguishable. But the
>>> Boltzmann brain-type states cannot reinforce the path that leads to
>>> coherent physics, because the continuations are disjoint.
>>>
>>> As to your most reasonable request for a non question begging notion of
>>> consistent in this context, my tentative answer rests on my remarks about
>>> the "struggle between remembering and forgetting". Here's where I use
>>> Hoyle's pigeon hole analogy, which is pretty much equivalent to Barbour's
>>> time capsule one (as he acknowledges in TEOT)
>>>
>>>
>>> Both Hoyle's pigeon holes and Barbour's time capsules assume that there
>>> is a coherent underlying physics with regular exceptionless laws. Until you
>>> have something like that, you cannot define consistent continuations.
>>>
>>>
>> I think that white noise -> consciousness -> white noise must be of an
>> infinitsimal measure for programs generating consciousness moments, because
>> it needs a lot of informations to go from consciousness moment to white
>> noise, more than from consciousness->consciousness-next moment, hence
>> there must be shorter program that goes from one instant to the next than
>> from one thing to totally something else...
>>
>>
>> What has program length got to do with the UD? All programs, of any
>> length at all are present, and there are infinitely more programs that will
>> go from some chance occurrence of a conscious moment to something that is
>> not conscious, be it white noise or white rabbits, than there are programs
>> that produce consistent physical laws generating the sequence of conscious
>> moments. I think Russell's "Occam's catastrophe" is relevant here. The
>> information content of white noise is much greater than the content of
>> physical laws -- program length is not an issue.
>>
>
> Something lie the speed prior... yes the UD has all of them, but the
> measure function (which we don't have) must render the consistency, thing
> like complexity and size could be a way to explain why consciousness->white
> noise have low measure.
>
>
> Those are just arbitrary assumptions, designed to give you some handle on
> what you want. For consistency, the definition of 'consistent
> continuations' for the measure must come from logic and/or arithmetic alone.
>
>
> as for the anthropic principle, you can use  the same as in MWI, you are
> not where you're dead, in fact you are only found in "you" moment, hence
> consciousness->white noise is not a valid continuation and should be
> discarded for measure counting.
>
>
> How do you know that you will not be white noise in the next instant?
>
> Because white noise by definition is not me, so I have to only consider me
> moment for continuation.
>
>
> You have some insight into the future? How do you know that "you" will
> continue in any form? If you use this criterion for consistency, then you
> have built physics into your derivation of physics.
>

If computationalism is true, every me moment exists, so of course, there
are next me moment with computationalism... no need to have future insight,
the thing is to have a measure function on them.

Quentin


>
>
> Bruce
>
>
> Quentin
>
>
>> The anthropic principle assumes regular physical laws, and those are
>> exactly what we are trying to derive.
>>
>> Bruce
>>
> --
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To post to this group, send email to [email protected].
> Visit this group at https://groups.google.com/group/everything-list.
> For more options, visit https://groups.google.com/d/optout.
>



-- 
All those moments will be lost in time, like tears in rain. (Roy
Batty/Rutger Hauer)

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to