Re: Compatibilism
On Tue, Nov 23, 2010 at 4:20 PM, 1Z wrote: > > > On Nov 21, 6:35 pm, Rex Allen wrote: >> On Fri, Nov 19, 2010 at 7:28 AM, 1Z wrote: >>> On Nov 18, 6:31 am, Rex Allen wrote: If there is a reason, then the reason determined the choice. No free will. >>> >>> Unless you determined the reason. >> >> How would you do that? By what means? According to what rule? Using >> what process? >> >> If you determined the reason, what determined you? Why are you in the >> particular state you're in? >> >> If there exists some rule that translates your specific state into >> some particular choice, then there's still no free will. The rule >> determined the choice. > > And if there isn't...you have an action that is reasoned yet > undetermined, as required If there is no rule that translates your specific state into some particular choice, then what is it connects the state to the choice? The state occurs. Then the choice occurs. But nothing connects them? That is accidentalism isn't it? >>> I.1.v Libertarianism — A Prima Facie case for free will >> >> As for the rest of it, I read it, but didn't find it convincing on any level. >> >> RIG + SIS <> Free Will >> >> A random process coupled to a deterministic process isn't free will. >> It's just a random process coupled to a deterministic process. > > If you insist that FW is a Tertium Datur that is fundamenally > different from both determinism and causation, then you > won't accept a mixture. However, I don;t think Tertium Datur > is a good definition of DW sinc e it is too question begging It seems to me that when people discuss free will, they are always really interested in "ultimate responsibility" for actions. Any defense of "free will" must allow for ultimate responsibility for actions. I say that ultimate responsibility is impossible, because neither caused actions nor random actions nor any combination of cause and randomness seems to result in "ultimate responsibility". Ultimate responsibility means that reward and punishment are justified for acts *even after* setting aside any utilitarian considerations. So *if* it were possible to be ultimately responsible for a bad act, we wouldn't need to justify the offender's punishment in terms of deterring future bad behavior by the offender or others. We wouldn't need to justify the offender's punishment in terms of rehabilitating the offender so that they don't commit similar bad acts in the future. We wouldn't need to justify the offender's punishment in terms of motivating better behavior by them or others in the future. We wouldn't need to justify the offender's punishment in terms of compensating their victims or insuring social stability. Instead, we could justify the offender's punishment purely in terms of their ultimate responsibility for it. Using their free will, they chose to commit the bad act, and therefore they deserve the punishment. End of story. So, given that the punishment would no longer need to be justified in terms of anything other than ultimate responsibility, how would one justify limits on the punishment's severity? -- You received this message because you are subscribed to the Google Groups "Everything List" group. To post to this group, send email to everything-l...@googlegroups.com. To unsubscribe from this group, send email to everything-list+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/everything-list?hl=en.
Re: Compatibilism
On Tue, Nov 23, 2010 at 4:12 PM, 1Z wrote: > On Nov 21, 6:43 pm, Rex Allen wrote: >> On Fri, Nov 19, 2010 at 7:36 AM, 1Z wrote: >>> >>> No-one is. They are just valid descriptions. There is no argument >>> to the effect that logic is causal or it is nothing. It is not >>> the case that causal explanation is the only form of explanagion >> >> “Valid descriptions” don’t account for why things are this way rather >> than some other way. >> > > If a higher level description is a valid description of > some microphysics, then it will be an explanation of > why the result happened given the initial conditions > > It won't solve the trilemma, but neither will > microphysical causality So Agrippa's Trilemma revolves around the question of how we can justify our beliefs. It seems to me that an entirely acceptable solution is just to accept that we can't justify our beliefs. >> As I said before, materialism could conceivably explain human ability >> and behavior, but in my opinion runs aground at human consciousness. >> Therefore, I doubt that humans are a complex sort of robot. > > Is human consciousness causally effective? I don't believe so, no. And claiming that consciousness is itself caused just runs into infinite regress, as you then need to explain what causes the cause of conscious experience, and so on. Therefore, taking the same approach as with Agrippa's Trilemma, it seems best to just accept that there is no cause for conscious experience either. Is it a useful answer? Maybe not. But where does it say that all answers have to be useful? Besides, what causes you to care about usefulness? Evolution. What causes evolution? Initial conditions and causal laws. What causes initial conditions and causal laws? And so on. We've been through this before I think. -- You received this message because you are subscribed to the Google Groups "Everything List" group. To post to this group, send email to everything-l...@googlegroups.com. To unsubscribe from this group, send email to everything-list+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/everything-list?hl=en.
Re: advice needed for Star Trek talk
On Wed, Nov 24, 2010 at 1:50 PM, ronaldheld wrote: > Jason: > I see what you are saying up at our level of understanding, I do not > know how to present that in a technically convincing matter. > Ronald > > Which message in particular do you think is difficult to present convincingly? Tegmark's ideas that everything is real, or the suggestion that computer simulation might be a legitimate tool for exploration? Jason -- You received this message because you are subscribed to the Google Groups "Everything List" group. To post to this group, send email to everything-l...@googlegroups.com. To unsubscribe from this group, send email to everything-list+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/everything-list?hl=en.
Re: Compatibilism
On Thu, Nov 25, 2010 at 3:38 PM, Rex Allen wrote: > > But I also deny that mechanism can account for consciousness (except > by fiat declaration that it does). > > Rex, I am interested in your reasoning against mechanism. Assume there is were an] mechanical brain composed of mechanical neurons, that contained the same information as a human brain, and processed it in the same way. The behavior between these two brains is in all respects identical, since the mechanical neurons react identically to their biological counterparts. However for some unknown reason the computer has no inner life or conscious experience. If you were to ask this mind if it is conscious it would have to say yes, but since it is not conscious, this would be a lie. However, the mechanical mind would not believe itself to be lying. It's neural activity would match the activity of a biological brain telling the truth. It not only is lying about it's claim of consciousness, but would be wrong in its belief that it is conscious. It would be wrong in believing it sees red when you hold a ripe tomato in front of it. My question is what could possibly make the mechanical mind wrong in these beliefs when the biological mind is right? The mechanical mind contains all the same information as the biological one; the information received from the red-sensitive cones in its eyes can be physically traced as it moves through the mechanical mind and leads to it uttering that it sees a tomato. How could this identical informational content be invalid, wrong, false in one representation of a mind, but true in another? Information can take many physical forms. The same digital photograph can be stored as differently reflective areas in a CD or DVD, as charges of electrons in Flash memory, as a magnetic encoding on a hard drive, as holes in a punch card, and yet the file will look the same regardless of how it is physically stored. Likewise the file can be sent in an e-mail which may transmit as fields over an electrical wire, laser pulses in an glass fiber, radio waves in the air, the physical implementation is irrelevant. Is the same not true for information contained within a conscious mind? Jason -- You received this message because you are subscribed to the Google Groups "Everything List" group. To post to this group, send email to everything-l...@googlegroups.com. To unsubscribe from this group, send email to everything-list+unsubscr...@googlegroups.com. For more options, visit this group at http://groups.google.com/group/everything-list?hl=en.
Re: Compatibilism
On Mon, Nov 22, 2010 at 11:40 AM, Bruno Marchal wrote: > On 21 Nov 2010, at 19:47, Rex Allen wrote: >> On Fri, Nov 19, 2010 at 8:32 AM, Bruno Marchal wrote: >>> >>> But your reasoning does not apply to free will in the sense I gave: the >>> ability to choose among alternatives that *I* cannot predict in advance >>> (so that *from my personal perspective* it is not entirely due to reason >>> nor do to randomness). >> >> So that is a good description of the subjective feeling of free will. > > I was not describing the subjective feeling of free will, which is another > matter, and which may accompany or not the experience of free will. > Free-will is the ability to choose among alternatives that *I* cannot > genuinely predict in advance so that reason fails, and yet it is not random. The ability to choose among unpredictable alternatives? What??? In no way does “ability to choose from unpredictable alternatives” match my conception of free will. Nor would you find many people in agreement amongst the general populace. You’re just redefining “free will” in a way that allows you to claim that it exists but which bears little relation to the original conception. In a deterministic universe, there are no alternatives. Things can only unfold one way. Our being unable to predict that unfolding is neither here nor there. Again, ignorance is not free will. Ignorance is just ignorance. >> But if you question most people closely, this isn't what they mean by >> “free will”. > > You have interpret too much quickly what I was describing. Free-will as I > define it is not the subjective feeling of having free-will. It is really > due to the fact that the choice I will make is not based on reason, nor on > randomness from my (real) perspective (which exists). I didn’t say that the options were choices based on “reason or randomness” I said: “Either there is *a reason* for what I choose to do, or there isn't.” By “a reason” I mean “a cause”. I don’t mean “reason” in the sense of rationality. > Subjective does not mean inexisting. Free-will is subjective or better > subject-related, but it exists and has observable consequences, like > purposeful murdering, existence of jails, etc. It is the root of moral > consciousness, or conscience. How does my inability to predict my choices or alternatives in advance serve as the root for moral conscience? >> They mean the ability to make choices that aren't random, but which >> also aren't caused. > > And this becomes, with the approach I gave: "the ability to make choices > that aren't random, but for which they have to ignore the cause". And I > insist: they might even ignore that they ignore the cause. They will say > "because I want do that" or things like that. The vast majority of the populace certainly does not equate free will with ignorance of causes. > I disagree that many people would accept your definition, because it would > entail (even for religious rationalist believers) that free-will does not > exist, and the debate would be close since a long time. If you ask “most people”, they will not agree that the human choice is random, and they will not agree that human choice can be explained by causal forces. Rather, they claim that human choice is something not random *and* not caused. Though they can’t get any more specific than that. The debate isn’t settled because they won’t admit that there is no third option. They feel free, therefore they *believe* that they must actually be free. Free from randomness and free from causal forces. “I feel free, therefore I must be free.” That reasoning is what keeps the free will debate alive. >> They have the further belief that since the choices aren't random or >> caused, the chooser bears ultimate responsibility for them. > > They are right. That is what the materialist eliminativist will deny, and > eventually that is why they will deny any meaning to notion like "person", > free-will, responsibility or even "consciousness". How does ignorance of what choice you will make lead to ultimate responsibility for that choice? I deny the possibility of ultimate responsibility and I’m not a eliminative materialist. But I also deny that mechanism can account for consciousness (except by fiat declaration that it does). As to “person”, I take a deflationary view of the term. There’s less to it than meets the eye. >> This further belief doesn't seem to follow from any particular chain >> of reasoning. It's just another belief that this kind of person has. > > Because as a person she is conscious and feel a reasonable amount of sense > of responsibility, which is genuine and legitimate from her first person > perspective (and from the perspective of machine having a similar level of > complexity). This comes back to my earlier point. She “feels” a sense of responsibility and therefore believes that she is genuinely and legitimately responsible. But the fact that she feels responsibility in no