On 10/15/2018 5:06 PM, Philip Thrift wrote:


On Monday, October 15, 2018 at 6:45:11 PM UTC-5, Brent wrote:



    On 10/14/2018 11:13 PM, Philip Thrift wrote:


    On Sunday, October 14, 2018 at 9:53:07 PM UTC-5, Brent wrote:



        On 10/14/2018 2:48 PM, John Clark wrote:

            >/And there are sound reasons for doubting the
            consciousness of computers -/

        Name one of them that could not also be used to doubt the
        consciousness of your fellow human beings.

        The reason for not doubting that other human beings are
        conscious is that (1) I am conscious and (2) other human
        beings are made of the same stuff in approximately the same
        way that I am and (3) they behave the same way in relation to
        what I am conscious of, e.g. they jump at a sudden loud sound.

        Brent



    The thought crossed my mind yesterday: I was helping a young man
    applying for a Ph.D. program in chemical engineering with his
    application, and we were talking about chemistry and
    consciousness*, and I mentioned a type of zombie - a being that
    could converse (like an advanced Google Assistant or Sophie
    Robot) but not be conscious - and I thought it was *possible* he
    was a zombie.

    * cf.
    *Experience processing*
    https://codicalist.wordpress.com/2018/10/14/experience-processing/
    <https://codicalist.wordpress.com/2018/10/14/experience-processing/>

    I suppose you've read Scott Aaronson's take down of Tononi's
    theory.  So I wonder why you would reference Tononi.

    One problem with the "experience is primary" theory is that
    there's no way for it to evolve.  If it's a property of matter why
    are organized information processing lumps of matter more capable
    of experience than unorganized lumps of the same composition...the
    obvious answer is that the former process information, and
    processing information is something natural selection can work
    on.  Smart animals reproduce better.  Animals with
    experiences...who cares?

    “Emotional-BDI agents are BDI agents whose behavior is guided not
    only by beliefs, desires and intentions, but also by the role of
    emotions in reasoning and decision-making." This makes a false
    assumption that emotions are something independent of beliefs,
    desire, intentions, reasoning, and decision making.  But this says
    nothing about the satisfaction and thwarting of desires and
    intentions.  Why are those enough to explain emotions.  I agree
    that emotions are necessary for reasoning in the sense that
    emotions are the value-weights given to events, including those
    imagined by foresight, and that some values are primitive.

    I think it is false that "Purely informational processing, which
    includes intentional agent programming (learning from experience,
    self-modeling), does not capture all true experiential processing
    (phenomenal consciousness). "  It is a cheat to put in "purely". 
    In fact all learning and intentional planning must include
    weighing alternatives and assigning value/emotion to them.  I
    don't see any need for a further primitive modality.  For example,
    a feeling of dizziness is a failure to maintain personal spacial
    orientation which is a value at a very low (subconscious) level. 
    Sure there are feelings and emotions...but I think they are all
    derivative from more primitive values that are derivative from
    evolution.

    I think the reason you are attracted to this idea is that it is
    closed within the computer/information/program frame. And that is
    why I use the example of the AI Mars Rover. Sure, emotion cannot
    be derived within a computer.  Emotion is something useful to a
    robot, an AI that works and strives within an external world which
    also acts on it.

    Brent




I'll include the reference to

*Why I Am Not An Integrated Information Theorist (or, The Unconscious Expander)*
   https://www.scottaaronson.com/blog/?p=1799

Since I say that IIT is still in the /information-oriented paradigm /and not in the /experience-oriented paradigm/, Aaronson's post helps my case.

I don't quite follow the rest. A person may feel pleasure (a modality) without reasoning "I need to feel pleasure now."


It's not a question of what it's possible to feel, but whether that it can be accounted for by information processing.  "I feel pleasure now." may well be a function of specific perceptions, values, and reasoning about them.  I don't see any attempt to prove this cannot be the case.  It seems that helping yourself to a primitive "experience" built into matter is just baseless speculation unless you have some project to measure or characterize this experience and show how it interacts with information...because we certainly know that information can give pain or pleasure.

Brent

--
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at https://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to