Matt,

Create a numeric "pleasure" variable in your mind, initialize it with
a positive number and then keep doubling it for some time. Done? How
do you feel? Not a big difference? Oh, keep doubling! ;-))

Regards,
Jiri Jelinek

On Nov 3, 2007 10:01 PM, Matt Mahoney <[EMAIL PROTECTED]> wrote:
> --- "Edward W. Porter" <[EMAIL PROTECTED]> wrote:
> > If bliss without intelligence is the goal of the machines you imaging
> > running the world, for the cost of supporting one human they could
> > probably keep at least 100 mice in equal bliss, so if they were driven to
> > maximize bliss why wouldn't they kill all the grooving humans and replace
> > them with grooving mice.  It would provide one hell of a lot more bliss
> > bang for the resource buck.
>
> Allow me to offer a less expensive approach.  Previously on the singularity
> and sl4 mailing lists I posted a program that can feel pleasure and pain: a 2
> input programmable logic gate trained by reinforcement learning.  You give it
> an input, it responds, and you reward it.  In my latest version, I automated
> the process.  You tell it which of the 16 logic functions you want it to learn
> (AND, OR, XOR, NAND, etc), how much reward to apply for a correct output, and
> how much penalty for an incorrect output.  The program then generates random
> 2-bit inputs, evaluates the output, and applies the specified reward or
> punishment.  The program runs until you kill it.  As it dies it reports its
> life history (its age, what it learned, and how much pain and pleasure it
> experienced since birth).
>
> http://mattmahoney.net/autobliss.txt  (to run, rename to autobliss.cpp)
>
> To put the program in an eternal state of bliss, specify two positive numbers,
> so that it is rewarded no matter what it does.  It won't learn anything, but
> at least it will feel good.  (You could also put it in continuous pain by
> specifying two negative numbers, but I put in safeguards so that it will die
> before experiencing too much pain).
>
> Two problems remain: uploading your mind to this program, and making sure
> nobody kills you by turning off the computer or typing Ctrl-C.  I will address
> only the first problem.
>
> It is controversial whether technology can preserve your consciousness after
> death.  If the brain is both conscious and computable, then Chalmers' fading
> qualia argument ( http://consc.net/papers/qualia.html ) suggests that a
> computer simulation of your brain would also be conscious.
>
> Whether you *become* this simulation is also controversial.  Logically there
> are two of you with identical goals and memories.  If either one is killed,
> then you are in the same state as you were before the copy is made.  This is
> the same dilemma that Captain Kirk faces when he steps into the transporter to
> be vaporized and have an identical copy assembled on the planet below.  It
> doesn't seem to bother him.  Does it bother you that the atoms in your body
> now are not the same atoms that made up your body a year ago?
>
> Let's say your goal is to stimulate your nucleus accumbens.  (Everyone has
> this goal; they just don't know it).  The problem is that you would forgo
> food, water, and sleep until you died (we assume, from animal experiments).
> The solution is to upload to a computer where this could be done safely.
>
> Normally an upload would have the same goals, memories, and sensory-motor I/O
> as the original brain.  But consider the state of this program after self
> activation of its reward signal.  No other goals are needed, so we can remove
> them.  Since you no longer have the goal of learning, experiencing sensory
> input, or controlling your environment, you won't mind if we replace your I/O
> with a 2 bit input and 1 bit output.  You are happy, no?
>
> Finally, if your memories were changed, you would not be aware of it, right?
> How do you know that all of your memories were not written into your brain one
> second ago and you were some other person before that?  So no harm is done if
> we replace your memory with a vector of 4 real numbers.  That will be all you
> need in your new environment.  In fact, you won't even need that because you
> will cease learning.
>
> So we can dispense with the complex steps of making a detailed copy of your
> brain and then have it transition into a degenerate state, and just skip to
> the final result.
>
> Step 1. Download, compile, and run autobliss 1.0 in a secure location with any
> 4-bit logic function and positive reinforcement for both right and wrong
> answers, e.g.
>
>   g++ autobliss.cpp -o autobliss.exe
>   autobliss 0110 5.0 5.0  (or larger numbers for more pleasure)
>
> Step 2. Kill yourself.  Upload complete.
>
>
>
> -- Matt Mahoney, [EMAIL PROTECTED]
>
> -----
> This list is sponsored by AGIRI: http://www.agiri.org/email
> To unsubscribe or change your options, please go to:
> http://v2.listbox.com/member/?&;
>

-----
This list is sponsored by AGIRI: http://www.agiri.org/email
To unsubscribe or change your options, please go to:
http://v2.listbox.com/member/?member_id=8660244&id_secret=60896909-433d26

Reply via email to