[note, I'm cc'ing kurzweilAI.net because it's rude to talk about someone
behind their back, at least I'm not being doubly-rude here].
I just finished reading the Big Weenie's new book. I was fairly happy
with the first 280 pages of the thing. It seemed like he was changing
his spiel to advocating a cloud based meta-cortex; very similar to what
I've been advocating recently, actually, minus the cloud part... (I
think that the idea of people putting ALL their information on the cloud
was enough to give certain key google execs the biggest boner of their
lives, hence he was hired.)
I think he just amused himself finding several dozen random quotations
for each chapter and subsection and let his ghost-writer do the rest. Or
that's how it felt.
He spent most of his ink talking about himself. That said, I still agree
with most of his AI ideas, fuzzy and pop-sci as they were. Writers of
pop-sci books should burn in hell for keeping people away from the real
inf0z. U guys suck.
On pages 281 and 282 he decided to burn all the credibility he had built
up through the rest of the book. Basically, what he said was outrageous
enough that I'd propose burning him at the stake.
LOOK, ASSHOLE, YOUR LAW OF ACCELERATING RETURNS IS NOT AS IMPORTANT AS
MY EXISTENCE AS A HUMANOID LIFEFORM. YOU CAN HAVE YOUR COMPUTRONIUM. YOU
CAN UPLOAD INTO YOUR COMPUTRONIUM. YOU CAN CONVERT THE ODD MOON HERE, A
GAS GIANT THERE, A NEBULA OR TWO, EVEN A BLACK HOLE IF YOU CAN FIND A
WAY TO EXPLOIT ONE. KNOCK YOURSELF OUT!!! BUT STOP THERE. LEAVE THE
TERRESTRIAL PLANETS TO THE PEOPLE, FLUFFY BUNNIES, AND FURRY LITTLE
MEOW-MONSTERS WHO NEED THEM!!!
At no point in my life have I ever granted him license to speak about my
destiny. =|
He doesn't explicitly propose an uploading paradigm, you kinda have to
piece it together from paragraphs scattered throughout the book. While I
remain anti-uploading, the outlines he drew were slightly less
unappealing than the standard picture. His proposal for emulation was
based on a medium-high level model of the brain, while still flawed,
actually could surpass the baseline's capabilities, as an AI. The
traditional proposal of low-level simulation would always be much worse.
The bulk of his uploading argument was on pages 240-247, though he
didn't present it as an uploading proposal. I kinda passed over it
because he didn't say anything outrageous in it. My objections to this
thought experiment are much more subtle. My objection is that at some
point during the replacement procedure a paradox will occur where you
are at the point of replacing the exact same part of the brain that
would normally perceive some kind of benefit from the procedure. The
paradox being that the benefit-receiving part of the brain would be
destroyed without receiving any benefit.
My counter-proposal is to apply the concept of mind coalescence and
mind-meld with an external AI, avoiding unnecessary damage to the
original brain. If the procedure is successful, then the mind will have
a much firmer foundation of redundancy to start chucking obsolete parts,
when and if it decides to do so.
This does not completely resolve the paradox, it only provides you with
the tools that will allow you to address it in a way that you might find
acceptable.
This is not to say that after my proposal you'd necessarily go live like
a fairy in computronium. No, you still have the full breadth and depth
of human choices. Your >>> choice <<< of mentality does not at all imply
any choices about your physicality, there you have all of your tastes,
preferences, and desires to guide you.
--
E T F
N H E
D E D
Powers are not rights.
-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription:
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com