On Saturday, September 22, 2012 9:10:30 AM UTC-4, Bruno Marchal wrote:
> On 21 Sep 2012, at 22:48, Craig Weinberg wrote:
> Post from my blog:
> Simple as that, really. From psychological discoveries of the subconscious
> and unconscious, to cognitive bias and logical fallacies, to
> quasi-religious faith in artificial intelligence, we seem to have a mental
> blind spot for emotional realities.
> What could be more human than making emotional mistakes or having one’s
> judgment cloud over because of favoritism or prejudice? Yet when it comes
> to assessing the feasibility of a sentient being composed of programmed
> functions, we tend to miss entirely this little detail: Personal
> preference. Opinion. Bias. It doesn’t bother us that machines completely
> lack this dimension and in all cases exhibit nothing but impersonal
> computation. This tends to lead the feel-blind intellect to unknowingly
> bond to the computer. The consistency of an automaton’s function is
> comforting to our cognitive self, who longs to be free of emotional bias,
> so much so that it is able to hide that longing from itself and project the
> clean lines of perfect consequences outward onto a program.
> It’s not that machines aren’t biased too - of course they are incredibly
> biased toward the most literal interpretations possible, but they are all
> biased in the same exact way so that is seems to us a decent tradeoff. The
> rootless consciousness of the prefrontal cortex thinks that is a small
> price to pay, and one which will inevitably be mitigated with improvements
> in technology. In its crossword puzzle universe of Boolean games, something
> like a lack of personhood or feeling is a minor glitch, an aesthetic ‘to be
> continued’ which need only be set aside for now while the more important
> problems of function can be solved.
> It seems that the ocean of feelings and dreams which were tapped into by
> Freud, Jung, and others in the 20th century have been entirely dismissed in
> favor of a more instrumental approach. Simulation of behaviors. Turing
> machine emulation. This approach has the fatal flaw of drawing the mind
> upside down, with intellect and logic at the base that builds up to complex
> mimicry of mood and inflection. The mind has an ego and doesn’t know it.
> Thinking has promoted itself to a cause of feeling and experience rather
> than a highly specialized and esoteric elaboration of personhood.
> We can see this of course in developmental psychology and anthropology.
> Babies don’t come out of the womb with a flashing cursor, ready to accept
> programming passively. Primitive societies don’t begin with impersonal
> state bureaucracies and progress to chiefdoms. We seem to have to learn
> this lesson again and again that our humanity is not a product of strategy
> and programming, but of authenticity and direct participation.
> When people talk about building advanced robots and computers which will
> be indistinguishable from or far surpass human beings, they always seem to
> project a human agenda on them. We define intelligence outside of ourselves
> as that which serves a function to us, not to the being itself. This again
> suggests to me the reflective quality of the mind, of being blinded by the
> reflection of our own eyes in our sunglasses. Thoughts have a hard time
> assessing the feeling behind themselves, and an even harder time admitting
> that it matters.
> I think we see this more and more in all areas of our lives - an
> overconfidence in theoretical approaches and a continuous disconnecting
> with the results. We keep hoping that it will work this time, even though
> we probably know that it never will. It’s as if our collective psyche is
> waiting for our deluded minds to catch up. Waiting for us to figure out
> that in spite of the graphs and tests and retooling, the machine is really
> not working any better.
> You are right. We have very often dismissed emotion, feelings and
> consciousness in human.
> Unfortunately, dismissing emotion feelings and consciousness in machine,
> will not help.
You don't see a connection between the two? There is no chance of machine
feelings being a psychological projection?
I'm not opposed to the idea of computers having emotions in theory, but the
evidence we've seen so far shows precisely the opposite. If inorganic
machines could grow and change and learn by themselves, then we would
likely see a single example of just that. What we see instead is that even
with many brilliant minds working hard with the finest technology, face a
perpetual uphill battle. In spite of Moore's Law and 30 years of commercial
explosion, there is still no sign of any authentic feeling or intentional
act by a program. What we see is exactly what I would expect from a
fundamentally flawed assumption being dragged out - like Ptolemaic
astronomy...it just isn't working out because we aren't approaching it the
right way. We are trying to build a house on top of a floating roof.
> You received this message because you are subscribed to the Google Groups
> "Everything List" group.
> To view this discussion on the web visit
> To unsubscribe from this group, send email to
> For more options, visit this group at
You received this message because you are subscribed to the Google Groups
"Everything List" group.
To view this discussion on the web visit
To post to this group, send email to firstname.lastname@example.org.
To unsubscribe from this group, send email to
For more options, visit this group at