Ed, they used to combine ritalin with lsd for psychotherapy. It
assists in absorbing insights achieved from psycholitic doses, which
is a term for doses that are not fully psychedelic. Those are edifying
on their own but are less organized. I don't know if you can get this
in a clinical setting today. But these molecules are gradually being
apprehended as tools

On 11/30/08, Ben Goertzel <[EMAIL PROTECTED]> wrote:
> Ed,
>
> Unfortunately to reply to your message in detail would absorb a lot of
> time, because there are two issues mixed up
>
> 1) you don't know much about computability theory, and educating you
> on it would take a lot of time (and is not best done on an email list)
>
> 2) I may not have expressed some of my weird philosophical ideas about
> computability and mind and reality clearly ... though Abram, at least,
> seemed to "get" them ;)  [but he has a lot of background in the area]
>
> Just to clarify some simple things though: Pi is a computable number,
> because there's a program that would generate it if allowed to run
> long enough....  Also, pi has been proved irrational; and, quantum
> theory really has nothing directly to do with uncomputability...
>
> About
>
>>How can several pounds of matter that is the human brain model
>> the true complexity of an infinity of infinitely complexity things?
>
> it is certainly thinkable that the brain is infinite not finite in its
> information content, or that it's a sort of "antenna" that receives
> information from some infinite-information-content source.  I'm not
> saying I believe this, just saying it's a logical possibility, and not
> really ruled out by available data...
>
> Your reply seems to assume that the brain is a finite computational
> system and that other alternatives don't make sense.  I think this is
> an OK working assumption for AGI engineers but it's not proved by any
> means.
>
> My main point in that post was, simply, that science and language seem
> intrinsically unable to distinguish computable from uncomputable
> realities.  That doesn't necessarily mean the latter don't "exist" but
> it means they're not really scientifically useful entities.  But, my
> detailed argument in favor of this point requires some basic
> understanding of computability math to appreciate, and I can't review
> those basics in an email, it's too much...
>
> ben g
>
> On Sun, Nov 30, 2008 at 4:20 PM, Ed Porter <[EMAIL PROTECTED]> wrote:
>> Ben,
>>
>>
>>
>> On November 19, 2008 5:39 you wrote the following under the above titled
>> thread:
>>
>>
>>
>> ----------------------
>>
>> Ed,
>>
>>
>>
>> I'd be curious for your reaction to
>>
>>
>>
>> http://multiverseaccordingtoben.blogspot.com/2008/10/are-uncomputable-entities-useless-forhtml
>>
>>
>>
>> which explores the limits of scientific and linguistic explanation, in
>>
>> a different but possibly related way to Richard's argument.
>>
>>
>>
>> ----------------------
>>
>>
>>
>> In the below email I asked you some questions about your article, which
>> capture my major problem in understanding it, and I don't think I ever
>> receive a reply
>>
>>
>>
>> The questions were at the bottom of such a long post you may well never
>> have
>> even seen them.  I know you are busy, but if you have time I would be
>> interested in hearing your answers to the following questions about the
>> following five quoted parts (shown in red if you are seeing this in rich
>> text) from you article.  If you are too busy to respond just say so,
>> either
>> on or off list.
>>
>>
>>
>> ---------------------
>>
>>
>>
>> (1) "In the simplest case, A2 may represent U directly in the language,
>> using a single expression"
>>
>>
>>
>> How, can "U" be directly represented in the language if it is
>> uncomputable?
>>
>>
>>
>> I assume you consider any irational number, such as pi to be uncomputable
>> (although, at least pi has a forumula that with enough computation can
>> approach it as a limit –I assume that for most real numbers if there is
>> such
>> a formula, we do not know it.) (By the way, do we know for a fact that pi
>> is
>> irational, and if so how do we know other than that we have caluclated it
>> to
>> millions of places and not yet found an exact solution?)
>>
>>
>>
>> Merely communicating the symbol pi only represents the number if the agent
>> receiving the communication has a more detailed definition, but any
>> definition, such as a formula for iteratively approaching pi, which
>> presumably is what you mean by "R_U" would only be an approximation.
>>
>>
>>
>> So U could never by fully represented unless one had infinite time --- and
>> I
>> generally consider it a waste of time to think about infinate time unless
>> there is something valuable about such considerations that has a use in
>> much
>> more human-sized chunks of time.
>>
>>
>>
>> In fact, it seems the major message of quantum mechanics is that even
>> physical reality doesn't have the time or machinery to compute
>> uncomputable
>> things, like a space constructed of dimensions each correspond to all the
>> real numbers within some astronomical range .  So the real number line is
>> not really real.  It is at best a construct of the human mind that can at
>> best only be approximated in part.
>>
>>
>>
>> (2) "complexity(U) < complexity(R_U)"
>>
>>
>>
>> Because I did not understand how U could be represented, and how R_U could
>> be anything other than an approximation for any practical purposes, I
>> didn't
>> understand the meaning of the above line from your article?
>>
>>
>>
>> If U and R_U have the meaning I guessed in my discussion of quote (1),
>> then
>> U could not be meaningfully representable in the language, other than by a
>> symbol that references some definition (presumably R_U), which, in order
>> even be able to approximate U's uncomputable complexity, would have to be
>> more complex than U itself.
>>
>>
>>
>> Thus, according to this understanding, wouldn't quote (2) always be true?
>>
>>
>>
>>
>>
>> (3) "complexity(real number line R) <>"
>>
>>
>>
>> I didn't understand this formula because I don't know what the "<>" symbol
>> means and I don't know if some text was supposed to follow after it.
>>
>>
>>
>> (4) "If NO, then it means the mind is better off using the axioms for R
>> than
>> using R directly. And, I suggest, that is what we actually do when using R
>> in calculus. We don't use R as an "actual entity" in any strong sense, we
>> use R as an abstract set of axioms."
>>
>>
>>
>> As is stated regarding quote (3) I don't understand what you are saying
>> "NO"
>> to.  But it seems pretty obvious that our minds, and even our computers,
>> do
>> not use R directly (after all the percent of R that is uncomputable would
>> appear to approach 100% as a limit --- even worse it contains an infinity
>> of
>> infinitely complex things), but we have a set of axioms and models about
>> it
>> that are quite useful?
>>
>>
>>
>>
>>
>> (5) "What would YES mean? It would mean that somehow we, as uncomputable
>> beings, used R as an internal source of intuition about continuity ... not
>> thus deriving any conclusions beyond the ones obtainable using the axioms
>> about R, but deriving conclusions in a way that we found subjectively
>> simpler."
>>
>>
>>
>> Again, from my discussion of Quote (3), I don't know what "YES" means. But
>> if by use R as an internal source of intuition about continuity, you mean
>> we
>> that we actually model the true complexity of R, I think that is absurd on
>> its face.  How can several pounds of matter that is the human brain model
>> the true complexity of an infinity of infinitely complexity things?
>>
>>
>>
>> ---------------------
>>
>>
>>
>> I don't understand what your paper on uncomputability has to do with my
>> questions and comments about Richard's paper, other than to highlight that
>> many things are uncomputable, some in theory and many more in practice,
>> and
>> that instead of dealing with many things, imagined or real, in their true
>> complexity our minds deal with simplifications of them.
>>
>>
>>
>> But such simplifications, particularly since they often let us apply more
>> complex analysis where it is most needed, can be very valuable.
>>
>>
>>
>> Furthermore, it is not clear to me that consciousness is not computable.
>> I
>> think it is, in fact, computed.  But I have always felt that a given
>> computation can never fully model or understand itself.
>>
>>
>>
>> Perhaps you are saying that we can never communicate the true complexity
>> of
>> our consciousness to someone else, except, to some extent, by reference to
>> their own consciousnesses --- that when we use words to describe our
>> consciousness we are sending symbols, somewhat like "U" in your article,
>> which is defined by reference to the actual sense of consciousness in
>> someone else that functions very roughly, somewhat like R_U in your paper
>>
>>
>>
>> Ed Porter
>>
>>
>>
>>
>>
>> -----Original Message-----
>> From: Ed Porter [mailto:[EMAIL PROTECTED]
>> Sent: Wednesday, November 19, 2008 9:38 PM
>> To: [email protected]
>> Subject: RE: [agi] A paper that actually does solve the problem of
>> consciousness
>>
>>
>>
>> Ben,
>>
>>
>>
>> I have never assumed language is all-powerful, in fact I have always
>> assumed
>> at least since boarding school, if not years before, that there are severe
>> limits to human understanding.
>>
>>
>>
>> I certainly agree there are limits to what we can understand about
>> consciousness.  A consciousness cannot completely model itself, because
>> that
>> would require the model created in the mind to be as complex as the
>> computation that is modeling it, which seem inherently impossible.
>>
>>
>>
>> But many aspects of reality can be meaningfully represented by models that
>> are substantial simplification of what they are modeling.  Since any
>> aspect
>> of physical reality that we can see or touch without the aid of
>> instruments
>> involves at least 10^20 atoms, each vibrating trillions of times a second,
>> and each having electrons whose Schrödinger equations vibrate something
>> like
>> 10^19 times a second --- we humans naturally perceive, understand, and
>> navigate the world only at the level of extremely gross generalizations.
>> But through the tools of science, including computers, we have been able
>> to
>> create and test models that operate at much finer, or much more complex
>> levels
>>
>>
>>
>> I totally disagree with the notion that consciousness and its relation to
>> the physical world are inexplicable.  Clearly certain aspect of
>> consciousness can be explained in terms of meaningful generalizations.
>> Psychology and brain science have already created many such meaningful
>> generalizations.  Richard seems to admit as much, when he dismissed all
>> the
>> examples I have given in this thread of scientific knowledge about
>> consciousness as merely examples of the easy problems of consciousness.
>> Easy or not, they are meaningful explanations about consciousness.
>>
>>
>>
>> Our understanding of the human mind has grown tremendously in the last
>> decade and the rate of of our learning on the subject is rapidly
>> accelerating.  This includes our understanding of the physical correlates
>> of
>> consciousness.
>>
>>
>>
>> So, do I think we will ever understand everything about consciousness ---
>> of
>> course not.  But do I think that within fifty years we will know much,
>> much
>> more about it – of course.
>>
>>
>>
>> In fact, I think we will come to understand the sense of awareness that we
>> experience in our own consciousnesses as a natural result of a certain
>> type
>> of computation, one which has an extremely rich, but somewhat coherently
>> controlled feedback loop with its own extremely complex internal state.
>>
>>
>>
>> Why does do our bodies sense reaility?  Because they are located within
>> it,
>> and have systems for sensing and affecting realities.
>>
>>
>>
>> Why do we experience consciousness?  Because computation in our mind is
>> located within the mind and has system for sensing and affecting its
>> states.
>>
>>
>>
>> Its not quite that simple, but that is a central part of the puzzle.  It
>> is
>> sort of a Zen thing.  But I hope people on this list open your mind to the
>> concept that a human consciousness is a special type of computation.  It
>> is
>> a computation that includes the generation of a sense of experiencing and
>> understanding a sequence of conscious concepts by simultaneous activation
>> of
>> prior experiences related to each such concept, being project into a mind
>> previously activated by the grounding of previously selected concepts, so
>> as
>> to provide a sense of grounding for those concepts that is appropriate to
>> the sequence of prior activations.
>>
>>
>>
>> The complexity of the brain, just in terms of neurons (10^11) is equal to
>> a
>> large football stadium (10^5) seats in which every single seat is itself a
>> 10 large football stadiums in which each seat corresponds to a agent with
>> connections to 100 to 10,000 other such agents, having memory at each of
>> those connections, and there are mechanism for communicating information
>> to
>> the whole stadium of stadiums at once, and there are lots of local
>> channels
>> and screens.
>>
>>
>>
>> Image, what a complexly dynamic crowd that could be.
>>
>>
>>
>> The more you think about it the more it makes sense.  I read a book called
>> the "Minds of Robots" in 1964 which said that conscious news was
>> computation, but it was not until after I was well into my '69-70
>> independent study on AI with comprehensive reading list from Minsky, and I
>> understood roughly the numbers associated with the computation of the
>> brain,
>> had understood experiential computing based on Minsky K-line theory, and
>> had
>> thought about it in a couple on acid trips that I truly started to
>> understand how straight forward such a statement is.
>>
>>
>>
>> (Please note I have not taken any acid in over three decades and I an not
>> advocating its use outside of the care of a responsible psychiatrist.)
>>
>>
>>
>> ----------------
>>
>>
>>
>> With regard to your paper, I read it, but I did not spend the time that
>> probably would be required for me to understand it.
>>
>>
>>
>> Unlike you, I was only briefly bothered by the fact that most of the real
>> number line was full of irrational numbers.  I had very little trouble
>> understanding the concept of a limit in calculus.  If some solution can be
>> reasonably be shown to have an error smaller than any you would ever be
>> concerned with, that's good enough for me.
>>
>>
>>
>> I guess this is because I have never been one for theoretical purity.  In
>> fact, I tend to instinctually distrust it.  In fact, I think all kids
>> should
>> be taught in school to distrust, at least to some degree, all theories (as
>> well as their own senses and memories).  I was originally in favor of
>> spending one day a year in high school science to discuss intelligent
>> design
>> because if it would be part of an honest discussion about why, when, and
>> to
>> what degree we should trust scientific theory.  Once I found that most of
>> the intelligent design texts were total, closed minded propaganda, I
>> changed
>> my mind.
>>
>>
>>
>> There were some parts of your paper I particularly did not under stand.
>> Let
>> me quote them and then ask you about them.
>>
>>
>>
>> ---------------------
>>
>>
>>
>> (1) In the simplest case, A2 may represent U directly in the language,
>> using
>> a single expression"
>>
>>
>>
>> How, can U be directly represented in the language if it is uncomputable?
>>
>>
>>
>> I assume you consider any irational number, such as pi to be uncomputable
>> (although, at least pi has a forumula that with enough computation can be
>> approach it as a limit – as assume the for most real numbers if there is
>> such a formula, we do not know it.)  Merely communicating the symbol pi
>> only
>> represents the number if the agent receiving the communication has a more
>> detailed definition, but any definition, such as a formula for iteratively
>> approaching pi, which presumably is what you mean by R_U would only be an
>> approximation.
>>
>>
>>
>> So U could never by fully represented unless one had infinite time --- and
>> I
>> generally consider it a waste of time to think about infinate time unless
>> there is something valuable about such considerations that has a use in
>> much
>> more human-sized chunks of time.
>>
>>
>>
>> In fact, it seems the major message of quantum mechanic is that even
>> physical reality doesn't have the time or machinery to compute
>> uncomputable
>> things.  So the real number line is not really real.  It is at best a
>> construct of the human mind that can at best only be approximated in part.
>>
>>
>>
>> (2) complexity(U) < complexity(R_U)
>>
>>
>>
>> Because I did not understand how U could be represented, and how R_U could
>> be anything other than an approximation for any practical purposes, I
>> didn't
>> understand the meaning of the above line from your article?
>>
>>
>>
>> If U and R_U have the meaning I guessed in my discussion of text quote
>> (1),
>> above, U could not be meaningfully representable in the language, other
>> than
>> by a symbol that reference some definition (presumably R_U), which, in
>> order
>> even be able to approximate U's uncomputable complexity, would have to be
>> more complex than U itself.
>>
>>
>>
>> So why wouldn't this inequality always be true?
>>
>>
>>
>> (3) complexity(real number line R) <>
>>
>>
>>
>> I didn't understand this formula because I don't know what the "<>" symbol
>> means and I don't know if some text was supposed to follow after it.
>>
>>
>>
>> (4) If NO, then it means the mind is better off using the axioms for R
>> than
>> using R directly. And, I suggest, that is what we actually do when using R
>> in calculus. We don't use R as an "actual entity" in any strong sense, we
>> use R as an abstract set of axioms.
>>
>>
>>
>> From quote (4) above it is clear I don't understand what you are saying
>> "NO"
>> to.  But it seems pretty obvious that our minds, and even our computers,
>> do
>> not use R directly (after all the percent of it that is uncomputable would
>> appear to approach 100% as a limit --- even worse it contains an infinity
>> of
>> infinitely complex things), but we have a set of axioms and models about
>> it
>> that are quite useful?
>>
>>
>>
>>
>>
>> (5) What would YES mean? It would mean that somehow we, as uncomputable
>> beings, used R as an internal source of intuition about continuity ... not
>> thus deriving any conclusions beyond the ones obtainable using the axioms
>> about R, but deriving conclusions in a way that we found subjectively
>> simpler.
>>
>>
>>
>> Again from Quote (4) I don't know what "YES" means. But if by use R as an
>> internal source of intuition about continuity, you mean we that we
>> actually
>> model the true complexity of R, I think that is absurd on its face.  How
>> can
>> several pounds of matter that is the human brain model an infinity of
>> infinitely complexity things?
>>
>>
>>
>> ---------------------
>>
>>
>>
>> I don't understand what your paper on uncomputability has to do with my
>> questions and comments about Richard's paper, other than to highlight
>> profoundly that many things are uncomputable.   But for at least since my
>> 1969-1970 study of AI I have always felt the true complexity of any one
>> human consciousness would be far beyond human comprehension.  After all I
>> am
>> the one attacking Richard's paper for not discussing depth and complexity
>> of
>> computation as a source of our perception of consciousnesses richness.
>>
>>
>>
>> As I said above, it is clearly uncomputable to have the brain be able to
>> understand itself completely, or even anything close to completely.  I
>> have
>> never doubted that.  But that does not necessarily mean that we cannot
>> come
>> to know the brain and mind, through the use of computers and models, with
>> as
>> much specificity as we can understand most aspects of physical reality
>> that
>> are anywhere nearly as complex.
>>
>>
>>
>> I do not think the human mind is all irreducible complexity.  Remember
>> that
>> chaos theory is the study of systems that are a mixture of randomness and
>> regularity. Despite its complexity, I think human consciousness has enough
>> regularities that in fifty years we will have a surprising and
>> philosophically transformative degree of understanding about it – although
>> not total understanding.
>>
>>
>>
>> Remember most people in the know are predicting human-level AI in 20
>> years,
>> so that would mean the level of understanding we would have of the human
>> mind in 50 years, barring a major collapse of civilization, would benefit
>> from a full 30 years of superhuman intelligence and immeasurably better
>> brain scanning and interfacing technology.
>>
>>
>>
>> Ed Porter
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>>
>> -----Original Message-----
>> From: Ben Goertzel [mailto:[EMAIL PROTECTED]
>> Sent: Wednesday, November 19, 2008 5:39 PM
>>
>> To: [email protected]
>> Subject: Re: [agi] A paper that actually does solve the problem of
>> consciousness
>>
>>
>>
>> Ed,
>>
>>
>>
>> I'd be curious for your reaction to
>>
>>
>>
>> http://multiverseaccordingtoben.blogspot.com/2008/10/are-uncomputable-entities-useless-for.html
>>
>>
>>
>> which explores the limits of scientific and linguistic explanation, in
>>
>> a different but possibly related way to Richard's argument.
>>
>>
>>
>> Science and language are powerful tools for explanation but there is
>>
>> no reason to assume they are all-powerful.  We should push them as far
>>
>> as we can, but no further...
>>
>>
>>
>> I agree with Richard that according to standard scientific notions of
>>
>> explanation, consciousness and its relation to the physical world are
>>
>> inexplicable.  My intuition and reasoning are probably not exactly the
>>
>> same as his, but there seems some similarity btw our views...
>>
>>
>>
>> -- Ben G
>>
>>
>>
>>
>>
>> On Wed, Nov 19, 2008 at 5:27 PM, Ed Porter <[EMAIL PROTECTED]> wrote:
>>
>>> Richard,
>>
>>>
>>
>>>
>>
>>>
>>
>>> (the second half of this post, that starting with the all capitalized
>>
>>> heading, is the most important)
>>
>>>
>>
>>>
>>
>>>
>>
>>> I agree with your extreme cognitive semantics discussion.
>>
>>>
>>
>>>
>>
>>>
>>
>>> I agree with your statement that one criterion for "realness" is the
>>
>>> directness and immediateness of something's phenomenology.
>>
>>>
>>
>>>
>>
>>>
>>
>>> I agree with your statement that, based on this criterion for "realness,"
>>
>>> many conscious phenomena, such as qualia, which have traditionally fallen
>>
>>> under the hard problem of consciousness seem to be "real."
>>
>>>
>>
>>>
>>
>>>
>>
>>> But I have problems with some of the conclusions you draw from these
>>> things,
>>
>>> particularly in your "Implications" section at the top of the second
>>> column
>>
>>> on Page 5 of your paper.
>>
>>>
>>
>>>
>>
>>>
>>
>>> There you state
>>
>>>
>>
>>>
>>
>>>
>>
>>> "…the correct explanation for consciousness is that all of its various
>>
>>> phenomenological facets deserve to be called as "real" as any other
>>> concept
>>
>>> we have, because there are no meaningful objective standards that we
>>> could
>>
>>> apply to judge them otherwise."
>>
>>>
>>
>>>
>>
>>>
>>
>>> That aspects of consciousness seem real does not provides much of an
>>
>>> "explanation for consciousness."  It says something, but not much.  It
>>> adds
>>
>>> little to Descartes' "I think therefore I am."  I don't think it provides
>>
>>> much of an answer to any of the multiple questions Wikipedia associates
>>> with
>>
>>> Chalmer's hard problem of consciousness.
>>
>>>
>>
>>>
>>
>>>
>>
>>> You further state that some aspects of consciousness have a unique status
>>> of
>>
>>> being beyond the reach of scientific inquiry and give a purported reason
>>> why
>>
>>> they are beyond such a reach. Similarly you say:
>>
>>>
>>
>>>
>>
>>>
>>
>>> "…although we can never say exactly what the phenomena of consciousness
>>> are,
>>
>>> in the way that we give scientific explanations for other things, we can
>>
>>> nevertheless say exactly why we cannot say anything: so in the end, we
>>> can
>>
>>> explain it."
>>
>>>
>>
>>>
>>
>>>
>>
>>> First, I would point out as I have in my prior papers that, given the
>>
>>> advances that are expected to be made in AGI, brain scanning and brain
>>
>>> science in the next fifty years, it is not clear that consciousness is
>>
>>> necessarily any less explainable than are many other aspects of physical
>>
>>> reality.  You admit there are easy problems of consciousness that can be
>>
>>> explained, just as there are easy parts of physical reality that can be
>>
>>> explained. But it is not clear that the percent of consciousness that
>>> will
>>
>>> remain a mystery in fifty years is any larger than the percent of basic
>>
>>> physical reality that will remain a mystery in that time frame.
>>
>>>
>>
>>>
>>
>>>
>>
>>> But even if we accept as true your statement that certain phenomena of
>>
>>> consciousness are beyond analysis, that does little to explain
>>
>>> consciousness.  In fact, it does not appear to answer any of the hard
>>
>>> problems of consciousness.  For example, just because (a) we are
>>> conscious
>>
>>> of the distinction used in our own mind's internal representation between
>>
>>> sensation of the colors red and blue, (b) we allegedly cannot analyze
>>> that
>>
>>> difference further, and (c) that distinction seems subjectively real to
>>> us
>>
>>> --- that does not shed much light on whether or not a p-zombie would be
>>
>>> capable of acting just like a human without having consciousness of red
>>> and
>>
>>> blue color qualia.
>>
>>>
>>
>>>
>>
>>>
>>
>>> It is not even clear to me that your paper shows consciousness is not an
>>
>>> "artifact, " as your abstract implies.  Just because something is "real"
>>
>>> does not mean it is not an "artifact", in many senses of the word, such
>>> as
>>
>>> an unintended, secondary, or unessential, aspect of something.
>>
>>>
>>
>>>
>>
>>>
>>
>>>
>>
>>>
>>
>>> THE REAL WEAKNESS OF YOUR PAPER IS THAT IS PUTS WAY TOO MUCH EMPHASIS ON
>>> THE
>>
>>> PART OF YOUR MOLECULAR FRAMEWORK THAT ALLEGEDLY BOTTOMS OUT, AND NOT
>>> ENOUGH
>>
>>> ON THE PART OF THE FRAMEWORK YOU SAY REPORTS A SENSE OF REALNESS DESPITE
>>
>>> SUCH BOTTOMING OUT  -- THE SENSE OF REALNESS THAT IS MOST ESSENTIAL TO
>>
>>> CONSCIOUSNESS.
>>
>>>
>>
>>>
>>
>>>
>>
>>> It is my belief that if you want to understand consciousness in the
>>> context
>>
>>> of the types of things discussed in your paper, you should focus the part
>>> of
>>
>>> the molecular framework, which you imply it is largely in the foreground,
>>
>>> that prevents the system from returning with no answer, even when trying
>>> to
>>
>>> analyze a node such as a lowest level input node for the color red in a
>>
>>> given portion of the visual field.
>>
>>>
>>
>>>
>>
>>>
>>
>>> This is the part of your molecular framework that
>>
>>>
>>
>>>
>>
>>>
>>
>>> "…because of the nature of the representations used in the foreground,
>>> there
>>
>>> is no way for the analysis mechanism to fail to return some kind of
>>> answer,
>>
>>> because a non-existent answer would be the same as representing the color
>>> of
>>
>>> red as "nothing," and in that case all colors would be the same." (Page
>>> 3,
>>
>>> Col.2, first full paragraph.)
>>
>>>
>>
>>>
>>
>>>
>>
>>> It is also presumably the part of your molecular framework that
>>
>>>
>>
>>>
>>
>>>
>>
>>> "…report that 'There is definitely something that it is like to be
>>
>>> experiencing the subjective essence of red, but that thing is ineffable
>>> and
>>
>>> inexplicable.' " (Page 3, Col. 2, 2nd full paragraph.)
>>
>>>
>>
>>>
>>
>>>
>>
>>> This is the part of your system that is providing the subjective
>>> experience
>>
>>> that you say is providing the "realness" to your conscious experience.
>>> This
>>
>>> is where your papers should focus.  How does it provide this sense of
>>
>>> realness.
>>
>>>
>>
>>>
>>
>>>
>>
>>> Unfortunately, your description of the molecular framework provides some,
>>
>>> but very little, insight into what might be providing this subjective
>>> sense
>>
>>> of experience, that is so key to the conclusions of your paper.
>>
>>>
>>
>>>
>>
>>>
>>
>>> In multiple prior posts on this thread I have said I believe the real
>>> source
>>
>>> of consciousness appears to lie in such a molecular framework, but that
>>> to
>>
>>> have anything approaching a human level of such consciousness this
>>
>>> framework, and its computations that give rise to consciousness, have to
>>> be
>>
>>> extremely complex.  I have also emphasized that brain scientist who have
>>
>>> already done research on the neural correlates of consciousness, tend to
>>
>>> indicate humans usually only report consciousness of things associated
>>> with
>>
>>> fairly broad spread neural activation, which would normally involve many
>>
>>> billions or trillions of inter-neuron messages per second.  I have
>>> posited
>>
>>> that widespread activation of the nodes directly and indirectly
>>> associated
>>
>>> with a given "conscious" node, provides dynamic grounding for the meaning
>>> of
>>
>>> the conscious node.
>>
>>>
>>
>>>
>>
>>>
>>
>>> As I have pointed out, we know of nothing about physical reality that is
>>
>>> anything other than computation (if you consider representation to be
>>> part
>>
>>> of computation).  Similarly there is nothing our subjective experience
>>> can
>>
>>> tell us about our own consciousnesses that is other than computation.
>>> One
>>
>>> of the key words we humans use to describe our consciousnesses is
>>
>>> "awareness."  Awareness is created by computation.  It is my belief that
>>
>>> this awareness comes from the complex, dynamically focused, and
>>> meaningful
>>
>>> way in which our thought processes compute in interaction with
>>> themselves.
>>
>>>
>>
>>>
>>
>>>
>>
>>> Ed Porter
>>
>>>
>>
>>>
>>
>>>
>>
>>> P.S. (With regard to the alleged bottoming out reported in your papert:
>>> as
>>> I
>>
>>> have pointed out in previous threads, even the lowest level nodes in any
>>
>>> system would normally have associations that would give them a type and
>>
>>> degree of grounding and, thus, further meaning  So that spreading
>>> activation
>>
>>> would normally not bottom out when it reaches the lowest level nodes.
>>> But
>>
>>> it would be subject to circularly, or a lack of information about lowest
>>
>>> nodes other than what could be learned from their associations with other
>>
>>> nodes in the system.)
>>
>>>
>>
>>>
>>
>>>
>>
>>>
>>
>>>
>>
>>>
>>
>>>
>>
>>> -----Original Message-----
>>
>>> From: Richard Loosemore [mailto:[EMAIL PROTECTED]
>>
>>> Sent: Wednesday, November 19, 2008 1:57 PM
>>
>>>
>>
>>> To: [email protected]
>>
>>> Subject: Re: [agi] A paper that actually does solve the problem of
>>
>>> consciousness
>>
>>>
>>
>>>
>>
>>>
>>
>>> Ben Goertzel wrote:
>>
>>>
>>
>>>> Richard,
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> I re-read your paper and I'm afraid I really don't grok why you think it
>>
>>>
>>
>>>> solves Chalmers' hard problem of consciousness...
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> It really seems to me like what you're suggesting is a "cognitive
>>
>>>
>>
>>>> correlate of consciousness", to morph the common phrase "neural
>>
>>>
>>
>>>> correlate of consciousness" ...
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> You seem to be stating that when X is an unanalyzable, pure atomic
>>
>>>
>>
>>>> sensation from the perspective of cognitive system C, then C will
>>
>>>
>>
>>>> perceive X as a raw quale ... unanalyzable and not explicable by
>>
>>>
>>
>>>> ordinary methods of explication, yet, still subjectively real...
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> But, I don't see how the hypothesis
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> "Conscious experience is **identified with** unanalyzable mind-atoms"
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> could be distinguished empirically from
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> "Conscious experience is **correlated with** unanalyzable mind-atoms"
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> I think finding cognitive correlates of consciousness is interesting,
>>
>>>
>>
>>>> but I don't think it constitutes solving the hard problem in Chalmers'
>>
>>>
>>
>>>> sense...
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> I grok that you're saying "consciousness feels inexplicable because it
>>
>>>
>>
>>>> has to do with atoms that the system can't explain, due to their role as
>>
>>>
>>
>>>> its primitive atoms" ... and this is a good idea, but, I don't see how
>>
>>>
>>
>>>> it bridges the gap btw subjective experience and empirical data ..
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> What it does is explain why, even if there *were* no hard problem,
>>
>>>
>>
>>>> cognitive systems might feel like there is one, in regard to their
>>
>>>
>>
>>>> unanalyzable atoms
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> Another worry I have is: I feel like I can be conscious of my son, even
>>
>>>
>>
>>>> though he is not an unanalyzable atom.  I feel like I can be conscious
>>
>>>
>>
>>>> of the unique impression he makes ... in the same way that I'm conscious
>>
>>>
>>
>>>> of redness ... and, yeah, I feel like I can't fully explain the
>>
>>>
>>
>>>> conscious impression he makes on me, even though I can explain a lot of
>>
>>>
>>
>>>> things about him...
>>
>>>
>>
>>>>
>>
>>>
>>
>>>> So I'm not convinced that atomic sensor input is the only source of raw,
>>
>>>
>>
>>>> unanalyzable consciousness...
>>
>>>
>>
>>>
>>
>>>
>>
>>> My first response to this is that you still don't seem to have taken
>>
>>>
>>
>>> account of what was said in the second part of the paper  -  and, at the
>>
>>>
>>
>>> same time, I can find many places where you make statements that are
>>
>>>
>>
>>> undermined by that second part.
>>
>>>
>>
>>>
>>
>>>
>>
>>> To take the most significant example:  when you say:
>>
>>>
>>
>>>
>>
>>>
>>
>>>  > But, I don't see how the hypothesis
>>
>>>
>>
>>>  >
>>
>>>
>>
>>>  > "Conscious experience is **identified with** unanalyzable mind-atoms"
>>
>>>
>>
>>>  >
>>
>>>
>>
>>>  > could be distinguished empirically from
>>
>>>
>>
>>>  >
>>
>>>
>>
>>>  > "Conscious experience is **correlated with** unanalyzable mind-atoms"
>>
>>>
>>
>>>
>>
>>>
>>
>>> ... there are several concepts buried in there, like [identified with],
>>
>>>
>>
>>> [distinguished empirically from] and [correlated with] that are
>>
>>>
>>
>>> theory-laden.  In other words, when you use those terms you are
>>
>>>
>>
>>> implictly applying some standards that have to do with semantics and
>>
>>>
>>
>>> ontology, and it is precisely those standards that I attacked in part 2
>>
>>>
>>
>>> of the paper.
>>
>>>
>>
>>>
>>
>>>
>>
>>> However, there is also another thing I can say about this statement,
>>
>>>
>>
>>> based on the argument in part one of the paper.
>>
>>>
>>
>>>
>>
>>>
>>
>>> It looks like you are also falling victim to the argument in part 1, at
>>
>>>
>>
>>> the same time that you are questioning its validity:  one of the
>>
>>>
>>
>>> consequences of that initial argument was that *because* those
>>
>>>
>>
>>> concept-atoms are unanalyzable, you can never do any such thing as talk
>>
>>>
>>
>>> about their being "only correlated with a particular cognitive event"
>>
>>>
>>
>>> versus "actually being identified with that cognitive event"!
>>
>>>
>>
>>>
>>
>>>
>>
>>> So when you point out that the above distinction seems impossible to
>>
>>>
>>
>>> make, I say:  "Yes, of course:  the theory itself just *said* that!".
>>
>>>
>>
>>>
>>
>>>
>>
>>> So far, all of the serious questions that people have placed at the door
>>
>>>
>>
>>> of this theory have proved susceptible to that argument.
>>
>>>
>>
>>>
>>
>>>
>>
>>> That was essentially what I did when talking to Chalmers.  He came up
>>
>>>
>>
>>> with an objection very like the one you gave above, so I said: "Okay,
>>
>>>
>>
>>> the answer is that the theory itself predicts that you *must* find that
>>
>>>
>>
>>> question to be a stumbling block ..... AND, more importantly, you should
>>
>>>
>>
>>> be able to see that the strategy I am using here is a strategy that I
>>
>>>
>>
>>> can flexibly d
>>
>> ...
>>
>> [Message clipped]
>
>
>
> --
> Ben Goertzel, PhD
> CEO, Novamente LLC and Biomind LLC
> Director of Research, SIAI
> [EMAIL PROTECTED]
>
> "I intend to live forever, or die trying."
> -- Groucho Marx
>
>
> -------------------------------------------
> agi
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>


-------------------------------------------
agi
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=8660244&id_secret=120640061-aded06
Powered by Listbox: http://www.listbox.com

Reply via email to