I didn't mean to say that using probability and weighted reasoning were
wrong or something.  I just meant that you cannot use probability without
the supposition of a logically sound frame (or something).
Jim Bromer

On Sun, Mar 17, 2013 at 9:44 AM, Jim Bromer <[email protected]> wrote:

> Charles,
> The notion of the probability of a prediction (an expectation in general
> human language) is nonsensical if you have ruled out assertions.  Since you
> are not ruling out assertions you are unwittingly allowing the notion of
> "truth" in the front door even as you are chasing it out the back. An
> assertion is a notion of the truth of the assertion as a likely
> possibility. If you are saying that an AGI program was capable of somewhat
> reliably deducing the probability of a prediction then you are asserting
> that the process was based on the strength and the truth of the application
> of the methods used to derive those probabilities.
>
> If it were easy for a computer program to attain the probability of an
> event based on observations of past events then this kind of discussion
> would not be relevant to AGI.  The problem is that events are actually
> complexities which are not only composed of distinct 'kinds' of events and
> some background 'noise' but of different variations of 'kinds' of events
> and a lot of other events.  The science of using probability and
> statistics is premised on the methodical actions of an agent who is not
> only intelligent but highly trained in the science of the applied
> statistical methods.  The idea that intelligence can be founded on
> statistics is backwards.
> Jim Bromer
>
> On Tue, Feb 19, 2013 at 7:56 PM, Charles Hixson <
> [email protected]> wrote:
>
>>  On 02/19/2013 11:03 AM, Piaget Modeler wrote:
>>
>>
>>  I'm sure this topic has been discussed before.  Sorry for rehashing it
>> if so. I have a specific question I'd like to answer.
>>
>>
>>  In designing a cognitive system, someone made a criticism that utterly
>> confounded me.  And got me thinking.
>>
>>  The system receives sensory data sets from the world and transforms
>> them into percept propositions which it asserts to
>> its memory.  Each percept proposition is activated when it is asserted.
>>  Infereneces are made from these percepts.
>> These initial percepts and its inferences are called "Observables".  All
>> observables can be activated, but there is only a
>> notion of activation.
>>
>>  Next, the system can predict that these observables will recur at some
>> point.  But the prediction refers only to predicting
>> the re-activation of observables.
>>
>>  Then some one asked, where is the notion of TRUTH in your system.  I
>> was flabbergasted.  Speechless. Then I asked
>> well what is truth?  I checked wikipedia.  (
>> http://en.wikipedia.org/wiki/Truth )
>>
>>
>>  It turns out that when someone says something is true, it means a very
>> many things:
>>
>>  a) It means that the statement is logically consistent (validity),
>> b) that the statement corresponds, concurs, or conforms to reality
>> (verity),
>> c) that one is sure of the statement (certainty / confidence),
>> d) that the statement is likely to occur rather than unlikely
>> (Likelihood), and
>> e) that we agree with the statement (agreement).
>>
>>  So my questions are:
>>
>>  (1) Is truth necessary or important to a cognitive system?
>> (2) Which notion of truth should a cognitive system model?
>> (3) How do we ascribe truth (values) to sensory input or inferences
>> derived from sensory input?
>>
>>  Your thoughts?
>>
>>  ~PM.
>>
>>
>> ------------------------------------------------------------------------------------------------------------------------------------------------
>> *Confidential *- *This message is meant solely for the intended
>> recipient. Please do not copy or forward this message without *
>> *the consent of the sender. If you have received this message in error,
>> please delete the message and notify the sender.*
>>     *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/232072-58998042> | 
>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>> <http://www.listbox.com>
>>
>> Truth is an illusion.  It is the belief that what you believe to be most
>> likely is, in fact, inevitable.
>>
>> An AI doesn't need the concept of truth...except to communicate with
>> people.  Internally it can operate off of graded degrees of probability,
>> cost, benefit, etc.  When communicating with people it needs to condense
>> that so that when something has more than a certain amount of probability,
>> and the benefit of asserting it is sufficiently large, and the cost of
>> being wrong is sufficiently small, then it synopsizes this as proclaiming
>> "truth".  It's my belief that people operate in the same way, though this
>> is disguised because different people use different constraints on things
>> like "What is probable enough?".  Also note that the cost and the benefit
>> are figured on the basis of the cost/benefit to the entity proclaiming a
>> truth rather than on those accepting it.
>>
>> So perhaps we would want a sufficiently capable AI to avoid talking about
>> truth, and instead talk about what the probabilities are, and what costs
>> and benefits can be expected.  It's a bit harder to understand, but it
>> strikes me as much safer.
>>
>> --
>> Charles Hixson
>>
>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>> <https://www.listbox.com/member/archive/rss/303/10561250-470149cf> |
>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>> <http://www.listbox.com>
>>
>
>



-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to