When I was reading Aristotle I remember some passage that basically
asked "how could nothing be true and nothing be false?"  Answer:  if
there were no people.  That's not the end of it, obviously, but
important.

On 3/17/13, Jim Bromer <[email protected]> wrote:
> An assertion is the notion of the truth of the assertion as something that
> can be considered (is what I meant to say of course.)
>
> On Sun, Mar 17, 2013 at 3:06 PM, Jim Bromer <[email protected]> wrote:
>
>> I didn't mean to say that using probability and weighted reasoning were
>> wrong or something.  I just meant that you cannot use probability without
>> the supposition of a logically sound frame (or something).
>> Jim Bromer
>>
>> On Sun, Mar 17, 2013 at 9:44 AM, Jim Bromer <[email protected]> wrote:
>>
>>> Charles,
>>> The notion of the probability of a prediction (an expectation in general
>>> human language) is nonsensical if you have ruled out assertions.  Since
>>> you
>>> are not ruling out assertions you are unwittingly allowing the notion of
>>> "truth" in the front door even as you are chasing it out the back. An
>>> assertion is a notion of the truth of the assertion as a likely
>>> possibility. If you are saying that an AGI program was capable of
>>> somewhat
>>> reliably deducing the probability of a prediction then you are asserting
>>> that the process was based on the strength and the truth of the
>>> application
>>> of the methods used to derive those probabilities.
>>>
>>> If it were easy for a computer program to attain the probability of an
>>> event based on observations of past events then this kind of discussion
>>> would not be relevant to AGI.  The problem is that events are actually
>>> complexities which are not only composed of distinct 'kinds' of events
>>> and
>>> some background 'noise' but of different variations of 'kinds' of events
>>> and a lot of other events.  The science of using probability and
>>> statistics is premised on the methodical actions of an agent who is not
>>> only intelligent but highly trained in the science of the applied
>>> statistical methods.  The idea that intelligence can be founded on
>>> statistics is backwards.
>>>  Jim Bromer
>>>
>>> On Tue, Feb 19, 2013 at 7:56 PM, Charles Hixson <
>>> [email protected]> wrote:
>>>
>>>>  On 02/19/2013 11:03 AM, Piaget Modeler wrote:
>>>>
>>>>
>>>>  I'm sure this topic has been discussed before.  Sorry for rehashing it
>>>> if so. I have a specific question I'd like to answer.
>>>>
>>>>
>>>>  In designing a cognitive system, someone made a criticism that utterly
>>>> confounded me.  And got me thinking.
>>>>
>>>>  The system receives sensory data sets from the world and transforms
>>>> them into percept propositions which it asserts to
>>>> its memory.  Each percept proposition is activated when it is asserted.
>>>>  Infereneces are made from these percepts.
>>>> These initial percepts and its inferences are called "Observables".
>>>> All
>>>> observables can be activated, but there is only a
>>>> notion of activation.
>>>>
>>>>  Next, the system can predict that these observables will recur at some
>>>> point.  But the prediction refers only to predicting
>>>> the re-activation of observables.
>>>>
>>>>  Then some one asked, where is the notion of TRUTH in your system.  I
>>>> was flabbergasted.  Speechless. Then I asked
>>>> well what is truth?  I checked wikipedia.  (
>>>> http://en.wikipedia.org/wiki/Truth )
>>>>
>>>>
>>>>  It turns out that when someone says something is true, it means a very
>>>> many things:
>>>>
>>>>  a) It means that the statement is logically consistent (validity),
>>>> b) that the statement corresponds, concurs, or conforms to reality
>>>> (verity),
>>>> c) that one is sure of the statement (certainty / confidence),
>>>> d) that the statement is likely to occur rather than unlikely
>>>> (Likelihood), and
>>>> e) that we agree with the statement (agreement).
>>>>
>>>>  So my questions are:
>>>>
>>>>  (1) Is truth necessary or important to a cognitive system?
>>>> (2) Which notion of truth should a cognitive system model?
>>>> (3) How do we ascribe truth (values) to sensory input or inferences
>>>> derived from sensory input?
>>>>
>>>>  Your thoughts?
>>>>
>>>>  ~PM.
>>>>
>>>>
>>>> ------------------------------------------------------------------------------------------------------------------------------------------------
>>>> *Confidential *- *This message is meant solely for the intended
>>>> recipient. Please do not copy or forward this message without *
>>>> *the consent of the sender. If you have received this message in error,
>>>> please delete the message and notify the sender.*
>>>>     *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>> <https://www.listbox.com/member/archive/rss/303/232072-58998042> |
>>>> Modify <https://www.listbox.com/member/?&;> Your Subscription
>>>> <http://www.listbox.com>
>>>>
>>>> Truth is an illusion.  It is the belief that what you believe to be
>>>> most
>>>> likely is, in fact, inevitable.
>>>>
>>>> An AI doesn't need the concept of truth...except to communicate with
>>>> people.  Internally it can operate off of graded degrees of
>>>> probability,
>>>> cost, benefit, etc.  When communicating with people it needs to
>>>> condense
>>>> that so that when something has more than a certain amount of
>>>> probability,
>>>> and the benefit of asserting it is sufficiently large, and the cost of
>>>> being wrong is sufficiently small, then it synopsizes this as
>>>> proclaiming
>>>> "truth".  It's my belief that people operate in the same way, though
>>>> this
>>>> is disguised because different people use different constraints on
>>>> things
>>>> like "What is probable enough?".  Also note that the cost and the
>>>> benefit
>>>> are figured on the basis of the cost/benefit to the entity proclaiming
>>>> a
>>>> truth rather than on those accepting it.
>>>>
>>>> So perhaps we would want a sufficiently capable AI to avoid talking
>>>> about truth, and instead talk about what the probabilities are, and
>>>> what
>>>> costs and benefits can be expected.  It's a bit harder to understand,
>>>> but
>>>> it strikes me as much safer.
>>>>
>>>> --
>>>> Charles Hixson
>>>>
>>>>    *AGI* | Archives <https://www.listbox.com/member/archive/303/=now>
>>>> <https://www.listbox.com/member/archive/rss/303/10561250-470149cf> |
>>>> Modify<https://www.listbox.com/member/?&;>Your Subscription
>>>> <http://www.listbox.com>
>>>>
>>>
>>>
>>
>
>
>
> -------------------------------------------
> AGI
> Archives: https://www.listbox.com/member/archive/303/=now
> RSS Feed: https://www.listbox.com/member/archive/rss/303/11943661-d9279dae
> Modify Your Subscription:
> https://www.listbox.com/member/?&;
> Powered by Listbox: http://www.listbox.com
>


-------------------------------------------
AGI
Archives: https://www.listbox.com/member/archive/303/=now
RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424
Modify Your Subscription: 
https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657
Powered by Listbox: http://www.listbox.com

Reply via email to