or even hugely.

On 13 June 2014 19:49, LizR <[email protected]> wrote:

> The closest I've seen to a computer programme behaving in what might be
> called an intelligent manner was in one of Douglas Hofstadter's books. (I
> think it designed fonts or something?) At least as he described it, it
> seemed to be doing something clever, but nowhere near the level needed to
> pass the Turing Test "for real" - but that's the point, I suppose. You
> can't expect to write a programme to pass the TT until you've written one
> that can do tiny bits of cleverness, and then another one that uses those
> tiny bits to be a bit more clever, and so on. In a way this is like the way
> that SF writers thought we'd have soon robot servants that were almost
> human, and might even rebel ... without realising that the process would
> have to be higely, mind-bogglingly incremental.
>
>
>
> On 13 June 2014 18:35, Pierz <[email protected]> wrote:
>
>> Meh. The whole thing really just illustrates a fundamental problem with
>> our current conception of AI -at least as it manifests in such 'tests'. It
>> is perfectly clear that the Eliza-like program here just has some bunch of
>> pre-prepared statements to regurgitate and the programmers have tried to
>> wire these responses up to questions in such a way that they appear to be
>> legitimate, spontaneous answers. But intelligence consists in the invention
>> of those responses. This is always the problem with computer programs, at
>> least as they exist today: they really just crystallize acts of human
>> intelligence into strict, repeatable procedures. Even chess programs, which
>> are arguably the closest thing we have to computer intelligence, depend on
>> this crystallized intelligence, because the pruning rules and strategic
>> heuristics they rely upon draw on deep human insights that the computer
>> could never have arrived at itself. As humans we resemble computers to the
>> extent that we have automated our behaviour - when we regurgitate a "good
>> how are you?" in response to a social enquiry as to how we are we are
>> fundamentally behaving like Eliza. But when we engage in real conversation
>> or any other form of novel problem solving, we don't seem very
>> computer-like at all, the point that Craig makes (ad nauseam).
>>
>> On Friday, June 13, 2014 5:20:16 AM UTC+10, John Clark wrote:
>>
>>> On Wed, Jun 11, 2014 at 4:22 PM, <[email protected]> wrote:
>>>
>>> > If the TT has been watered down, then the first question for me would
>>>> be "doesn't this logically pre-assume a set of explicit standards existed
>>>> in the first place"?
>>>>
>>>
>>> My answer is "no". So am I a human or a computer?
>>>
>>> > Has there ever been a robust set of standards?
>>>>
>>>
>>> No, except that whatever procedure you use to judge the level of
>>> intelligence of your fellow Human Beings it is only fair that you use the
>>> same procedure when judging machines. I admit this is imperfect, humans can
>>> turn out to be smarter or dumber than originally thought, but it's the only
>>> tool we have for judging such things. If the judge is a idiot then the
>>> Turing Test doesn't work very well, or if the subject is a genius but
>>> pretending to be a idiot you well also probably end up making the wrong
>>> judgement but such is life, you do the best you can with the tools at hand.
>>>
>>> By the way, for a long time machines have been able to beautifully
>>> emulate the behavior of two particular types of humans, those in a coma and
>>> those that are dead.
>>>
>>>    John K Clark
>>>
>>>
>>>
>>>  --
>> You received this message because you are subscribed to the Google Groups
>> "Everything List" group.
>> To unsubscribe from this group and stop receiving emails from it, send an
>> email to [email protected].
>> To post to this group, send email to [email protected].
>> Visit this group at http://groups.google.com/group/everything-list.
>> For more options, visit https://groups.google.com/d/optout.
>>
>
>

-- 
You received this message because you are subscribed to the Google Groups 
"Everything List" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To post to this group, send email to [email protected].
Visit this group at http://groups.google.com/group/everything-list.
For more options, visit https://groups.google.com/d/optout.

Reply via email to