Can it dream?
Harry

On Mon, Apr 10, 2023 at 11:49 AM Alain Sepeda <alain.sep...@gmail.com>
wrote:

> There are works to allow LLM to discuss in order to have reflection...
> I've seen reference to an architecture where two GPT instances talk to
> each other, with different roles, one as a searcher, the other as a
> critic...
> Look at this article.
> LLM may just be the building block of something bigger...
>
> https://www.nextbigfuture.com/2023/04/gpt4-with-reflexion-has-a-superior-coding-score.html
>
> add to that, they can use external applications (plugin), talk to
> generative AI like Dall-E...
>
> Many people say it is not intelligent, but are we ?
> I see AI making mistakes very similar to the one I do when I'm tired, or
> beginner...
>
> The real difference is that today, AI are not the fruit of a Darwinian
> evolution, with struggle to survive, dominate, eat or be eaten, so it's
> less frightening than people or animals.
> The only serious fear I've heard is that we become so satisfied by those
> AIs, that we delegate our genetic evolution to them, and we lose our
> individualistic Darwinian struggle to survive, innovate, seduce a partner,
> enjoying a bee-Hive mentality, at the service of the AI system, like
> bee-workers and bee-queen... The promoter of that theory estimate it will
> take a millennium.
> Anyway there is nothing to stop, as if a majority decide to stop
> developing AI, a minority will develop them at their service, and China is
> ready, with great experts and great belief in the future. Only the West is
> afraid. (there is a paper on that circulating, where fear of AI is linked
> to GDP/head)
>
>
> Le lun. 10 avr. 2023 à 16:47, Jed Rothwell <jedrothw...@gmail.com> a
> écrit :
>
>> I wrote:
>>
>>
>>> Food is contaminated despite our best efforts to prevent that.
>>> Contamination is a complex process that we do not fully understand or
>>> control, although of course we know a lot about it. It seems to me that as
>>> AI becomes more capable it may become easier to understand, and more
>>> transparent.
>>>
>>
>> My unfinished thought here is that knowing more about contamination and
>> seeing more complexity in it has improved our ability to control it.
>>
>>
>> Sean True <sean.t...@gmail.com> wrote:
>>
>> I think it’s fair to say no AGI until those are designed in, particularly
>>> the ability to actually learn from experience.
>>>
>>
>> Definitely! ChatGPT agrees with you!
>>
>

Reply via email to