I wrote a simple reinforcement learner which includes the line of code:
printf("Ouch!\n");
So I don't see communication of qualia as a major obstacle to AGI.
Or do you mean something else by qualia?
On Mon, Sep 24, 2018, 5:21 AM John Rose <[email protected]> wrote:
> > -----Original Message-----
> > From: Matt Mahoney via AGI <[email protected]>
> >
> > I was applying John's definition of qualia, not agreeing with it. My
> definition is
> > qualia is what perception feels like. Perception and feelings are both
> > computable. But the feelings condition you to believing there is
> something
> > magical and mysterious about it.
> >
>
> And what I'm saying is that the communication of qualia is important for
> general intelligence in a system of agents. And how do agents interpret the
> signals, process and recommunicate them.
>
> But without fully understanding qualia since they're intimately intrinsic
> to agent experience we can still explore their properties by answering
> questions such as: What is an expression of the information distance
> between qualia of differing agents with same stimuli? How do qualia map to
> modeled environment? How do they change over time in a system of learning
> agents? What is the compressional loss into communication? And how do
> multi-agent models change over time from communicated and decompressed
> qualia.
>
> And what is the topology of qualia variance within an agent related to the
> complexity classes of environmental strategy?
>
> And move on to questions such as can there be enhancements to agent
> language to accelerate learning in a simulated system? And enhancements to
> agent structure?
>
> John
>
------------------------------------------
Artificial General Intelligence List: AGI
Permalink:
https://agi.topicbox.com/groups/agi/T9c94dabb0436859d-M2f5ee95344c2383d5154fb35
Delivery options: https://agi.topicbox.com/groups/agi/subscription