9pm EET works for me.. is that 1.5 hours from now or 25.5 hours from now?

On Thu, Apr 1, 2021 at 8:19 AM Michele Thiella <[email protected]>
wrote:

> Could it be around 9pm EET?
> it's a completely different time but should it be available for everyone?
>
> Michele
> Il giorno giovedì 1 aprile 2021 alle 16:29:53 UTC+2 Nil ha scritto:
>
>> Sure! The place is
>>
>> https://meet.jit.si/proto-agi
>>
>> the time is
>>
>> 10:45am EET
>>
>> Unfortunately probably too early if you're in the US.
>>
>> Michele, maybe we could do a last minute change to fit the US timezone
>> as well? With the risk of adding confusion though.
>>
>> I'll try to record the call, BTW.
>>
>> Nil
>>
>> On 4/1/21 5:08 PM, Douglas Miles wrote:
>> > May I sit in on the meeting as a fly on the wall?
>> > If so, when/how shall I connect?
>> >
>> > Thanks in advance!
>> > Douglas Miles
>> >
>> > On Thu, Apr 1, 2021 at 2:52 AM Michele Thiella <[email protected]
>> > <mailto:[email protected]>> wrote:
>> >
>> > Hi Nil,
>> > you're right! currently EET corresponds to the Italian time!
>> > Great, then I might be a few minutes late because I have a lesson
>> > first. But surely 10.45am EET can work!
>> >
>> > Also for me, no problems for those who want to join!
>> > Thanks for the PLN link. See you tomorrow.
>> >
>> > Michele
>> > Il giorno mercoledì 31 marzo 2021 alle 14:40:38 UTC+2 Nil ha scritto:
>> >
>> > Hi Michele,
>> >
>> > On 3/27/21 12:12 PM, Michele Thiella wrote:
>> > > Is there any recommended book/paper to study before the code
>> > of PLN rules?
>> >
>> > Search for Probabilistic Logic Networks in
>> >
>> >
>> https://wiki.opencog.org/w/Background_Publications#Books_Directly_Related_to_OpenCog_AI
>> > <
>> https://wiki.opencog.org/w/Background_Publications#Books_Directly_Related_to_OpenCog_AI>
>>
>> >
>> >
>> > > For the meeting, could it be at 11.30am EET?
>> >
>> > 11:30am EET works for me. But maybe you mean 10:30am EET. With
>> > daylight saving time it seems EET corresponds to Italy time. I'm
>> > not
>> > sure so double check but anyway 10:30am Italy time works for me.
>> >
>> > Nil
>> >
>> > >
>> > > Michele
>> > >
>> > > Il giorno venerdì 26 marzo 2021 alle 08:56:11 UTC+1 Nil ha
>> > scritto:
>> > >
>> > > On 3/25/21 9:03 PM, Michele Thiella wrote:
>> > > > Can I ask you to say something about tree of decisions in Eva?
>> > > Was it a
>> > > > separate scheme/python module that analyzed SequentialAnd?
>> > > > While i'm at it, I can't place some components in your
>> > architecture:
>> > > > I read Moshe Looks thesis on MOSES and what I found on
>> > OpenPsi.
>> > > But in
>> > > > practice what were they used for?
>> > >
>> > > MOSES is a program learner. In principle it could learn any
>> > program, in
>> > > practice it is mostly used to learn multivariable boolean
>> > functions (as
>> > > it doesn't work very well on anything else, so far anyway).
>> > >
>> > > See for more info
>> > >
>> > >
>> > https://wiki.opencog.org/w/Meta-Optimizing_Semantic_Evolutionary_Search
>> > <
>> https://wiki.opencog.org/w/Meta-Optimizing_Semantic_Evolutionary_Search>
>> >
>> > >
>> > <
>> https://wiki.opencog.org/w/Meta-Optimizing_Semantic_Evolutionary_Search
>> > <
>> https://wiki.opencog.org/w/Meta-Optimizing_Semantic_Evolutionary_Search>>
>>
>> >
>> > >
>> > >
>> > > > Finally, in practice what does PLN do/have more than URE?
>> > >
>> > > The URE is a generic rewriting system, that needs a rule set to
>> > > operate.
>> > >
>> > > See for more info
>> > >
>> > > https://wiki.opencog.org/w/Unified_rule_engine
>> > <https://wiki.opencog.org/w/Unified_rule_engine>
>> > > <https://wiki.opencog.org/w/Unified_rule_engine
>> > <https://wiki.opencog.org/w/Unified_rule_engine>>
>> > >
>> > > Such rule set can be PLN, which has been specifically
>> > tailored to
>> > > handle
>> > > uncertain reasoning
>> > >
>> > > https://github.com/opencog/pln
>> > <https://github.com/opencog/pln> <https://github.com/opencog/pln
>> > <https://github.com/opencog/pln>>
>> > >
>> > > or the Miner, which is has been tailored to find frequent
>> > subgraphs
>> > >
>> > > https://github.com/opencog/miner
>> > <https://github.com/opencog/miner>
>> > <https://github.com/opencog/miner
>> > <https://github.com/opencog/miner>>
>> > >
>> > > or more, though these are the two most used/mature.
>> > >
>> > > Nil
>> > >
>> > > >
>> > > >
>> > > > Before reasoning is possible, one must have a world-model.
>> > This
>> > > > model has several parts to it:
>> > > > * The people in the room, and their 3D coordinates
>> > > > * The objects on the table and their 3D coordinates.
>> > > > * The self-model (current position of robot, and of its
>> > arms, etc.)
>> > > > The above is updated rapidly, by sensor information.
>> > > >
>> > > > Then there is some long-term knowledge:
>> > > > * The names of everyone who is known. A dictionary linking
>> > names to
>> > > > faces.
>> > > >
>> > > > Then there is some common-sense knowledge:
>> > > > * you can talk to people,
>> > > > * you can pick up bottles on a table
>> > > > * you cannot talk to bottles
>> > > > * you cannot pick up people.
>> > > > * bottles can be picked up with the arm.
>> > > > * facial expressions and arm movements can be used to
>> > communicate
>> > > > with people.
>> > > >
>> > > > The world model needs to represent all of this. It also
>> > needs to
>> > > > store all of the above in a representation that is
>> > accessible to
>> > > > natural language, so that it can talk about the position of
>> > its arm,
>> > > > the location of the bottle, and the name of the person it is
>> > > talking to.
>> > > >
>> > > > Reasoning is possible only *after* all of the above has been
>> > > > satisfied, not before.  Attempts to do reasoning before the
>> > above
>> > > > has been built will always come up short, because some
>> > important
>> > > > piece of information will be missing, or will be stored
>> > somewhere,
>> > > > in some format that the reasoning system does not have
>> > access to it.
>> > > >
>> > > > The point here is that people have been building "reasoning
>> > systems"
>> > > > for the last 30 or 40 years. They are always frail and
>> > fragile. They
>> > > > are always missing key information.  I think it is
>> > important to try
>> > > > to understand how to represent information in a uniform
>> > manner, so
>> > > > that reasoning does not stumble.
>> > > >
>> > > >
>> > > > Atomspace:
>> > > >
>> > > >   Concepts: "name" - "3D pose"
>> > > >   - bottle - Na
>> > > >   - table - Na
>> > > >   (Predicate: "over" List ("bottle") ("table"))
>> > > >   Actions:
>> > > >   - Go random
>> > > >   - Go to coord
>> > > >   - Grab obj
>> > > >
>> > > > Goal: (bottle in hand)    // = grab bottle
>> > > >
>> > > > Inference rules: all the necessary rules, i.e.
>> > > > * grab-rule: preconditions: (robot-coord = obj-coord) ...,
>> > > > effects: (obj in hand) ...
>> > > > * coord-rule: if x is in "coord1" and y is over x then y is in
>> > > > "coord1"
>> > > >
>> > > > -> So, robot try backward chaining to find the behavior
>> > tree to
>> > > > run. It doesn't find it, it lacks knowledge, it doesn't know
>> > > > where the bottle is (let's leave out partial trees).
>> > > > -> Go random ...
>> > > > -> Vision sensor recognizes table
>> > > > -> atomspace update: table in coord (1,1,1)
>> > > > -> forward chaining -> bottle in coord (1,1,1)
>> > > > -> backward chaining finds a tree, that is
>> > > > Go to coord (1,1,1) + Grap obj
>> > > > -> goal achieved
>> > > >
>> > > >
>> > > > This is a more-or-less textbook robotics homework
>> > assignment. It has
>> > > > certainly been solved in many different ways by many different
>> > > > people using many different technologies, over the last
>> > 40-60 years.
>> > > > Algorithms like A-star search are one of the research
>> > results of
>> > > > trying to solve the above. The AtomSpace would be a horrible
>> > > > technology to solve the above problem, its too slow, too
>> > bulky, too
>> > > > complicated.
>> > > >
>> > > > The chaining steps can be called "inference", but it is
>> > inference
>> > > > devoid of natural language, devoid of "true understanding".
>> > My goal
>> > > > is to have a conversation with the robot:
>> > > >
>> > > > "What do you see?"
>> > > > "A bottle"
>> > > > "where is it?"
>> > > > "on the table"
>> > > > "can you reach it?"
>> > > > "no"
>> > > > "could you reach it if you move to a different place?"
>> > > > "yes"
>> > > > "where would you move?"
>> > > > "closer to the bottle"
>> > > > "can you please move closer to the bottle?"
>> > > > (robot moves)
>> > > >
>> > > >
>> > > > This is now clear to me, but why natural language?
>> > > > if i didn't want interactions with humans could i do it
>> > differently?
>> > > > A certain variation of the sensor values already represents
>> > "the
>> > > forward
>> > > > movement", I do not need to associate a name with it if I
>> > don't
>> > > speak,
>> > > > also for the Atom "bottle" I could use its ID instead.
>> > > > I don't understand why removing natural language implies
>> > having an
>> > > > inference devoid of "true understanding".
>> > > >
>> > > > Stupid example: If I speak Italian with a French, neither
>> > of us
>> > > > understands the other. But a bottle remains a bottle for
>> > both and
>> > > if I
>> > > > give him my hand he will probably do it too ... or he will
>> > leave
>> > > without
>> > > > saying goodbye.
>> > > >
>> > > > I'm probably missing something big, but until I don't bang
>> > my head
>> > > > against it, I don't see.
>> > > >
>> > > >
>> > > > This can be solved by carefully hand-crafting a chatbot
>> > dialog tree.
>> > > > (The ghost chatbot system in opencog was designed to allow
>> > such
>> > > > dialog trees to be created) Over the decades, many chatbots
>> > have
>> > > > been written. Again: there are common problems:
>> > > >
>> > > > -- the text is hard-coded, and not linguistic.  Minor
>> > changes in
>> > > > wording cause the chatbot to get confused.
>> > > > -- there is no world-model, or it is ad hoc and scattered
>> > over many
>> > > > places
>> > > > -- no ability to perform reasoning
>> > > > -- no memory of the dialog ("what were we talking about?" -
>> > well,
>> > > > chatbots do have a one-word "topic" variable, so the
>> > chatbot can
>> > > > answer "we are talking about baseball", but that's it.
>> > There is no
>> > > > "world model" of the conversation, and no "world model" of
>> > who the
>> > > > conversation was with ("On Sunday, I talked to John about a
>> > bottle
>> > > > on a table and how to grasp it")
>> > > >
>> > > > Note that ghost has all of the above problems. It's not
>> > linguistic,
>> > > > it has no world-model, it has no defined representation
>> > that can be
>> > > > reasoned over, and it has no memory.
>> > > >
>> > > > 20 years ago, it was hard to build a robot that could grasp a
>> > > > bottle. It was hard to create a good chatbot.
>> > > >
>> > > > What is the state of the art, today? Well, Tesla has
>> > self-driving
>> > > > cars, and Amazon and Apple have chatbots that are very
>> > > > sophisticated.  There is no open source for any of this,
>> > and there
>> > > > are no open standards, so if you are a university grad
>> > student (or a
>> > > > university professor) it is still very very hard to build a
>> > robot
>> > > > that can grasp a bottle, or a robot that you can talk to.
>> > And yet,
>> > > > these basic tasks have become "engineering"; they are no
>> > longer
>> > > > "science".  The science resides at a more abstract level.
>> > > >
>> > > > --linas
>> > > >
>> > > >
>> > > > I find the abstract level incredible, both in terms of
>> > beauty and
>> > > > difficulty!
>> > > >
>> > > > Michele
>> > > >
>> > > > --
>> > > > You received this message because you are subscribed to the
>> > Google
>> > > > Groups "opencog" group.
>> > > > To unsubscribe from this group and stop receiving emails
>> > from it,
>> > > send
>> > > > an email to [email protected]
>> > > > <mailto:[email protected]>.
>> > > > To view this discussion on the web visit
>> > > >
>> > >
>> >
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com
>> > <
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com>
>>
>> >
>> > >
>> > <
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com
>> > <
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com>>
>>
>> >
>> > >
>> > > >
>> > >
>> > <
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com?utm_medium=email&utm_source=footer
>> > <
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com?utm_medium=email&utm_source=footer>
>>
>> >
>> > >
>> > <
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com?utm_medium=email&utm_source=footer
>> > <
>> https://groups.google.com/d/msgid/opencog/5ac81cf1-c4cd-40cd-9438-55d8dc3d95f5n%40googlegroups.com?utm_medium=email&utm_source=footer>>>.
>>
>> >
>> > >
>> > >
>> > > --
>> > > You received this message because you are subscribed to the
>> > Google
>> > > Groups "opencog" group.
>> > > To unsubscribe from this group and stop receiving emails from
>> > it, send
>> > > an email to [email protected]
>> > > <mailto:[email protected]>.
>> > > To view this discussion on the web visit
>> > >
>> >
>> https://groups.google.com/d/msgid/opencog/f8d77746-4855-491d-bf65-4bc73d45ca39n%40googlegroups.com
>> > <
>> https://groups.google.com/d/msgid/opencog/f8d77746-4855-491d-bf65-4bc73d45ca39n%40googlegroups.com>
>>
>> >
>> > >
>> > <
>> https://groups.google.com/d/msgid/opencog/f8d77746-4855-491d-bf65-4bc73d45ca39n%40googlegroups.com?utm_medium=email&utm_source=footer
>> > <
>> https://groups.google.com/d/msgid/opencog/f8d77746-4855-491d-bf65-4bc73d45ca39n%40googlegroups.com?utm_medium=email&utm_source=footer>>.
>>
>> >
>> >
>> > --
>> > You received this message because you are subscribed to the Google
>> > Groups "opencog" group.
>> > To unsubscribe from this group and stop receiving emails from it,
>> > send an email to [email protected]
>> > <mailto:[email protected]>.
>> > To view this discussion on the web visit
>> >
>> https://groups.google.com/d/msgid/opencog/c1d4319f-70de-4cda-a6e5-a91c8dd53946n%40googlegroups.com
>> > <
>> https://groups.google.com/d/msgid/opencog/c1d4319f-70de-4cda-a6e5-a91c8dd53946n%40googlegroups.com?utm_medium=email&utm_source=footer>.
>>
>> >
>> > --
>> > You received this message because you are subscribed to the Google
>> > Groups "opencog" group.
>> > To unsubscribe from this group and stop receiving emails from it, send
>> > an email to [email protected]
>> > <mailto:[email protected]>.
>> > To view this discussion on the web visit
>> >
>> https://groups.google.com/d/msgid/opencog/CAER3M5%3D1%2B7jNyXa6zjcoWf8qJ%2BVWkjjZ_x_V%3Dbmp8QrhUVHNTw%40mail.gmail.com
>> > <
>> https://groups.google.com/d/msgid/opencog/CAER3M5%3D1%2B7jNyXa6zjcoWf8qJ%2BVWkjjZ_x_V%3Dbmp8QrhUVHNTw%40mail.gmail.com?utm_medium=email&utm_source=footer>.
>>
>>
> --
> You received this message because you are subscribed to the Google Groups
> "opencog" group.
> To unsubscribe from this group and stop receiving emails from it, send an
> email to [email protected].
> To view this discussion on the web visit
> https://groups.google.com/d/msgid/opencog/a2120141-b415-41f7-8b29-c7cba23db310n%40googlegroups.com
> <https://groups.google.com/d/msgid/opencog/a2120141-b415-41f7-8b29-c7cba23db310n%40googlegroups.com?utm_medium=email&utm_source=footer>
> .
>

-- 
You received this message because you are subscribed to the Google Groups 
"opencog" group.
To unsubscribe from this group and stop receiving emails from it, send an email 
to [email protected].
To view this discussion on the web visit 
https://groups.google.com/d/msgid/opencog/CAER3M5kfoO6YkXYz9dWkKvoLOJgyfSFRPWC1YBHE6AYENPX%2Bvg%40mail.gmail.com.

Reply via email to