Thanks Matt, that was insightful. ~PM > Date: Sat, 24 Jan 2015 14:53:09 -0500 > Subject: Re: [agi] IS > From: [email protected] > To: [email protected] > > On Sat, Jan 24, 2015 at 11:59 AM, Piaget Modeler via AGI > <[email protected]> wrote: > > How do you represent IS ? Do you differentiate IS from TYPE-OF (i.e., > > IS-A), or INSTANCE-OF ? > > > > Take for example, > > > > IS(apple, fruit) - TYPE-OF > > IS(John_Smith, Politician) - INSTANCE-OF > > IS(my_coat, green) - ??? > > IS could also be Islamic State. > > Language evolves. Knowledge representation systems that assign a fixed > set of meanings to words have a long history of failure. I don't know > why anyone still pursues this approach. > > I understand that a structured knowledge representation doesn't > require a supercomputer like a neural language model. Initially it > looks like the right approach too, because rule coverage has a power > law distribution, with the IS-A construct ranked right at the top. You > can cover half of the language with just a few hundred rules. The > problem is that nobody knows how many rules you need to cover the > other half. Doug Lenat (Cyc) has been plugging away at it for over 30 > years. Apparently it was a lot more than he thought. > > First, our brains evolved to be able to learn language. Then language > evolved to have a structure that can be learned in a few years on a > noisy, massively parallel 10 petaflop computer with 100 terabits of > memory. > > The rules (I believe there are 10^8 to 10^9 of them) can be grouped > roughly into lexical, semantics, and grammar. Rules in each set can be > learned after learning a large portion of the previous set. Note that > I listed semantics before grammar, which is the opposite of the way > most parsers work (or actually, don't work). Children learn the rules > for splitting continuous text into words by age 7 to 10 months. They > learn to associate words with other words and with nonverbal > perceptions (grounding) starting around 1 year. They start forming > grammatically correct sentences around age 2-3. > > We can divide grammar rules into categorization (X is a noun) and > rules for ordering words (adjectives precede nouns in English). Most > rules are very specific. For example, we say "salt and pepper", not > "pepper and salt". We use high level grammar rules to solve math > problems, so there is an obvious learning hierarchy here too. > > I am not sure how much this helps. Most of us don't have the resources > to do the 10^24 operations needed to properly learn natural language, > other than in our own brains. We usually compromise and do something > we can afford, but there is an obvious tradeoff between CPU, memory, > and text prediction accuracy which I have documented at > http://mattmahoney.net/dc/text.html > A highly optimized program running for a week on a high end desktop > with 32 GB of memory still falls well short of what humans can do. > > -- > -- Matt Mahoney, [email protected] > > > ------------------------------------------- > AGI > Archives: https://www.listbox.com/member/archive/303/=now > RSS Feed: https://www.listbox.com/member/archive/rss/303/19999924-4a978ccc > Modify Your Subscription: https://www.listbox.com/member/?& > Powered by Listbox: http://www.listbox.com
------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
