There are many papers published on learning of formal grammars from examples... it's a well-understood problem, addressable via neural nets, statistical learning methods, genetic programming and so forth...
On Thu, Oct 23, 2008 at 10:02 AM, Mark Waser <[EMAIL PROTECTED]> wrote: > Hi. I don't understand the following statements. Could you explain it > some more? > > - Natural language can be learned from examples. Formal language can not. > I think that you're basing this upon the methods that *you* would apply to > each of the types of language. It makes sense to me that because of the > regularities of a formal language that you would be able to use more > effective methods -- but it doesn't mean that the methods used on natural > language wouldn't work (just that they would be as inefficient as they are > on natural languages. > > I would also argue that the same argument applies to the first statement of > following the following two. > > - Formal language must be parsed before it can be understood. Natural > language must be understood before it can be parsed. > > > ----- Original Message ----- > *From:* Matt Mahoney <[EMAIL PROTECTED]> > *To:* agi@v2.listbox.com > *Sent:* Wednesday, October 22, 2008 9:23 PM > *Subject:* Lojban (was Re: [agi] constructivist issues) > > Why would anyone use a simplified or formalized English (with regular > grammar and no ambiguities) as a path to natural language understanding? > Formal language processing has nothing to do with natural language > processing other than sharing a common lexicon that make them appear > superficially similar. > > - Natural language can be learned from examples. Formal language can not. > - Formal language has an exact grammar and semantics. Natural language does > not. > - Formal language must be parsed before it can be understood. Natural > language must be understood before it can be parsed. > - Formal language is designed to be processed efficiently on a fast, > reliable, sequential computer that neither makes nor tolerates errors, > between systems that have identical, fixed language models. Natural language > evolved to be processed efficiently by a slow, unreliable, massively > parallel computer with enormous memory in a noisy environment between > systems that have different but adaptive language models. > > So how does yet another formal language processing system help us > understand natural language? This route has been a dead end for 50 years, in > spite of the ability to always make some initial progress before getting > stuck. > > -- Matt Mahoney, [EMAIL PROTECTED] > > --- On *Wed, 10/22/08, Ben Goertzel <[EMAIL PROTECTED]>* wrote: > > From: Ben Goertzel <[EMAIL PROTECTED]> > Subject: Re: [agi] constructivist issues > To: agi@v2.listbox.com > Cc: [EMAIL PROTECTED] > Date: Wednesday, October 22, 2008, 12:27 PM > > > This is the standard Lojban dictionary > > http://jbovlaste.lojban.org/ > > I am not so worried about word meanings, they can always be handled via > reference to WordNet via usages like run_1, run_2, etc. ... or as you say by > using rarer, less ambiguous words > > Prepositions are more worrisome, however, I suppose they can be handled in > a similar way, e.g. by defining an ontology of preposition meanings like > with_1, with_2, with_3, etc. > > In fact we had someone spend a couple months integrating existing resources > into a preposition-meaning ontology like this a while back ... the so-called > PrepositionWordNet ... or as it eventually came to be called the LARDict or > LogicalArgumentRelationshipDictionary ... > > I think it would be feasible to tweak RelEx to recognize these sorts of > subscripts, and in this way to recognize a highly controlled English that > would be unproblematic to map semantically... > > We would then say e.g. > > I ate dinner with_2 my fork > > I live in_2 Maryland > > I have lived_6 for_3 41 years > > (where I suppress all _1's, so that e.g. ate means ate_1) > > Because, RelEx already happily parses the syntax of all simple sentences, > so the only real hassle to deal with is disambiguation. We could use > similar hacking for reference resolution, temporal sequencing, etc. > > The terrorists_v1 robbed_v2 my house. After that_v2, the jerks_v1 > urinated in_3 my yard. > > I think this would be a relatively pain-free way to communicate with an AI > that lacks the common sense to carry out disambiguation and reference > resolution reliably. Also, the log of communication would provide a nice > training DB for it to use in studying disambiguation. > > -- Ben G > > > On Wed, Oct 22, 2008 at 12:00 PM, Mark Waser <[EMAIL PROTECTED]> wrote: > >> >> IMHO that is an almost hopeless approach, ambiguity is too integral >> to English or any natural language ... e.g preposition ambiguity >> Actually, I've been making pretty good progress. You just always use big >> words and never use small words and/or you use a specific phrase as a >> "word". Ambiguous prepositions just disambiguate to one of >> three/four/five/more possible unambiguous words/phrases. >> >> The problem is that most previous subsets (Simplified English, Basic >> English) actually *favored* the small tremendously over-used/ambiguous words >> (because you got so much more "bang for the buck" with them). >> >> Try only using big unambiguous words and see if you still have the same >> opinion. >> >> >> If you want to take this sort of approach, you'd better start with >> Lojban instead.... Learning Lojban is a pain but far less pain than you'll >> have trying to make a disambiguated subset of English. >> >> My first reaction is . . . . Take a Lojban dictionary and see if you can >> come up with an unambiguous English word or very short phrase for each >> Lojban word. If you can do it, my approach will work and will have the >> advantage that the output can be read by anyone (i.e. it's the equivalent of >> me having done it in Lojban and then added a Lojban -> English translation >> on the end) though the input is still *very* problematical (thus the need >> for a semantically-driven English->subset translator). If you can't do it, >> then my approach won't work. >> >> Can you do it? Why or why not? If you can, do you still believe that my >> approach won't work? Oh, wait . . . . a Lojban-to-English dictionary *does* >> attempt to come up with an unambiguous English word or very short phrase for >> each Lojban word. :-) >> >> Actually, hmmmm . . . . a Lojban dictionary would probably help me focus >> my efforts a bit better and highlight things that I may have missed . . . . >> do you have a preferred dictionary or resource? (Google has too many for me >> to do a decent perusal quickly) >> >> >> >> ----- Original Message ----- >> *From:* Ben Goertzel <[EMAIL PROTECTED]> >> *To:* agi@v2.listbox.com >> *Sent:* Wednesday, October 22, 2008 11:11 AM >> *Subject:* Re: [agi] constructivist issues >> >> >> >> >>> >>> Personally, rather than starting with NLP, I think that we're going to >>> need to start with a formal language that is a disambiguated subset of >>> English >> >> >> >> IMHO that is an almost hopeless approach, ambiguity is too integral to >> English or any natural language ... e.g preposition ambiguity >> >> If you want to take this sort of approach, you'd better start with Lojban >> instead.... Learning Lojban is a pain but far less pain than you'll have >> trying to make a disambiguated subset of English. >> >> ben g >> >> ------------------------------ >> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/> | >> Modify<https://www.listbox.com/member/?&>Your Subscription >> <http://www.listbox.com> >> >> ------------------------------ >> *agi* | Archives <https://www.listbox.com/member/archive/303/=now> >> <https://www.listbox.com/member/archive/rss/303/> | >> Modify<https://www.listbox.com/member/?&>Your Subscription >> <http://www.listbox.com> >> > > > > -- > Ben Goertzel, PhD > CEO, Novamente LLC and Biomind LLC > Director of Research, SIAI > [EMAIL PROTECTED] > > "A human being should be able to change a diaper, plan an invasion, butcher > a hog, conn a ship, design a building, write a sonnet, balance accounts, > build a wall, set a bone, comfort the dying, take orders, give orders, > cooperate, act alone, solve equations, analyze a new problem, pitch manure, > program a computer, cook a tasty meal, fight efficiently, die gallantly. > Specialization is for insects." -- Robert Heinlein > > > ------------------------------ > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > > ------------------------------ > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > > ------------------------------ > *agi* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Ben Goertzel, PhD CEO, Novamente LLC and Biomind LLC Director of Research, SIAI [EMAIL PROTECTED] "A human being should be able to change a diaper, plan an invasion, butcher a hog, conn a ship, design a building, write a sonnet, balance accounts, build a wall, set a bone, comfort the dying, take orders, give orders, cooperate, act alone, solve equations, analyze a new problem, pitch manure, program a computer, cook a tasty meal, fight efficiently, die gallantly. Specialization is for insects." -- Robert Heinlein ------------------------------------------- agi Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/ Modify Your Subscription: https://www.listbox.com/member/?member_id=8660244&id_secret=117534816-b15a34 Powered by Listbox: http://www.listbox.com