Fatmah, On Sun, Mar 24, 2013 at 6:22 AM, Fatmah <[email protected]> wrote:
> we developed some units > What did they do? Steve > > ------------------------------ > *From:* Steve Richfield <[email protected]> > *To:* AGI <[email protected]> > *Sent:* Saturday, March 23, 2013 6:57 PM > > *Subject:* Re: [agi] 40 years of parsing NL... > > PM (and Logan), > > You said in a previous posting that you have experience with L-A. What > have you (or others) done with it? > > I ask because once you sidestep semantic units, it seems to me like you > have thrown the baby out with the bathwater, at least for the usual > applications needing some degree of "understanding". Maybe I just haven't > noticed a good application that doesn't need semantic units, or I haven't > understood a good way to live without them. Sure you can "parse" while > ignoring them, but then of what use is the resulting parse?!!! > > Idioms (of which there are thousands) are a sort of ill-behaved semantic > unit. How do you handle idioms while sidestepping semantic units? > > Logan: Have you been following this discussion? RADP is close enough to > what I am planning to have the same semantic unit needs. Can you help make > sense of this? > > What (if anything) am I missing here? > > Steve > ================= > On Fri, Mar 22, 2013 at 7:08 PM, Steve Richfield < > [email protected]> wrote: > > PM, > > On Fri, Mar 22, 2013 at 5:27 PM, Piaget Modeler <[email protected] > > wrote: > > > Actually, it's more than making a chatbot. It's having a real robot > respond to a person based on linking utterances > (made by either the robot or the person) to the current context (milieu > entities and events). > > I think before you make your Worldcomp presentation it would behoove you > to read the *NEWCAT *and > *Computation of Language* books so that you can adequately articulate the > differences in your approach. > > > We seem to be talking past each other here. My presentation at Worldcomp > need not compare with anything, most especially character-based methods > that don't seem to even recognize what parsing applications need from a > parser, let alone squarely addressing the how to provide what those > applications need. There is SO much that these methods don't on first > glance address. > > Each parsing method seems to need a champion, and you seem to be the > resident champion for L-A grammar here. I know you want to just send me > some hyperlinks and tell me to go away and read some books, but here on > this forum we each learn our own particular areas, and defend against > stones tossed by people defending nearby areas. I tossed a stone your way > when I claimed blinding speed. You tossed a stone back when you explained > that all that was needed to parse was to move about though L-A map of > English grammar. I tossed the stone back, pointing out that losing the > semantic elements (many of which are idioms that don't make much > grammatical sense) throws the baby out with the bath water, because > applications (other than machine translation) are only interested in > semantics, not syntax. Dragging semantics out of a parse tree is a really > BIG job, requiring the SAME tests as other parsing methods. Sure you > produce a parse in a hurry by not doing the job of other parsers, but then > doing that job loses the speed advantage. > > To illustrate some of the challenges, I took a large idiom dictionary and > tried looking up idioms that I commonly use in everyday speech, and only > found about half of them. So much for quality control. How does L-A deal > with idioms? Once you have discarded the low-level semantic elements as > part of putting words into parse trees, recognizing idioms could become > quite difficult. Further. many idioms are ungrammatical. Are you planning > to include idioms as part of the map of the language?!!! > > Anyway, I **DO** want to understand L-A enough to see if it is > significant, or have you understand my method enough to be able to compare > the two, so we can both see the relationships between these two VERY > different things. > > Steve > > > ------------------------------ > Date: Fri, 22 Mar 2013 15:30:59 -0700 > Subject: Re: [agi] 40 years of parsing NL... > > From: [email protected] > To: [email protected] > > PM, > > This guy is talking about a different approach for making a chatbot - > right? If so, he doesn't show any indication of knowing about present > chatbots. Present technology is to have a variety of sentence skeletons, > into which appropriate words and phrases are placed, which seems to work > quite well. > > I would think that promoting a technology would best be done with FREE > documents and other supporting material. I already have the 10,000 most > commonly used words in a file in order of frequency of use, if you or > anyone else wants a copy. > > I believe that my approach will be fast enough to keep up with the > Internet, and I haven't seen any other approach that promises such blinding > speed. In theory, all I need do is get the word out, and wait for folks at > Google, Yahoo, and Facebook to discover it, which is my present plan. > > I also plan to present this at the next WORLDCOMP conference. > > BTW, ***THANKS*** for holding my feet to the fire!!! I plan to adapt > these discussions into the paper I present at WORLDCOMP. > > Steve > =================== > On Fri, Mar 22, 2013 at 1:39 PM, Piaget Modeler <[email protected] > > wrote: > > Roland's next step: > > > http://www.amazon.com/Computational-Linguistics-Talking-Robots-Processing/dp/3642224318/ref=sr_1_1?ie=UTF8&qid=1363984424&sr=8-1&keywords=talking+robots+roland+hausser > > Computational Linguistics and Talking Robots: Processing Content in > Database Semantics > > Publication Date: September 14, 2011 | ISBN-10: 3642224318 | ISBN-13: > 978-3642224317 | Edition: 2011 > The practical task of building a talking robot requires a theory of how > natural language communication works. Conversely, the best way to > computationally verify a theory of natural language communication is to > demonstrate its functioning concretely in the form of a talking robot, the > epitome of human–machine communication. To build an actual robot requires > hardware that provides appropriate recognition and action interfaces, and > because such hardware is hard to develop the approach in this book is > theoretical: the author presents an artificial cognitive agent with > language as a software system called database semantics (DBS). Because a > theoretical approach does not have to deal with the technical difficulties > of hardware engineering there is no reason to simplify the system – instead > the software components of DBS aim at completeness of function and of data > coverage in word form recognition, syntactic–semantic interpretation and > inferencing, leaving the procedural implementation of elementary concepts > for later. In this book the author first examines the universals of natural > language and explains the Database Semantics approach. Then in Part I he > examines the following natural language communication issues: using > external surfaces; the cycle of natural language communication; memory > structure; autonomous control; and learning. In Part II he analyzes the > coding of content according to the aspects: semantic relations of > structure; simultaneous amalgamation of content; graph-theoretical > considerations; computing perspective in dialogue; and computing > perspective in text. The book ends with a concluding chapter, a > bibliography and an index. The book will be of value to researchers, > graduate students and engineers in the areas of artificial intelligence and > robotics, in particular those who deal with natural language processing. > > > For you, Steve, the next step is to write a book about your approach and > sell it for $100 a pop, or $75 for the e-book, > and do a book tour (if possible). > > Then gain some early adopters and market traction. > > The point is to make money WHILE promoting your idea. > > Cheers, > > ~PM > > ------------------------------ > Date: Fri, 22 Mar 2013 12:13:23 -0700 > Subject: [agi] 40 years of parsing NL... > From: [email protected] > To: [email protected] > > > Piaget, Logan, et al, > > We have had some interesting discussions about which method is best and > fastest, but is it even possible?!!! > > My own big wake-up call came many years ago, when I recorded a class I > presented, and had it transcribed with instructions "don't edit it, just > transcribe what I said". It was FULL of fragments, missing words, and even > misstatements, but the class had NO problem grokking what I had said. > > Similarly, just take any unedited posting (you can easily recognize > editing by the lack of ANY spelling errors) and try hand-diagramming its > sentences. They will be better than spoken sentences, but still, you will > have problems with around half of them. > > Several early NL projects set out with dictionaries that identified every > part of speech that each word could be, and programmatically set about > identifying a set of assumptions wherein each sentence would hang together. > Unfortunately, few sentences had exactly one solution, and the presence of > any presumed words fractured the entire process. > > More recently, "ontological" approaches have attempted to sub-divide the > parts of speech, e.g. identifying whether a particular noun can have color, > weight, etc., to assist in assigning the targets of adjectives and adverbs. > > The present consensus seems to be that speech is made to a particular > audience with a particular set of presumed knowledge to use to fill in the > gaps, and an automated listener/reader will NOT be able to understand > "plain English" without similar real-world experience as an intended > reader. Without that experience, lots of gaps and disambiguation errors > will persist regardless of how much programming effort is expended. > > Language translation can skirt many/most of these issues, by preserving > the semantic ambiguities in the translation, to let the reader/listener > figure out what the computer failed to figure out. > > No, there will never ever be "full understanding", if for no other reason > than some of what I say simply doesn't make sense. Instead, what can be > done, and what is needed for present applications, are various forms of > partial understanding. You can see this in throwing some numerical problems > at WolframAlpha.com and watching the parsing of it. It picks out key words > and tries ways of relating them together. Similarly, DrEliza.com picks out > key words and phrases that are associated with symptoms and conditions it > knows about. > > The MOST important part of "understanding" is often identifying what the > writer does NOT know (and the computer does know), sort of a reverse > analysis. I refer to these as "statements of ignorance" and this is an > important part of DrEliza.com > > My parsing proposal was made as a component in a larger system in support > of problem solving and sales (it is just one box among many in figure 1 in > my patent application). My approach appears to be general purpose and > applicable to other applications. Given that a universal parser appears to > be impossible until it can walk among us, and even then will have some > problems, each application must consider what it needs to obtain from the > text/speech to do its job. > > So, when relating performance of parsers, it is important to disambiguate > just WHAT is being performed, e.g. just WHAT is "parsing", and what > applications will a particular approach work best for? > > Logan, what do you see are the "best fit" applications for reverse ascent > descent parsing? > > Piaget, what do you see are the "best fit" applications for LA parsing? > > Any thoughts? > > Steve > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/19999924-5cfde295> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com/> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com/> > > > > > -- > Full employment can be had with the stoke of a pen. Simply institute a six > hour workday. That will easily create enough new jobs to bring back full > employment. > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/19999924-5cfde295> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com/> > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com/> > > > > > -- > Full employment can be had with the stoke of a pen. Simply institute a six > hour workday. That will easily create enough new jobs to bring back full > employment. > > > > > -- > Full employment can be had with the stoke of a pen. Simply institute a six > hour workday. That will easily create enough new jobs to bring back full > employment. > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/18570668-a1f923df> | > Modify <https://www.listbox.com/member/?&> Your Subscription > <http://www.listbox.com/> > > > *AGI* | Archives <https://www.listbox.com/member/archive/303/=now> > <https://www.listbox.com/member/archive/rss/303/10443978-6f4c28ac> | > Modify<https://www.listbox.com/member/?&>Your Subscription > <http://www.listbox.com> > -- Full employment can be had with the stoke of a pen. Simply institute a six hour workday. That will easily create enough new jobs to bring back full employment. ------------------------------------------- AGI Archives: https://www.listbox.com/member/archive/303/=now RSS Feed: https://www.listbox.com/member/archive/rss/303/21088071-f452e424 Modify Your Subscription: https://www.listbox.com/member/?member_id=21088071&id_secret=21088071-58d57657 Powered by Listbox: http://www.listbox.com
