W dniu 02.02.2016 o 18:08, Andriy Rysin pisze: > Hey Marcin > > this is great addition, though I have one remark. Besides valency > information some other type of information could be useful too (if we > starting to head this direction). E.g. I have rules in Ukrainian that > suggests superlative form for adjective when "самий" (very) + base > form is used. Currently I have the relation between base form and > comparative/superlative forms encoded in the dictionary but in general > this is higher-level information that should be stored outside of the > tag dictionary.
I would argue that in some languages (at least in Polish and English) this is not a semantic-level information, this is a grammatical information, or morphosyntactic information. > > I am wondering if we could develop more generic approach for such > additional (semantic) information, e.g. split each type of this info > into category and allow generic references in the token/exception, > something like this: > > <token > semantic_info="<semantic_category_name>:<WHATEVER_STRING_FROM_THAT_CATEGORY>"/> > > or even as a subelement (I assume semantic information can get pretty > long/complicated so child element may be better choice and will allow > to add new attributes easily on it later) > > <token> > <semantic category="<semantic_category_name1>" > value="<WHATEVER_STRING_FROM_THAT_CATEGORY>"/> > <semantic category="<semantic_category_name2>" > value="<WHATEVER_STRING_FROM_THAT_CATEGORY>"/> > </token> > > so in valency case you described (1st case) it could be: > > <token postag="verb"> > <semantic category="valence" value="WHATEVER_STRING"/> > </token> Valency is definitely not a semantic category: https://en.wikipedia.org/wiki/Valency_(linguistics) But your approach seems quite elegant. I would argue that valency is one kind of information that should be treated as key-value <token postag="verb"> <valency category="noun phrase" value="accusative"/> </token> This would match a verb that takes an accusative noun phrase (of course, the values would be defined per valency lexicon in a language). There are free valency lexicons for many languages beside Polish. > > Thus if we add other semantic information into LT we can use this info > in the logic without changing the LT core. The core XML parsing will have to be changed anyway. Best, Marcin > > Thanks > Andriy > > 2016-01-28 7:30 GMT-05:00 Marcin Miłkowski <list-addr...@wp.pl>: >> Hi all, >> >> To allow for better disambiguation and have better rules, I need to >> include a valency dictionary with LT. These are dictionaries that >> specify which grammatical cases or prepositions go with which verbs etc. >> There are such resources for many languages that we support. And using >> these resources, we could enrich POS tag disambiguation a lot (I'm using >> a horribly long regular expression right now instead of a dictionary, >> for example), and write up a lot of important rules. >> >> The obvious choice for representing the dictionary (which is available >> for Polish on a fairly liberal license) is to use a finite-state lexicon >> that we normally use for taggers. The dictionary will be applied after >> tagging because valency dictionary will require POS tag + lexeme >> information. In Polish, the entries look like this: >> >> absurdalny: pewny: : : : {prepnp(dla,gen)} >> absurdalny: pewny: : : : {prepnp(w,loc)} >> absurdalny: pewny: : pred: : {prepnp(dla,gen)}+{cp(gdy)} >> absurdalny: pewny: : pred: : {prepnp(dla,gen)}+{cp(int)} >> absurdalny: potoczny: : pred: : {prepnp(dla,gen)}+{cp(jak)} >> absurdalny: pewny: : pred: : {prepnp(dla,gen)}+{cp(jeśli)} >> absurdalny: pewny: : pred: : {prepnp(dla,gen)}+{cp(kiedy)} >> absurdalny: pewny: : pred: : {prepnp(dla,gen)}+{cp(że)} >> absurdalny: pewny: : pred: : {prepnp(dla,gen)}+{cp(żeby)} >> >> But for French (see http://bach.arts.kuleuven.be/dicovalence/) they are >> paragraph-based: >> >> VAL$ abaisser: P0 P1 >> VTYPE$ predicator simple >> VERB$ ABAISSER/abaisser >> NUM$ 10 >> EG$ il faudra abaisser la persienne >> TR_DU$ laten zakken, neerhalen, neerlaten, doen dalen >> TR_EN$ let down, lower >> FRAME$ subj:pron|n:[hum], obj:pron|n:[nhum,?abs] >> P0$ (que), qui, je, nous, elle, il, ils, on, (ça), (ceci), celui-ci, >> ceux-ci >> P1$ que, la, le, les, en Q, ça, ceci, celui-ci, ceux-ci >> RP$ passif être, se passif >> AUX$ avoir >> >> >> VAL$ abaisser: P0 (P1) >> VTYPE$ predicator simple >> VERB$ ABAISSER/abaisser >> NUM$ 20 >> EG$ il a raconté cette anecdote pour m'abaisser >> TR_DU$ vernederen, kleineren >> TR_EN$ humiliate >> FRAME$ subj:pron|n:[hum], ?obj:pron|n:[hum] >> P0$ (que), qui, je, nous, elle, il, ils, on, (ça), celui-ci, ceux-ci, >> (ça(de_inf)) >> P1$ 0, qui, te, vous, la, le, les, se réc., en Q, celui-ci, ceux-ci, >> l'un l'autre >> RP$ passif être, se faire passif >> AUX$ avoir >> >> >> I would also add a new list of valency attributes to every >> AnalyzedToken, simply as a string value (parsing the string would be >> overkill because there might be different ways of encoding valency >> information for different languages), with appropriate getters and setters. >> >> To use this, we will need some additional attributes in XML elements: >> token and exception. The following notation seems to be fairly fine: >> >> <token postag="verb" valence="WHATEVER_STRING"/> >> >> I think matching valence should, by default, use regular expressions. >> I'm not sure if negation is needed. >> >> Alternatively, we could go for attribute-value pairs but the problem >> would be how to make this language-independent and use this in our XML >> notation. The easiest way would be then probably define types and their >> ids just the way I did this for unification. For French, this could look >> like: >> >> <token>abaisser<valency id="AUX">avoir</valency><valency >> id="VTYPE">predicator simple</valency></token> >> >> Of course, the textual values of <valency> could be made into regexes >> and negated. >> >> For this, of course we would need to write up a key-value parser for all >> particular valency dictionaries. But this would definitely speed up >> matching and would make writing rules much easier. >> >> Best >> >> ------------------------------------------------------------------------------ >> Site24x7 APM Insight: Get Deep Visibility into Application Performance >> APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month >> Monitor end-to-end web transactions and take corrective actions now >> Troubleshoot faster and improve end-user experience. Signup Now! >> http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140 >> _______________________________________________ >> Languagetool-devel mailing list >> Languagetool-devel@lists.sourceforge.net >> https://lists.sourceforge.net/lists/listinfo/languagetool-devel > > ------------------------------------------------------------------------------ > Site24x7 APM Insight: Get Deep Visibility into Application Performance > APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month > Monitor end-to-end web transactions and take corrective actions now > Troubleshoot faster and improve end-user experience. Signup Now! > http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140 > _______________________________________________ > Languagetool-devel mailing list > Languagetool-devel@lists.sourceforge.net > https://lists.sourceforge.net/lists/listinfo/languagetool-devel > ------------------------------------------------------------------------------ Site24x7 APM Insight: Get Deep Visibility into Application Performance APM + Mobile APM + RUM: Monitor 3 App instances at just $35/Month Monitor end-to-end web transactions and take corrective actions now Troubleshoot faster and improve end-user experience. Signup Now! http://pubads.g.doubleclick.net/gampad/clk?id=267308311&iu=/4140 _______________________________________________ Languagetool-devel mailing list Languagetool-devel@lists.sourceforge.net https://lists.sourceforge.net/lists/listinfo/languagetool-devel