[ 
https://issues.apache.org/jira/browse/LUCENE-1692?page=com.atlassian.jira.plugin.system.issuetabpanels:comment-tabpanel&focusedCommentId=12721475#action_12721475
 ] 

Robert Muir commented on LUCENE-1692:
-------------------------------------

probably, fixed it and testing with ant now. ill upload it at least so you can 
verify the behavior i've discovered.

do you want me to include patch with the two bugfixes (chinese empty token and 
thai offsets), or give you something separate for those?

for the other 2 bugs:
fixing the Thai tokentype bug, well its really a bug in the standardtokenizer 
grammar. i wasn't sure you wanted to change that at this moment, but if you 
want it fixed let me know!
in my opinion: fix for DutchAnalyzer is to deprecate/remove the contrib 
completely, since it claims to do snowball stemming, why shouldnt someone just 
use the Dutch snowball stemmer from the contrib/snowball package!

  


> Contrib analyzers need tests
> ----------------------------
>
>                 Key: LUCENE-1692
>                 URL: https://issues.apache.org/jira/browse/LUCENE-1692
>             Project: Lucene - Java
>          Issue Type: Test
>          Components: contrib/analyzers
>            Reporter: Robert Muir
>            Assignee: Michael McCandless
>             Fix For: 2.9
>
>         Attachments: LUCENE-1692.txt, LUCENE-1692.txt, LUCENE-1692.txt, 
> LUCENE-1692.txt
>
>
> The analyzers in contrib need tests, preferably ones that test the behavior 
> of all the Token 'attributes' involved (offsets, type, etc) and not just what 
> they do with token text.
> This way, they can be converted to the new api without breakage.

-- 
This message is automatically generated by JIRA.
-
You can reply to this email to add a comment to the issue online.


---------------------------------------------------------------------
To unsubscribe, e-mail: java-dev-unsubscr...@lucene.apache.org
For additional commands, e-mail: java-dev-h...@lucene.apache.org

Reply via email to