Hello Petr,

> we have changed the default word tokenizer to properly account for
> czech accents and now need to rebuild all the indexes. All went well
> except that the global virtual index refuses to reindex.

I'm unsure that the way to tackle this is in the word tokenizer;
shouldn't it be done in the strip_accents funcion?  Some years ago I
proposed to change its implementation:

 https://github.com/inveniosoftware/invenio/issues/425

I did not popose a patch because I don't know how to implement the
tests.

Best regards,

Ferran

Reply via email to