Hello Petr, > we have changed the default word tokenizer to properly account for > czech accents and now need to rebuild all the indexes. All went well > except that the global virtual index refuses to reindex.
I'm unsure that the way to tackle this is in the word tokenizer; shouldn't it be done in the strip_accents funcion? Some years ago I proposed to change its implementation: https://github.com/inveniosoftware/invenio/issues/425 I did not popose a patch because I don't know how to implement the tests. Best regards, Ferran

