Hi everyone!
My name is Fran and I am a new UIMA user, attracted by its capabilities
in natural language processing.
I am trying to develop a component that should receive tokens as input
in order to create its own annotations as output. I though that uima
modularity permits the use of my component by adding it in any pipeline
including a tokenizer before. However, I have to define what java
implementation of token is expected by my annotator. In this way, my
annotator would be dependent on the tokenizer so it only understands
those tokenizers that produce a concrete implementation of token defined
in my annotator. This does not permit the reuse of my annotator in
different pipelines, for example, when a tokenizer produces TokenA.java
and another one TokenB.java
My question is: is there any standard implementation of token so I can
use it as input for my annotator? If not, what is the strategy for
reusing uima components in different pipelines?
I hope that my question is clear. Thanks in advance.