I have been trying to find a similar evaluation tool, but since each
domain is different, and measuring degree of domain-coverage is
application dependent, this is difficult and subjective.
The only tool I found, based on quality metrics, is OOPS! Pitfall
scanner. This will tell you the number and type of bad design choices in
each ontology. However, domain-wise, you will have to assess by your own
criteria.
Best,
Natalia
--
Natalia Díaz Rodríguez. PhD. Student, Computer Engineering
Department of IT. Åbo Akademi University, Turku, Finland
Dept. of Computer Science and Artificial Intelligence. University of Granada,
Spain
+34 669685055
https://research.it.abo.fi/personnel/ndiaz
On 20.2.2014 22:04, Ghislain Atemezing wrote:
Hi Bernadette,
Many thanks for your answer.
Can you be a bit more specific? Are you looking for evaluation criteria?
Yes, actually I am looking for evaluation criteria. Let’s say two
ontologies X and Y are built for the same domain D, without any
Competency Questions (CQ). X and Y reused differently terms from other
namespaces. If I was asked to run an experiment with users to evaluate
(both qualitative and quantitative) those two ontologies , are they
already framework to help me in this task?
There is of course basic vocabulary considerations aimed at helping
people review a vocabulary to evaluate its usefulness in the Best
Practices document but presumably you're looking for more or
something different?? [1]
Thanks again for your advices.
Cheers,
Ghislain